CN112101217B - Pedestrian re-identification method based on semi-supervised learning - Google Patents

Pedestrian re-identification method based on semi-supervised learning Download PDF

Info

Publication number
CN112101217B
CN112101217B CN202010970306.5A CN202010970306A CN112101217B CN 112101217 B CN112101217 B CN 112101217B CN 202010970306 A CN202010970306 A CN 202010970306A CN 112101217 B CN112101217 B CN 112101217B
Authority
CN
China
Prior art keywords
samples
sample
pedestrian
new
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010970306.5A
Other languages
Chinese (zh)
Other versions
CN112101217A (en
Inventor
葛永新
高志顺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhenjiang Qidi Digital World Technology Co ltd
Original Assignee
Zhenjiang Qidi Digital World Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhenjiang Qidi Digital World Technology Co ltd filed Critical Zhenjiang Qidi Digital World Technology Co ltd
Priority to CN202010970306.5A priority Critical patent/CN112101217B/en
Publication of CN112101217A publication Critical patent/CN112101217A/en
Application granted granted Critical
Publication of CN112101217B publication Critical patent/CN112101217B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • G06V40/25Recognition of walking or running movements, e.g. gait recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2132Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
    • G06F18/21322Rendering the within-class scatter matrix non-singular
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2132Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
    • G06F18/21322Rendering the within-class scatter matrix non-singular
    • G06F18/21328Rendering the within-class scatter matrix non-singular involving subspace restrictions, e.g. nullspace techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a pedestrian re-identification method based on semi-supervised learning, which comprises the following steps that S100 learns a projection matrix U epsilon R d×c to project an original d-dimensional feature space into a c-dimensional subspace, so that U TX∈Rc×N meets the following conditions in a new subspace: the Euclidean distance between pairs of samples from the same pedestrian is smaller, and the Euclidean distance between pairs of samples from different pedestrians is larger; samples from the same pedestrian are defined as homogeneous samples, and samples from different pedestrians are defined as heterogeneous samples; s200, projecting the new sample into a new subspace by adopting a projection matrix U epsilon R d×c to obtain a predicted sample sequence, wherein the predicted sample sequence is arranged according to the Euclidean distance between the new sample and the samples in the training sample set from small to large. The method fully utilizes the accurately marked labeled sample, and the positive sample is restrained by using the contrast loss function, and simultaneously, the negative sample pair can be fully utilized, so that the identification speed is high, and the identification accuracy is higher.

Description

Pedestrian re-identification method based on semi-supervised learning
Technical Field
The invention relates to the technical field of pedestrian re-identification, in particular to a pedestrian re-identification method based on semi-supervised learning.
Background
Pedestrian re-recognition (Person-identification), also known as pedestrian re-recognition, is a technique that uses computer vision techniques to determine whether a particular pedestrian is present in an image or video sequence. Widely recognized as a sub-problem of image retrieval, given a monitored pedestrian image, the pedestrian image is retrieved across devices. The camera is used for making up the visual limitation of the fixed camera, can be combined with pedestrian detection and pedestrian tracking technologies, and can be widely applied to the fields of intelligent video monitoring, intelligent security and the like.
Although in recent years computer vision practitioners have proposed a number of algorithms from different perspectives for pedestrian re-recognition tasks, attempting to continually boost the recognition rate on the public data set, pedestrian re-recognition remains a very challenging task due to the effects of several realistic factors.
At present, a semi-supervised learning method is generally used for solving the task of re-identifying pedestrians, and the flow is approximately as follows: firstly, automatically labeling a label-free sample; and secondly, uniformly training the labeled samples and the automatically labeled samples, and optimizing the model to ensure that the model has better discrimination capability. And two problems exist in the automatic labeling and the utilization of labeled non-labeled samples:
① The idea of the method used in the automatic labeling of unlabeled samples is to label the new space that is being mapped using a K-Nearest Neighbor (KNN) algorithm. This makes the errors after automatic labeling relatively large if the new spatial discrimination capability learned is insufficient. When the model is trained by using the data with larger labeling errors, the model is very likely to have better generalization capability due to the increase of training samples, but the worse the discrimination capability of the model is caused by the more training.
② When training is performed, only positive sample pairs in the training set are constrained, and negative sample pairs are not concerned, so that the training samples are not fully utilized.
Disclosure of Invention
Aiming at the problems existing in the prior art, the technical problems to be solved by the invention are as follows: the existing method for solving the task of re-identifying pedestrians by using semi-supervised learning has the problems that automatic labeling errors are easily influenced by new spaces obtained by learning and the utilization of training samples is insufficient.
In order to solve the technical problems, the invention adopts the following technical scheme: the pedestrian re-identification method based on semi-supervised learning comprises the following steps:
S100: learning a projection matrix U ε R d×c projects the original d-dimensional feature space into the c-dimensional subspace such that U TX∈Rc×N satisfies in the new subspace: the Euclidean distance between pairs of samples from the same pedestrian is smaller, and the Euclidean distance between pairs of samples from different pedestrians is larger; samples from the same pedestrian are defined as homogeneous samples, and samples from different pedestrians are defined as heterogeneous samples;
S200: and projecting the new sample into a new subspace by adopting a projection matrix U epsilon R d×c to obtain a predicted sample sequence, wherein the predicted sample sequence is arranged according to the Euclidean distance between the new sample and the samples in the training sample set from small to large.
Preferably, the method for learning the projection matrix U e R d×c in S100 specifically includes:
S110, building a training sample set, wherein the training sample set comprises a plurality of samples, the samples comprise labeled samples and unlabeled samples, and the labels of the samples of the same pedestrian in the labeled samples are the same;
Let x= [ X L,XU]∈Rd×N ] denote all training samples, where N is the number of all pictures contained in the training set, d is the length of the feature vector, Representing N L tagged samples,Representing N U unlabeled exemplars;
s120, establishing an objective function as follows:
Wherein L (U) is a regression function, Omega (U) is a regularized constraint, alpha, lambda > 0 is a balance coefficient;
S130, the labeled sample loss function is a contrast loss function: n P sample pairs for sampling And/>If/>And/>Samples from the same pedestrian, then in the new projection space/>And/>The Euclidean distance d n between should be as small as possible, close to 0; conversely, d n should be at least greater than a predetermined threshold margin >0, which would result in a loss if the above conditions are not met;
S140, labeling labels by using label-free sample labels, namely labeling labels by using a method of K nearest neighbors, wherein the loss function of label-free sample labels is as follows:
wherein if U Txi and U Txj meet K nearest neighbors to each other and x i and x j are from different cameras, then take
Otherwise W ij = 0; (8);
after labeling the label-free sample, further restricting the existing subspace by using the labeled sample, wherein the restricting weight is the cosine distance of the two samples in the new projection space;
s150: regularization term: the projection matrix U is constrained using L2,1 norm:
Ω(U)=||U||2,1 (4)。
Preferably, the labeled sample loss function of S130 is:
Wherein:
Preferably, the N P samples sampled in S130 are sampled with a sampling strategy that maximizes the top-k recognition rate, i.e. for each image, all samples with k nearest neighbors are sampled.
Preferably, in S140, the method for labeling the label on the label-free sample by adopting the method of K nearest neighbors includes:
the K nearest neighbor N (x, K) defining sample x is as follows:
N(x,k)={x1,x2,...,xk},|N(p,k)|=k (5);
Where |·| represents the number of samples in the set, then K nearest neighbors R (x, K) to each other are defined as follows:
R(x,k)={xi|(xi∈N(x,k))∧(x∈N(xi,k))} (6)。
compared with the prior art, the invention has at least the following advantages:
(1) The invention uses K nearest neighbors to make the automatic labeling result of the unlabeled sample more reliable.
(2) And fully utilizing the accurately marked label sample. The contrast loss function commonly used in training the deep neural network can be used for restraining the positive sample and fully utilizing the negative sample pair. It should be noted that any type of loss for identification or classification may be used as an alternative to the labeled sample loss function.
(3) In order to enable convenient migration of the subsequent models to the depth model, an end-to-end training approach is used herein, and the training strategy uses a random gradient descent approach. The batch generation strategy for maximizing the top-k recognition rate is provided, and the problems that the convergence speed of the paired training strategy is low under random batches, the model is prevented from being fitted excessively and the like are solved.
Drawings
The K mutual nearest neighbor sampling strategy of the pedestrian re-recognition problem of fig. 1. First row: one picture to be retrieved and its 10 nearest neighbors, where P1-P4 are positive samples and N1-N6 are negative samples. Second row: every two columns are the 10 nearest neighbor images that correspond to the first row of images. The thick line non-chamfered rectangular frame and the thin line rectangular frame with chamfer represent the retrieved picture and the positive sample picture, respectively.
The negative sample closest to the image to be retrieved in fig. 2 is the most difficult negative sample; the first positive sample just smaller than the most difficult negative sample is the moderate positive sample; the intra-block samples are the herein sampling strategy.
Fig. 3 is a suitable positive sample sampling.
Detailed Description
The invention is described in further detail below with reference to the accompanying drawings.
Referring to fig. 1-3, the pedestrian re-recognition method based on semi-supervised learning comprises the following steps:
S100: learning a projection matrix U ε R d×c projects the original d-dimensional feature space into the c-dimensional subspace such that U TX∈Rc×N satisfies in the new subspace: the Euclidean distance between pairs of samples from the same pedestrian is smaller, and the Euclidean distance between pairs of samples from different pedestrians is larger; samples from the same pedestrian are defined as homogeneous samples, and samples from different pedestrians are defined as heterogeneous samples.
The method for learning the projection matrix U epsilon R d×c specifically comprises the following steps:
S110, building a training sample set, wherein the training sample set comprises a plurality of samples, the samples comprise labeled samples and unlabeled samples, and the labels of the samples of the same pedestrian in the labeled samples are the same;
Let x= [ X L,XU]∈Rd×N ] denote all training samples, where N is the number of all pictures contained in the training set, d is the length of the feature vector, Representing N L tagged samples,Representing N U unlabeled exemplars;
s120, establishing an objective function as follows:
Where L (U) is a regression function, in order to make the labeled samples satisfy the same label sample pair in the new space mapped to closer distances, the uncorrelated label sample pair farther distances, As a weighted regression function, the discriminant of the model can be improved by using unlabeled samples, omega (U) is a regularized constraint, features with more discriminant capability can be selected from the original feature space, overfitting is avoided, and alpha, lambda > 0 are balance coefficients;
S130, a labeled sample loss function: the purpose of this constraint is to make full use of the tag information of the tagged sample. In order to simultaneously utilize the positive and negative sample pair constraint, we use a training contrast loss function
Wherein:
n P sample pairs for sampling And/>If/>And/>Samples from the same pedestrian, then in the new projection space/>And/>The Euclidean distance d n between should be as small as possible, close to 0; conversely, d n should be at least greater than a predetermined threshold margin >0, which would result in a loss if the above conditions are not met;
S140, labeling labels by using label-free sample labels, namely labeling labels by using a method of K nearest neighbors, wherein the loss function of label-free sample labels is as follows:
In order to effectively utilize the discrimination information of the unlabeled exemplars and reduce the adverse effect of error labeling on the model, K is adopted to label the unlabeled exemplars instead of K nearest neighbors, and only positive exemplar pairs are constrained in the term. The specific loss function is as follows:
wherein if U Txi and U Txj meet K nearest neighbors to each other and x i and x j are from different cameras, then take
Otherwise W ij = 0 (8);
The significance of this term is that it is believed that the pairs of K mutually nearest neighbors in the learned subspace with discriminatory power are most likely from the same pedestrian. And then, after labeling the unlabeled samples, further restricting the existing subspace by using the labeled samples, wherein the restricting weight is the cosine distance of the two samples in the new projection space.
S150: regularization term: the regularization term is added to make the learned projection matrix more sparse while avoiding the occurrence of overfitting. Here we use L2,1 norm to constrain the projection matrix U:
Ω(U)=||U||2,1 (4)。
As an improvement, the N P samples sampled in S130 is a sampling strategy that maximizes the top-k recognition rate, i.e. for each image, all samples with k nearest neighbors are sampled. In this way, the over-fitting is avoided, and meanwhile, the discrimination information of the labeled sample can be utilized to the maximum extent.
When optimizing using random gradient descent, all samples need to be fed into the model in batches. All samples were randomly sampled, a small portion of the categories were randomly selected each time, and two images were taken for each category. At the time of loss calculation, all pairs of samples that all images in each batch may make up participate in the calculation. In this way, while many pairs of samples can be calculated at a time, the direction of optimization of such sampling may not be the direction that can cause the target to drop most quickly due to the randomness in the class sampling. Each time the optimization is completed, the distances of all samples under the current model will be calculated. To make the target drop faster, only a pair of the most difficult negative samples under the current model are selected for each image, as in fig. 2.
It is noted that some positive pairs of samples have too large intra-class differences due to drastic changes, which would most likely overfit the model if they were trained, as in fig. 3. To avoid this overfitting, a modest positive sample (moderate positive sample) is taken for each picture in the manner of fig. 2, i.e., sampling just less than the first positive sample of the most difficult negative samples. In order to utilize as much information as possible provided by the tagged samples, we propose a sampling strategy that maximizes the top-k recognition rate. As shown in fig. 2, for each image, all samples of its k nearest neighbors are sampled, thus avoiding overfitting while maximizing the discrimination information for labeled samples.
As an improvement, the method for marking the label on the label-free sample by adopting the K nearest neighbors method in S140 is as follows:
As shown in fig. 1, P1-P4 are four positive samples of the picture to be retrieved, but are not arranged in the first four bits of the nearest neighbor picture, and a large error is introduced if the K nearest neighbor result is directly used. However, it is worth noting that the picture to be retrieved and the four positive samples are each K nearest neighbors of each other, which we will refer to as K nearest neighbors of each other. Labeling unlabeled data in this manner reduces error introduction to some extent.
The K nearest neighbor N (x, K) defining sample x is as follows:
N(x,k)={x1,x2,...,xk},|N(p,k)|=k (5);
Where |·| represents the number of samples in the set, then K nearest neighbors R (x, K) to each other are defined as follows:
R(x,k)={xi|(xi∈N(x,k))∧(x∈N(xi,k))} (6)。
s200: and projecting the new sample into a new subspace by adopting a projection matrix U epsilon R d×c to obtain a predicted sample sequence, wherein the predicted sample sequence is arranged according to the Euclidean distance between the new sample and the samples in the training sample set from small to large. The predictive sample is ranked first, indicating that the probability of being the same person between the new sample and the predictive sample is highest.
Experiment and analysis:
feature selection: in order to quickly verify the validity of the proposed method, the LOMO features and GOG features commonly used in pedestrian re-recognition tasks are used herein.
Parameter setting: the algorithm is implemented by using theano framework. Wherein the minimum interval margin is taken to be 0.5, the balance coefficients α, λ are taken to be 0.005 and 0.0001, respectively, mapped to the subspace dimension c is taken to be 512, and the batch size, learning rate and k are taken to be 32, 1 and 10, respectively.
VIPeR database test results and analysis
VIPeR database is one of the most popular databases for pedestrian re-identification tasks. It contains 1264 images of 632 pedestrians acquired by two cameras with different illumination conditions with 90 ° angle of view. The training set is composed of 316 pedestrians, the test set is composed of the rest 316 pedestrians, and semi-supervision and full-supervision experimental settings are respectively carried out.
Semi-supervised experiments: for semi-supervised setting, we randomly take 1/3 of the pictures of pedestrians in the training set, wipe out the labels as unlabeled samples, and the remaining 2/3 of the pictures of pedestrians as labeled samples. The experimental results are shown in Table 4.1. By comparing the method with SSCDL and DLLAP, the performance of the method is greatly improved, and particularly the recognition rate of Rank-1 can reach 47.5% after LOMO features and GOG features are combined.
Table 4.1 VIPeR comparison of recognition rates of semi-supervised learning methods on database
Rank 1 5 10 20
SSCDL 25.6 53.7 68.2 83.6
DLLAP 32.5 61.8 74.3 84.1
LOMO+Our 34.2 65.2 76.4 85.4
GOG+Our 42.4 73.4 83.9 91.0
LOMO+GOG+Our 47.5 78.3 86.9 92.1
Full supervision experiment: we also set up the method presented here with full supervision, i.e. using labels of all training samples. The experimental results are shown in Table 4.2. Comparing DLLAP with L1Graph, it was found that there was a greater improvement in the methods presented herein when using GOG features and when using both LOMO and GOG features in combination. Comparing with semi-supervised setup, it can be seen that using LOMO and gos features, a recognition rate of 47.5% can be achieved with only 2/3 training sample labels, which is only 3% different than in the fully supervised case, fully demonstrating the effectiveness of the methods presented herein.
Table 4.2 VIPeR comparison of recognition rates under full supervision setting on database
Rank 1 5 10 20
DLLAP[41] 38.5 70.8 78.5 86.1
L1Graph[42] 41.5 - - -
LOMO+Our 36.1 68.2 79.6 88.5
GOG+Our 48.6 77.1 87.3 92.9
LOMO+GOG+Our 50.5 79.6 88.8 94.3
The method uses a contrast loss function to fully utilize the label information of the labeled sample, and uses a K mutual nearest neighbor method to replace a K nearest neighbor method to label the unlabeled sample. The experimental results on the pedestrian re-identification public dataset VIPeR confirm the effectiveness of the method.
Finally, it is noted that the above embodiments are only for illustrating the technical solution of the present invention and not for limiting the same, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications and equivalents may be made thereto without departing from the spirit and scope of the technical solution of the present invention, which is intended to be covered by the scope of the claims of the present invention.

Claims (3)

1. The pedestrian re-identification method based on semi-supervised learning is characterized by comprising the following steps of:
S100: learning a projection matrix U ε R d×c projects the original d-dimensional feature space into the c-dimensional subspace such that U TX∈Rc ×N satisfies in the new subspace: the Euclidean distance between pairs of samples from the same pedestrian is smaller, and the Euclidean distance between pairs of samples from different pedestrians is larger; samples from the same pedestrian are defined as homogeneous samples, and samples from different pedestrians are defined as heterogeneous samples;
The method for learning the projection matrix U epsilon R d×c in the S100 specifically comprises the following steps:
S110, building a training sample set, wherein the training sample set comprises a plurality of samples, the samples comprise labeled samples and unlabeled samples, and the labels of the samples of the same pedestrian in the labeled samples are the same;
Let x= [ X L,XU]∈Rd×N ] denote all training samples, where N is the number of all pictures contained in the training set, d is the length of the feature vector, Representing N L tagged samples,/>Representing N U unlabeled exemplars;
s120, establishing an objective function as follows:
Wherein L (U) is a regression function, W (U) is a regularized constraint, and alpha, lambda >0 is a balance coefficient;
S130, the labeled sample loss function is a contrast loss function: n P sample pairs for sampling And/>If/>And/>Samples from the same pedestrian, then in the new projection space/>And/>The Euclidean distance d n between should be as small as possible, close to 0; conversely, d n should be at least greater than a predetermined threshold margin >0, which would result in a loss if the above conditions are not met;
the labeled sample loss function of S130 is:
Wherein:
S140, labeling labels by using label-free sample labels, namely labeling labels by using a method of K nearest neighbors, wherein the loss function of label-free sample labels is as follows:
wherein if U Txi and U Txj meet K nearest neighbors to each other and x i and x j are from different cameras, then take
Otherwise W ij = 0; (8);
after labeling the label-free sample, further restricting the existing subspace by using the labeled sample, wherein the restricting weight is the cosine distance of the two samples in the new projection space;
s150: regularization term: the projection matrix U is constrained using L2,1 norm:
W(U)=||U||2,1 (4);
S200: and projecting the new sample into a new subspace by adopting a projection matrix U epsilon R d×c to obtain a predicted sample sequence, wherein the predicted sample sequence is arranged according to the Euclidean distance between the new sample and the samples in the training sample set from small to large.
2. The pedestrian re-recognition method based on semi-supervised learning of claim 1, wherein the N P sample sampling strategies sampled in S130 are sampling strategies that maximize top-k recognition rate, i.e., for each image, all samples with k nearest neighbors are sampled.
3. The pedestrian re-recognition method based on semi-supervised learning as set forth in claim 1, wherein the method for marking the label on the unlabeled exemplar by adopting the method of K nearest neighbors in S140 is as follows:
the K nearest neighbor N (x, K) defining sample x is as follows:
N(x,k)={x1,x2,...,xk},|N(p,k)|=k (5);
Where |·| represents the number of samples in the set, then K nearest neighbors R (x, K) to each other are defined as follows:
R(x,k)={xi|(xi∈N(x,k))∧(x∈N(xi,k))} (6)。
CN202010970306.5A 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning Active CN112101217B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010970306.5A CN112101217B (en) 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010970306.5A CN112101217B (en) 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning

Publications (2)

Publication Number Publication Date
CN112101217A CN112101217A (en) 2020-12-18
CN112101217B true CN112101217B (en) 2024-04-26

Family

ID=73758623

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010970306.5A Active CN112101217B (en) 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning

Country Status (1)

Country Link
CN (1) CN112101217B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113657176A (en) * 2021-07-22 2021-11-16 西南财经大学 Pedestrian re-identification implementation method based on active contrast learning
CN116052095B (en) * 2023-03-31 2023-06-16 松立控股集团股份有限公司 Vehicle re-identification method for smart city panoramic video monitoring

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8527432B1 (en) * 2008-08-08 2013-09-03 The Research Foundation Of State University Of New York Semi-supervised learning based on semiparametric regularization
CN107145827A (en) * 2017-04-01 2017-09-08 浙江大学 Across the video camera pedestrian recognition methods again learnt based on adaptive distance metric
CN109522956A (en) * 2018-11-16 2019-03-26 哈尔滨理工大学 A kind of low-rank differentiation proper subspace learning method
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN110008828A (en) * 2019-02-21 2019-07-12 上海工程技术大学 Pairs of constraint ingredient assay measures optimization method based on difference regularization
CN110175511A (en) * 2019-04-10 2019-08-27 杭州电子科技大学 It is a kind of to be embedded in positive negative sample and adjust the distance pedestrian's recognition methods again of distribution
CN111027442A (en) * 2019-12-03 2020-04-17 腾讯科技(深圳)有限公司 Model training method, recognition method, device and medium for pedestrian re-recognition
CN111027421A (en) * 2019-11-26 2020-04-17 西安宏规电子科技有限公司 Graph-based direct-push type semi-supervised pedestrian re-identification method
CN111033509A (en) * 2017-07-18 2020-04-17 视语智能有限公司 Object re-identification
CN111144451A (en) * 2019-12-10 2020-05-12 东软集团股份有限公司 Training method, device and equipment of image classification model
CN111353516A (en) * 2018-12-21 2020-06-30 华为技术有限公司 Sample classification method and model updating method for online learning
CN111563424A (en) * 2020-04-20 2020-08-21 清华大学 Pedestrian re-identification method and device based on semi-supervised learning

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3979007B2 (en) * 2000-12-22 2007-09-19 富士ゼロックス株式会社 Pattern identification method and apparatus
US9116894B2 (en) * 2013-03-14 2015-08-25 Xerox Corporation Method and system for tagging objects comprising tag recommendation based on query-based ranking and annotation relationships between objects and tags
US9471847B2 (en) * 2013-10-29 2016-10-18 Nec Corporation Efficient distance metric learning for fine-grained visual categorization
US11537817B2 (en) * 2018-10-18 2022-12-27 Deepnorth Inc. Semi-supervised person re-identification using multi-view clustering

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8527432B1 (en) * 2008-08-08 2013-09-03 The Research Foundation Of State University Of New York Semi-supervised learning based on semiparametric regularization
CN107145827A (en) * 2017-04-01 2017-09-08 浙江大学 Across the video camera pedestrian recognition methods again learnt based on adaptive distance metric
CN111033509A (en) * 2017-07-18 2020-04-17 视语智能有限公司 Object re-identification
CN109522956A (en) * 2018-11-16 2019-03-26 哈尔滨理工大学 A kind of low-rank differentiation proper subspace learning method
CN111353516A (en) * 2018-12-21 2020-06-30 华为技术有限公司 Sample classification method and model updating method for online learning
CN110008828A (en) * 2019-02-21 2019-07-12 上海工程技术大学 Pairs of constraint ingredient assay measures optimization method based on difference regularization
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN110175511A (en) * 2019-04-10 2019-08-27 杭州电子科技大学 It is a kind of to be embedded in positive negative sample and adjust the distance pedestrian's recognition methods again of distribution
CN111027421A (en) * 2019-11-26 2020-04-17 西安宏规电子科技有限公司 Graph-based direct-push type semi-supervised pedestrian re-identification method
CN111027442A (en) * 2019-12-03 2020-04-17 腾讯科技(深圳)有限公司 Model training method, recognition method, device and medium for pedestrian re-recognition
CN111144451A (en) * 2019-12-10 2020-05-12 东软集团股份有限公司 Training method, device and equipment of image classification model
CN111563424A (en) * 2020-04-20 2020-08-21 清华大学 Pedestrian re-identification method and device based on semi-supervised learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Center Based Pseudo-Labeling For Semi-Supervised Person Re-Identification;G. Ding等;2018 IEEE International Conference on Multimedia & Expo Workshops (ICMEW);20181129;第1-6页 *
行人重识别研究综述;张化祥等;山东师范大学学报(自然科学版);20181231;第33卷(第04期);第379-387页 *

Also Published As

Publication number Publication date
CN112101217A (en) 2020-12-18

Similar Documents

Publication Publication Date Title
Ren et al. Instance-aware, context-focused, and memory-efficient weakly supervised object detection
CN111814584B (en) Vehicle re-identification method based on multi-center measurement loss under multi-view environment
CN110414368B (en) Unsupervised pedestrian re-identification method based on knowledge distillation
Liu et al. Stepwise metric promotion for unsupervised video person re-identification
CN108960140B (en) Pedestrian re-identification method based on multi-region feature extraction and fusion
Dong et al. Deep metric learning-based for multi-target few-shot pavement distress classification
CN111950372B (en) Unsupervised pedestrian re-identification method based on graph convolution network
CN106682696B (en) The more example detection networks and its training method refined based on online example classification device
CN110263697A (en) Pedestrian based on unsupervised learning recognition methods, device and medium again
CN109961089A (en) Small sample and zero sample image classification method based on metric learning and meta learning
CN111598004B (en) Progressive reinforcement self-learning unsupervised cross-domain pedestrian re-identification method
US20210319215A1 (en) Method and system for person re-identification
US20140307958A1 (en) Instance-weighted mixture modeling to enhance training collections for image annotation
CN112101217B (en) Pedestrian re-identification method based on semi-supervised learning
CN109635708B (en) Unsupervised pedestrian re-identification method based on three-data-set cross migration learning
CN109299707A (en) A kind of unsupervised pedestrian recognition methods again based on fuzzy depth cluster
CN103150580A (en) Method and device for Hyperspectral image semi-supervised classification
CN113095442A (en) Hail identification method based on semi-supervised learning under multi-dimensional radar data
CN113158955B (en) Pedestrian re-recognition method based on clustering guidance and paired measurement triplet loss
Zheng et al. Adaptive boosting for domain adaptation: Toward robust predictions in scene segmentation
CN114329031B (en) Fine-granularity bird image retrieval method based on graph neural network and deep hash
CN114882534B (en) Pedestrian re-recognition method, system and medium based on anti-facts attention learning
CN110751005A (en) Pedestrian detection method integrating depth perception features and kernel extreme learning machine
CN114495004A (en) Unsupervised cross-modal pedestrian re-identification method
CN115294510A (en) Network training and recognition method and device, electronic equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant