CN111126360A - Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model - Google Patents

Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model Download PDF

Info

Publication number
CN111126360A
CN111126360A CN202010143811.2A CN202010143811A CN111126360A CN 111126360 A CN111126360 A CN 111126360A CN 202010143811 A CN202010143811 A CN 202010143811A CN 111126360 A CN111126360 A CN 111126360A
Authority
CN
China
Prior art keywords
pictures
loss function
picture
sample
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010143811.2A
Other languages
Chinese (zh)
Other versions
CN111126360B (en
Inventor
田玉敏
杨芸
吴自力
王笛
潘蓉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Publication of CN111126360A publication Critical patent/CN111126360A/en
Application granted granted Critical
Publication of CN111126360B publication Critical patent/CN111126360B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a cross-domain pedestrian re-identification method based on an unsupervised combined multi-loss model, and mainly solves the problem that the existing unsupervised cross-domain pedestrian re-identification method is low in identification rate. The scheme is that 1) a data set is obtained and divided into a training set and a testing set; 2) carrying out various preprocessing and expansion on the training set; 3) selecting a residual error network as a reference network model, initializing network parameters, and adjusting a network structure; 4) constructing a target domain loss function; 5) fusing a target domain loss function, a source domain loss function and a triple loss function to obtain a total loss function; 6) training the residual error network by using the total loss function to obtain a trained network model; 7) and inputting the test set into the trained network model, and outputting a recognition result. The invention improves the recognition rate of unsupervised cross-domain pedestrian re-recognition, effectively avoids the occurrence of over-fitting condition, and can be used for the intelligent security of suspect target searching and cross-camera tracking of pedestrians.

Description

Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model
Technical Field
The invention belongs to the field of computer vision and deep learning, and particularly relates to a cross-domain pedestrian re-identification method which can be used for the intelligent security field of suspect target searching and pedestrian cross-camera tracking.
Background
With the construction of digital cities, surveillance videos are more and more widely applied to various places where people live, the massive surveillance videos play an important role in public safety fields such as city public security and crime tracking, and how to better process the data is a great challenge in the future.
Pedestrian re-recognition, also known as pedestrian re-recognition, is a technique that uses computer vision techniques to determine whether a particular pedestrian is present in an image or video sequence, and is widely recognized as a sub-problem in image retrieval. Given a monitored pedestrian image, the pedestrian image is retrieved across the device.
In real life, under the influence of factors such as the visual angle of a camera, the posture of a pedestrian, shielding and illumination, images of the same person under different cameras are greatly different, and images of different pedestrians are likely to be very similar. This makes pedestrian re-identification a very challenging hot topic in the field of computer vision. The traditional pedestrian re-identification method is mainly researched from two aspects, namely, a pedestrian feature representation method is researched, and robust pedestrian features are designed to enable the pedestrian feature representation method to have invariance to factors such as visual angle change of a camera, pedestrian posture change, illumination change and background interference. And secondly, measurement learning, namely, the distance between different images of the same pedestrian is smaller and the distance between different pedestrians is larger by learning a distance measurement function.
In recent years, deep learning is rapidly developed, more and more researchers use deep learning technology to research pedestrian re-identification, and due to the good performance of deep learning in image high-level feature extraction and the appearance of large-scale data sets, the performance of the pedestrian re-identification method based on deep learning on each data set is better than that of the traditional pedestrian re-identification method. The existing pedestrian re-identification method based on deep learning comprises supervised learning and unsupervised learning, in the supervised learning, all data sets used for training need to be provided with labels, however, the labeling of data needs to consume a large amount of manpower and material resources, in real life, massive data are not provided with labels, and in the case of huge data, the manual labeling of the data is almost impossible. The pedestrian re-identification model learned through supervised learning has very poor performance on a label-free test set, the generalization capability of the model is very weak, and the expandability and the practicability are lacked in practical application. However, the pure utilization of unsupervised learning has poor performance of the learned network model due to no label of data, and cannot be applied in actual situations. By combining the two methods, how to obtain a better recognition result on a network model trained by a source domain with a label in a target domain without the label is a hotspot of pedestrian re-recognition research.
Hehe Fan, Liang Zheng et al, published in "Unsupervised person re-identification," Clustering and fining ". The thesis learns a deep network model as an initial pedestrian feature extractor using a labeled source domain data set, and then improves the network model on a target domain data set through unsupervised clustering. But this method cannot fully utilize the data information of the tagged source domain.
The paper Image-Image domain adaptation with predicted self-similarity and domain-similarity for person re-identification published by Wei jian Deng et al mainly studies how to reduce the difference of two cross-domain datasets at the pixel level. Although the method has good identification effect, the pedestrian re-identification error rate is high under the crossed cameras due to the fact that the intra-domain data characteristic distribution of the target domain is ignored.
Disclosure of Invention
The invention aims to provide a cross-domain pedestrian re-identification method based on an unsupervised joint multi-loss model, aiming at solving the problems in the prior art, so that the error rate of pedestrian re-identification on a label-free target domain data set is reduced, and the identification precision is improved.
In order to achieve the purpose, the technical scheme of the invention comprises the following steps:
1) acquiring a target domain data set without a label and a source domain data set with a label from an open website, randomly selecting a part of pictures in the target domain data set as a test set, dividing the test set into a query set and a candidate set, and taking the rest pictures in the target domain data set and the pictures in the source domain data set as training sets;
2) sequentially turning over, cutting and randomly erasing the pictures in the training set, preprocessing the data, and expanding the preprocessed pictures in the training set through cross-camera style migration; unifying the pictures of the training set to the same size;
3) selecting a residual error network as a reference network model, initializing network parameters, and adjusting the reference network structure;
4) constructing a target domain loss function:
4a) measuring the characteristics of each sample picture in the target domain data set by adopting cosine distance, sorting the distance from large to small, selecting the first k pictures as sample neighbors, and giving weight w to the neighbors of each picturei,j
Figure BDA0002400010980000031
Wherein i is the sample number, j is the sample neighbor number, k is the neighbor number, xt,iIs a target domain sampleThis picture, M (x)t,iK) is a set of numbers of k neighbors of the sample picture;
4b) according to the weight wi,jConstructing a target domain loss function Ltgt
Figure BDA0002400010980000032
Wherein the content of the first and second substances,
Figure BDA0002400010980000033
is a randomly sampled picture in the training set, ntIs the number of pictures of the batch training target domain,
Figure BDA0002400010980000034
is a picture
Figure BDA0002400010980000035
Probability of belonging to neighbor j;
5) fusing the target domain loss function with the existing source domain loss function and the triple loss function to obtain a total loss function L:
L=λ1Lsrc2Ltgt3LT
wherein L issrcAs a source domain loss function, LTIs a triplet loss function, λ1Is a weighted value, λ, of the proportion of the source domain loss function2Is the proportional weighted value of the loss function of the target domain, lambda3The weight value of the proportion of the triple loss function is calculated;
6) inputting the training set picture into a residual error network, training the residual error network by using a total loss function L and optimizing a training process to obtain a trained residual error network model;
7) inputting the pictures in the query set and the candidate set into a trained residual error network model for feature extraction, respectively calculating Euclidean distances between the features of the target pictures in the query set and the features of the pictures in the candidate set, sequencing the calculated Euclidean distances, and selecting the picture closest to the target in the query set in the candidate set as a recognition result.
The invention has the following advantages:
firstly, the invention effectively avoids the occurrence of the over-fitting condition and improves the generalization capability of the model because the data is subjected to various pretreatments of overturning, cutting and random erasing;
secondly, the data set is subjected to cross-camera style migration, so that the error rate of pedestrian re-identification under the view angle of the crossed cameras is reduced while data is expanded;
thirdly, the method greatly excavates the characteristic distribution characteristics of the label-free target domain data set due to the construction of the target domain loss function, fuses the source domain loss function, the target domain loss function and the triple loss function, and obviously improves the recognition rate of the unsupervised cross-domain pedestrian re-recognition.
Drawings
FIG. 1 is a flow chart of an implementation of the present invention;
FIG. 2 is a diagram of a residual network architecture in accordance with the present invention;
fig. 3 is a schematic diagram of a residual error network training process in the present invention.
Detailed Description
The invention is described in detail below with reference to the figures and specific embodiments. It is to be understood that the embodiments described are only a few embodiments of the present invention, and not all embodiments. All other embodiments that can be obtained by a person skilled in the art based on the embodiments of the present invention without any inventive step shall fall within the scope of the present invention.
Referring to fig. 1, the specific implementation of this example is as follows:
the method comprises the steps of firstly, acquiring a target domain data set and a source domain data set, and organizing the acquired data sets into a training set and a testing set.
Downloading a Market-1501 data set and a DukeMTMC-reiD data set from a public website, taking the Market-1501 data set as a target domain data set and taking the DukeMTMC-reiD data set as a source domain data set;
using DukeMTMC-reiD data set as the first training set R1
The Market-1501 data set is divided into 6: 4 into a second training set R2And the test set T is further processed by the following steps of: 2 into a candidate set G and a query set Q;
the first training set R1And a second training set R2Forming a total training set R.
Step two, preprocessing the pictures in the training set R and expanding the second training set R2And unifying the picture size.
2.1) carrying out the following preprocessing on the training set R:
randomly extracting a plurality of pictures on the training set R for turning, including horizontal turning and vertical turning;
randomly extracting a plurality of pictures on the training set R for cutting, wherein the pictures are 2/3 of the length and the width of the original picture;
randomly extracting a plurality of pictures on the training set R to erase random areas in random sizes;
2.2) using the cyclic generation to generate a countermeasure network to carry out cross-camera style migration, and using a second training set R2Pictures shot by different cameras are input into the cyclic generation countermeasure network to obtain pictures with converted camera visual angles, the converted pictures are stored in a folder, and the pictures in the folder and the pictures before conversion are added into a second training set R2To the second training set R2The expansion is performed and all pictures are unified to 256 x 128 in size.
Referring to fig. 2, the structure of the residual error network selected in this example is described in further detail:
and step three, selecting a reference network model, initializing network parameters and adjusting the reference network structure.
3.1) selecting a residual error network ResNet-50 as a reference network model, and initializing the residual error network ResNet-50 by using model parameters pre-trained on ImageNet by utilizing transfer learning;
the residual error network ResNet-50 comprises five convolution layers, five pooling layers and a full connection layer, and the structural relationship is as follows in sequence: the first convolution layer, the first normalization processing layer, the first pooling layer, the second convolution layer, the second normalization processing layer, the second pooling layer, the third convolution layer, the third normalization processing layer, the third pooling layer, the fourth convolution layer, the fourth normalization processing layer, the fourth pooling layer, the fifth convolution layer, the fifth normalization processing layer, the fifth pooling layer and the 1000-dimensional full-connection layer;
3.2) adjusting the structure of ResNet-50:
3a) all layers before the fifth pooling layer of ResNet-50 are reserved, the last 1000-dimensional full-link layer is removed, and a 4096-dimensional output full-link layer is added;
3b) sequentially adding a normalization processing layer, a linear rectification function layer, a random discarding layer and two components after the 4096-dimensional output full connection layer, wherein the first component is a classification module of source domain data and comprises an M-dimensional full connection layer and an activation function layer; the second component is a sample memory module of the target domain, which is a key value pair structure with a key slot for storing picture characteristics and a value slot for storing the input number of the picture, and the key has uniqueness.
Referring to fig. 3, the training process of the residual error network in this example is described in further detail:
and step four, constructing a target domain loss function.
Based on the following three considerations, the objective domain loss function needs to be constructed:
regarding each sample of the target domain as an independent class, wherein each sample belongs to the class of the sample and is far away from other samples as far as possible;
the image after the cross-camera style migration and the image before the migration represent the same identity of the pedestrian, and the distance between the two images needs to be shortened;
for each sample picture, a sample with the same identity as the pedestrian exists in the neighbor of the sample picture, and the distance between the sample picture and the neighbor with the same identity is required to be shortened;
the objective loss function is constructed as follows:
4a) measuring the characteristics of each sample picture in the target domain data set by adopting cosine distanceThe cosine distance is the product of multiplying two eigenvectors and dividing the product by the modulus of the two eigenvectors, the larger the cosine distance is, the smaller the included angle between the two eigenvectors is, the closer the two eigenvectors are, otherwise, the farther the two eigenvectors are; according to the sorting of the distances from big to small, the first k pictures are selected as sample neighbors, and the neighbors of each picture are given a weight wi,j
Figure BDA0002400010980000061
Wherein i is the sample number, j is the sample neighbor number, k is the neighbor number, xt,iIs a target domain sample picture, M (x)t,iK) is a sample picture xt,iThe number of k neighbors, k being 6 in this example;
4b) according to the weight wi,jTo obtain a target domain loss function Ltgt
Figure BDA0002400010980000062
Wherein the content of the first and second substances,
Figure BDA0002400010980000063
is a randomly sampled picture in the training set, ntIs the number of pictures of the batch training target domain,
Figure BDA0002400010980000064
is a picture
Figure BDA0002400010980000065
Probability of belonging to neighbor j;
and step five, fusing the target domain loss function with the existing source domain loss function and the triple loss function.
5.1) Source Domain loss function:
the source domain loss function is a cross entropy loss function, and the formula is as follows:
Figure BDA0002400010980000066
in the formula, i is the sample number, m is the number of source domain pictures in the training batch, and xs,iFor source domain sample pictures, ys,iIs xs,iLabel information carried by the specimen, p (y)s,i|xs,i) Is xs,iBelong to ys,iThe probability of (d);
5.2) triple loss function:
because the source domain data set and the target domain data set do not have the same pedestrian identity, a picture before the target domain data set is subjected to style migration is randomly selected as an anchor sample, the picture after the picture is subjected to style migration is a positive sample, the picture of the source domain data set is taken as a negative sample, and the anchor sample, the positive sample and the negative sample form a three-tuple object.
Constructing a triple loss function by using the triple objects, wherein the formula is as follows:
Figure BDA0002400010980000071
wherein D is an edge parameter representing the minimum distance between the anchor point and the positive sample and the distance between the anchor point and the negative sample, D (.) is the Euclidean distance, and x istFor the picture before the target domain data set format migration,
Figure BDA0002400010980000072
for pictures after style migration, xsThe source domain picture is taken, X is the size of the batch training picture set, and X is taken as 128 in this example;
5.3) fusing the target domain loss function, the source domain loss function and the triple loss function to obtain a total loss function L:
L=λ1Lsrc2Ltgt3LT
wherein L issrcAs a source domain loss function, LTIs a triplet loss function, λ1Is a weighted value, λ, of the proportion of the source domain loss function2Is the proportional weighted value of the loss function of the target domain, lambda3For the proportion weighted value of the triple loss function, lambda is taken in the example1=0.5,λ2=0.3,λ3=0.2。
And step six, inputting the training set R into the residual error network ResNet-50, and training the residual error network ResNet-50 to obtain a trained residual error network model.
The training process is as follows:
6a) all the features stored in the sample memory are initialized to 0, and the sample memory is updated by the following formula:
K[i]←αK[i]+(1-α)f(xt,i)
wherein α is a hyperparameter, α epsilon [0,1 ∈ ]]Controlling the update frequency, K [ i ]]For sample picture xt,iFeatures stored in the sample memory, f (x)t,i) Is xt,iThe feature vector of the sample, which is α in this example, is 0.01 as the initial value;
6b) setting the total iteration times of training as 60 times, the number of batch pictures as 128, selecting the optimizer as a random gradient optimizer, setting the learning rate as 0.1 in the first 40 iterations, and dividing the learning rate by 10 in each iteration in the last 20 iterations;
6c) dividing a training set R into a plurality of small-batch data sets by taking 128 as a base number by adopting a small-batch iteration method, sequentially sending each small-batch data set into a residual error network ResNet-50 to extract a characteristic vector, inputting the obtained characteristic vector into a total loss function L to calculate loss, and calculating the gradient of the loss obtained by calculation;
6d) inputting the obtained gradient into a random gradient optimizer to update the weight and the threshold of the network;
6e) and repeating the steps 6c) to 6d), and stopping training when the total iteration times reach 60 times to obtain a trained model.
And step seven, testing the trained model and outputting a recognition result.
7.1) inputting the query set Q and the candidate set G into a trained residual error network model for feature extraction;
7.2) respectively calculating the Euclidean distance of the characteristics of the target picture in the query set Q and the characteristics of each picture in the candidate set G, wherein the formula is as follows:
Figure BDA0002400010980000081
wherein x is the target picture, yiN is the number of pictures in the candidate set G for inquiring the pictures;
7.3) ordering the calculated Euclidean distances from small to large, and selecting the picture closest to the target picture in the query set Q from the candidate set G as an identification result.
The effects of the present invention can be further explained by the evaluation results of the following experiments.
1. Experimental methods and evaluation criteria
The method used in the experiment was the present invention and the existing HHL method, an unsupervised cross-domain pedestrian re-identification method, HHL for short, proposed in the paper "general aperson regenerative model-and homogeneous", published by Zhuin Zhong, et al 2018 at the ECCV (European Conference on Computer Vision) Conference.
Evaluation criteria, using three criteria mAP, rank-1, rank-5 in the area of pedestrian re-identification, wherein:
mAP is average precision mean value, which is to sum and average the average precision in the multi-classification task;
rank-1 is the probability that the result of the 1 st graph in the search results is correct;
rank-5 is the probability that the results of the first 5 pictures in the search results are correct;
2. contents and results of the experiments
The results of the tests on the Market-1501 data set using the method of the present invention and the prior HHL method are shown in table 1.
TABLE 1
Method of producing a composite material mAP rank-1 rank-5
The invention 43.1 75.2 87.9
HHL 31.4 66.4 78.8
As can be seen from Table 1, compared with the prior art, the average accuracy mean mAP of the method is improved by 11.7, the probability rank-1 that the result of the 1 st graph in the search result is correct is improved by 8.8, and the probability rank-5 that the result of the first 5 graphs in the search result is correct is improved by 9.1.

Claims (8)

1. A cross-domain pedestrian re-identification method based on an unsupervised combined multi-loss model is characterized by comprising the following steps:
1) acquiring a target domain data set without a label and a source domain data set with a label from an open website, randomly selecting a part of pictures in the target domain data set as a test set, dividing the test set into a query set and a candidate set, and taking the rest pictures in the target domain data set and the pictures in the source domain data set as training sets;
2) sequentially turning over, cutting and randomly erasing the pictures in the training set, preprocessing the data, and expanding the preprocessed pictures in the training set through cross-camera style migration; unifying the pictures of the training set to the same size;
3) selecting a residual error network as a reference network model, initializing network parameters, and adjusting the reference network structure;
4) constructing a target domain loss function:
4a) measuring the characteristics of each sample picture in the target domain data set by adopting cosine distance, sorting the distance from large to small, selecting the first k pictures as sample neighbors, and giving weight w to the neighbors of each picturei,j
Figure FDA0002400010970000011
Wherein i is the sample number, j is the sample neighbor number, k is the neighbor number, xt,iIs a target domain sample picture, M (x)t,iK) is a set of numbers of k neighbors of the sample picture;
4b) according to the weight wi,jConstructing a target domain loss function Ltgt
Figure FDA0002400010970000012
Wherein the content of the first and second substances,
Figure FDA0002400010970000013
is a randomly sampled picture in the training set, ntIs the number of pictures of the batch training target domain,
Figure FDA0002400010970000014
is a picture
Figure FDA0002400010970000015
Probability of belonging to neighbor j;
5) fusing the target domain loss function with the existing source domain loss function and the triple loss function to obtain a total loss function L:
L=λ1Lsrc2Ltgt3LT
wherein L issrcAs a source domain loss function, LTIs a triplet loss function, λ1Is a weighted value, λ, of the proportion of the source domain loss function2Occupied for target domain loss functionProportional weight value, λ3The weight value of the proportion of the triple loss function is calculated;
6) inputting the training set picture into a residual error network, training the residual error network by using a total loss function L and optimizing a training process to obtain a trained residual error network model;
7) inputting the pictures in the query set and the candidate set into a trained residual error network model for feature extraction, respectively calculating Euclidean distances between the features of the target pictures in the query set and the features of the pictures in the candidate set, sequencing the calculated Euclidean distances, and selecting the picture closest to the target in the query set in the candidate set as a recognition result.
2. The method of claim 1, wherein 2) the cyclic generation countermeasure network is utilized to perform the cross-camera style migration, the pictures taken by different cameras in the target domain data set are input into the cyclic generation countermeasure network, the pictures with converted camera view angles are obtained, and the converted pictures and the pictures before conversion are added into the target domain data set together to expand the data set.
3. The method of claim 1, wherein in 3), migration learning is used, model parameters pre-trained on ImageNet are used for initializing the residual error network, and the structure of the residual error network is adjusted, so that the following is achieved:
3a) removing the last full connection layer of the residual error network and adding a 4096-dimensional output full connection layer;
3b) sequentially adding a normalization processing layer, a linear rectification function layer, a random discarding layer and two components after the 4096-dimensional output full connection layer, wherein the first component is a classification module of source domain data and comprises an M-dimensional full connection layer and an activation function layer; the second component is a sample memory module of the target domain, which is a key value pair structure with a key slot for storing picture characteristics and a value slot for storing the input number of the picture, the key having uniqueness.
4. The method of claim 1, wherein the cosine distance in 4) is a product of multiplying two eigenvectors and dividing the product by a modulus of the two eigenvectors, and wherein the larger the cosine distance, the smaller the included angle between the two eigenvectors, the closer the two eigenvectors are, and vice versa, the farther the distance is.
5. The method of claim 1, wherein the source domain loss function in 5) is a cross-entropy loss function, and the formula is as follows:
Figure FDA0002400010970000031
in the formula, i is the sample number, m is the number of source domain pictures in the training batch, and xs,iFor source domain sample pictures, ys,iIs xs,iLabel information carried by the specimen, p (y)s,i|xs,i) Is xs,iBelong to ys,iThe probability of (c).
6. The method according to claim 1, wherein 5) a picture before the target domain data set style migration is randomly selected as an anchor sample, the picture after the picture style migration is a positive sample, the picture of the source domain data set is a negative sample, and a triplet object is formed by the anchor sample, the positive sample and the negative sample;
constructing a triple loss function by using the triple objects, wherein the formula is as follows:
Figure FDA0002400010970000032
wherein D is an edge parameter representing the minimum distance between the anchor point and the positive sample and the distance between the anchor point and the negative sample, D (.) is the Euclidean distance, and x istFor the picture before the target domain data set format migration,
Figure FDA0002400010970000033
for pictures after style migration, xsIs the source domain picture, X is the batch training picture set size.
7. The method of claim 1, wherein 6) the residual network is trained using the total loss function L, and the method is implemented as follows:
6a) all the features stored in the sample memory are initialized to 0, and the sample memory is updated by the following formula:
K[i]←αK[i]+(1-α)f(xt,i)
wherein α is a hyperparameter, α epsilon [0,1 ∈ ]]Controlling the update frequency, K [ i ]]For sample picture xt,iFeatures stored in the sample memory, f (x)t,i) Is xt,iA feature vector of the sample;
6b) setting the total iteration times, the learning rate and the number of batch pictures of training, and selecting an optimizer;
6c) dividing a data set into a plurality of small-batch data sets by using a small-batch iteration method based on the number of batch pictures, sequentially sending each small-batch data set into a residual error network to extract a characteristic vector, and inputting the obtained characteristic vector into a total loss function L to calculate loss;
6d) solving the gradient of the calculated loss, and inputting the obtained gradient into the selected optimizer to update the network parameters;
6e) and repeating the steps 6c) to 6d), and stopping training when the total iteration times are reached to obtain a trained model.
8. The method according to claim 1, wherein 7) the euclidean distances are calculated for the features of the target pictures in the query set and the picture features in the candidate set respectively by the following formula:
Figure FDA0002400010970000041
wherein Q is a query set, G is a candidate set, x is a target picture, and yiFor querying pictures, n is the number of pictures.
CN202010143811.2A 2019-11-15 2020-03-04 Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model Active CN111126360B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2019111175727 2019-11-15
CN201911117572 2019-11-15

Publications (2)

Publication Number Publication Date
CN111126360A true CN111126360A (en) 2020-05-08
CN111126360B CN111126360B (en) 2023-03-24

Family

ID=70493521

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010143811.2A Active CN111126360B (en) 2019-11-15 2020-03-04 Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model

Country Status (1)

Country Link
CN (1) CN111126360B (en)

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111598004A (en) * 2020-05-18 2020-08-28 北京星闪世图科技有限公司 Progressive-enhancement self-learning unsupervised cross-domain pedestrian re-identification method
CN111783576A (en) * 2020-06-18 2020-10-16 西安电子科技大学 Pedestrian re-identification method based on improved YOLOv3 network and feature fusion
CN111814655A (en) * 2020-07-03 2020-10-23 浙江大华技术股份有限公司 Target re-identification method, network training method thereof and related device
CN111814854A (en) * 2020-06-28 2020-10-23 北京交通大学 Target re-identification method adaptive to unsupervised domain
CN111881714A (en) * 2020-05-22 2020-11-03 北京交通大学 Unsupervised cross-domain pedestrian re-identification method
CN111931637A (en) * 2020-08-07 2020-11-13 华南理工大学 Cross-modal pedestrian re-identification method and system based on double-current convolutional neural network
CN111950411A (en) * 2020-07-31 2020-11-17 上海商汤智能科技有限公司 Model determination method and related device
CN112069929A (en) * 2020-08-20 2020-12-11 之江实验室 Unsupervised pedestrian re-identification method and device, electronic equipment and storage medium
CN112084895A (en) * 2020-08-25 2020-12-15 南京邮电大学 Pedestrian re-identification method based on deep learning
CN112215280A (en) * 2020-10-12 2021-01-12 西安交通大学 Small sample image classification method based on meta-backbone network
CN112287089A (en) * 2020-11-23 2021-01-29 腾讯科技(深圳)有限公司 Classification model training and automatic question-answering method and device for automatic question-answering system
CN112347995A (en) * 2020-11-30 2021-02-09 中国科学院自动化研究所 Unsupervised pedestrian re-identification method based on fusion of pixel and feature transfer
CN112381056A (en) * 2020-12-02 2021-02-19 山西大学 Cross-domain pedestrian re-identification method and system fusing multiple source domains
CN112434599A (en) * 2020-11-23 2021-03-02 同济大学 Pedestrian re-identification method based on random shielding recovery of noise channel
CN112560925A (en) * 2020-12-10 2021-03-26 中国科学院深圳先进技术研究院 Complex scene target detection data set construction method and system
CN112633071A (en) * 2020-11-30 2021-04-09 之江实验室 Pedestrian re-identification data domain adaptation method based on data style decoupling content migration
CN112669827A (en) * 2020-12-28 2021-04-16 清华大学 Joint optimization method and system for automatic speech recognizer
CN112733695A (en) * 2021-01-04 2021-04-30 电子科技大学 Unsupervised key frame selection method in pedestrian re-identification field
CN112818175A (en) * 2021-02-07 2021-05-18 中国矿业大学 Factory worker searching method and training method of worker recognition model
CN112861705A (en) * 2021-02-04 2021-05-28 东北林业大学 Cross-domain pedestrian re-identification method based on hybrid learning
CN113065516A (en) * 2021-04-22 2021-07-02 中国矿业大学 Unsupervised pedestrian re-identification system and method based on sample separation
CN113095263A (en) * 2021-04-21 2021-07-09 中国矿业大学 Method and device for training heavy identification model of pedestrian under shielding and method and device for heavy identification of pedestrian under shielding
CN113095229A (en) * 2021-04-14 2021-07-09 中国矿业大学 Unsupervised domain self-adaptive pedestrian re-identification system and method
CN113128466A (en) * 2021-05-11 2021-07-16 深圳大学 Pedestrian re-identification method, system, electronic device and storage medium
CN113139536A (en) * 2021-05-12 2021-07-20 哈尔滨工业大学(威海) Text verification code identification method and equipment based on cross-domain meta learning and storage medium
CN113221656A (en) * 2021-04-13 2021-08-06 电子科技大学 Cross-domain pedestrian re-identification model based on domain invariant features and method thereof
CN113255823A (en) * 2021-06-15 2021-08-13 中国人民解放军国防科技大学 Unsupervised domain adaptation method and unsupervised domain adaptation device
CN113378620A (en) * 2021-03-31 2021-09-10 中交第二公路勘察设计研究院有限公司 Cross-camera pedestrian re-identification method in surveillance video noise environment
CN113537403A (en) * 2021-08-14 2021-10-22 北京达佳互联信息技术有限公司 Training method and device and prediction method and device of image processing model
CN113723345A (en) * 2021-09-09 2021-11-30 河北工业大学 Domain-adaptive pedestrian re-identification method based on style conversion and joint learning network
CN113837256A (en) * 2021-09-15 2021-12-24 深圳市商汤科技有限公司 Object recognition method, network training method and device, equipment and medium
CN113936301A (en) * 2021-07-02 2022-01-14 西北工业大学 Target re-identification method based on central point prediction loss function
CN114220124A (en) * 2021-12-16 2022-03-22 华南农业大学 Near-infrared-visible light cross-modal double-flow pedestrian re-identification method and system
CN115100591A (en) * 2022-06-17 2022-09-23 哈尔滨工业大学 Multi-target tracking and target re-identification system and method based on joint learning
CN115147774A (en) * 2022-07-05 2022-10-04 中国科学技术大学 Pedestrian re-identification method in degradation environment based on feature alignment
CN115345259A (en) * 2022-10-14 2022-11-15 北京睿企信息科技有限公司 Optimization method, equipment and storage medium for training named entity recognition model
EP4136580A4 (en) * 2020-06-29 2024-01-17 Zhejiang Dahua Technology Co Target re-identification method, network training method thereof, and related device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109948561A (en) * 2019-03-25 2019-06-28 广东石油化工学院 The method and system that unsupervised image/video pedestrian based on migration network identifies again
WO2019128367A1 (en) * 2017-12-26 2019-07-04 广州广电运通金融电子股份有限公司 Face verification method and apparatus based on triplet loss, and computer device and storage medium
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN110321813A (en) * 2019-06-18 2019-10-11 南京信息工程大学 Cross-domain pedestrian recognition methods again based on pedestrian's segmentation
CN110414462A (en) * 2019-08-02 2019-11-05 中科人工智能创新技术研究院(青岛)有限公司 A kind of unsupervised cross-domain pedestrian recognition methods and system again

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019128367A1 (en) * 2017-12-26 2019-07-04 广州广电运通金融电子股份有限公司 Face verification method and apparatus based on triplet loss, and computer device and storage medium
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN109948561A (en) * 2019-03-25 2019-06-28 广东石油化工学院 The method and system that unsupervised image/video pedestrian based on migration network identifies again
CN110321813A (en) * 2019-06-18 2019-10-11 南京信息工程大学 Cross-domain pedestrian recognition methods again based on pedestrian's segmentation
CN110414462A (en) * 2019-08-02 2019-11-05 中科人工智能创新技术研究院(青岛)有限公司 A kind of unsupervised cross-domain pedestrian recognition methods and system again

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王兴柱等: "三元组在非监督行人再识别中的应用", 《仪表技术》 *

Cited By (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111598004A (en) * 2020-05-18 2020-08-28 北京星闪世图科技有限公司 Progressive-enhancement self-learning unsupervised cross-domain pedestrian re-identification method
CN111598004B (en) * 2020-05-18 2023-12-08 江苏星闪世图科技(集团)有限公司 Progressive reinforcement self-learning unsupervised cross-domain pedestrian re-identification method
CN111881714A (en) * 2020-05-22 2020-11-03 北京交通大学 Unsupervised cross-domain pedestrian re-identification method
CN111881714B (en) * 2020-05-22 2023-11-21 北京交通大学 Unsupervised cross-domain pedestrian re-identification method
CN111783576A (en) * 2020-06-18 2020-10-16 西安电子科技大学 Pedestrian re-identification method based on improved YOLOv3 network and feature fusion
CN111783576B (en) * 2020-06-18 2023-08-18 西安电子科技大学 Pedestrian re-identification method based on improved YOLOv3 network and feature fusion
CN111814854B (en) * 2020-06-28 2023-07-28 北京交通大学 Target re-identification method without supervision domain adaptation
CN111814854A (en) * 2020-06-28 2020-10-23 北京交通大学 Target re-identification method adaptive to unsupervised domain
WO2022001489A1 (en) * 2020-06-28 2022-01-06 北京交通大学 Unsupervised domain adaptation target re-identification method
EP4136580A4 (en) * 2020-06-29 2024-01-17 Zhejiang Dahua Technology Co Target re-identification method, network training method thereof, and related device
CN111814655B (en) * 2020-07-03 2023-09-01 浙江大华技术股份有限公司 Target re-identification method, network training method thereof and related device
CN111814655A (en) * 2020-07-03 2020-10-23 浙江大华技术股份有限公司 Target re-identification method, network training method thereof and related device
CN111950411A (en) * 2020-07-31 2020-11-17 上海商汤智能科技有限公司 Model determination method and related device
CN111931637B (en) * 2020-08-07 2023-09-15 华南理工大学 Cross-modal pedestrian re-identification method and system based on double-flow convolutional neural network
CN111931637A (en) * 2020-08-07 2020-11-13 华南理工大学 Cross-modal pedestrian re-identification method and system based on double-current convolutional neural network
CN112069929A (en) * 2020-08-20 2020-12-11 之江实验室 Unsupervised pedestrian re-identification method and device, electronic equipment and storage medium
CN112069929B (en) * 2020-08-20 2024-01-05 之江实验室 Unsupervised pedestrian re-identification method and device, electronic equipment and storage medium
CN112084895A (en) * 2020-08-25 2020-12-15 南京邮电大学 Pedestrian re-identification method based on deep learning
CN112084895B (en) * 2020-08-25 2022-07-29 南京邮电大学 Pedestrian re-identification method based on deep learning
CN112215280A (en) * 2020-10-12 2021-01-12 西安交通大学 Small sample image classification method based on meta-backbone network
CN112287089A (en) * 2020-11-23 2021-01-29 腾讯科技(深圳)有限公司 Classification model training and automatic question-answering method and device for automatic question-answering system
CN112434599A (en) * 2020-11-23 2021-03-02 同济大学 Pedestrian re-identification method based on random shielding recovery of noise channel
CN112287089B (en) * 2020-11-23 2022-09-20 腾讯科技(深圳)有限公司 Classification model training and automatic question-answering method and device for automatic question-answering system
CN112347995A (en) * 2020-11-30 2021-02-09 中国科学院自动化研究所 Unsupervised pedestrian re-identification method based on fusion of pixel and feature transfer
CN112633071A (en) * 2020-11-30 2021-04-09 之江实验室 Pedestrian re-identification data domain adaptation method based on data style decoupling content migration
CN112347995B (en) * 2020-11-30 2022-09-23 中国科学院自动化研究所 Unsupervised pedestrian re-identification method based on fusion of pixel and feature transfer
CN112381056A (en) * 2020-12-02 2021-02-19 山西大学 Cross-domain pedestrian re-identification method and system fusing multiple source domains
CN112381056B (en) * 2020-12-02 2022-04-01 山西大学 Cross-domain pedestrian re-identification method and system fusing multiple source domains
CN112560925A (en) * 2020-12-10 2021-03-26 中国科学院深圳先进技术研究院 Complex scene target detection data set construction method and system
CN112669827B (en) * 2020-12-28 2022-08-02 清华大学 Joint optimization method and system for automatic speech recognizer
CN112669827A (en) * 2020-12-28 2021-04-16 清华大学 Joint optimization method and system for automatic speech recognizer
CN112733695A (en) * 2021-01-04 2021-04-30 电子科技大学 Unsupervised key frame selection method in pedestrian re-identification field
CN112861705A (en) * 2021-02-04 2021-05-28 东北林业大学 Cross-domain pedestrian re-identification method based on hybrid learning
CN112861705B (en) * 2021-02-04 2022-07-05 东北林业大学 Cross-domain pedestrian re-identification method based on hybrid learning
CN112818175A (en) * 2021-02-07 2021-05-18 中国矿业大学 Factory worker searching method and training method of worker recognition model
CN112818175B (en) * 2021-02-07 2023-09-01 中国矿业大学 Factory staff searching method and training method of staff identification model
CN113378620B (en) * 2021-03-31 2023-04-07 中交第二公路勘察设计研究院有限公司 Cross-camera pedestrian re-identification method in surveillance video noise environment
CN113378620A (en) * 2021-03-31 2021-09-10 中交第二公路勘察设计研究院有限公司 Cross-camera pedestrian re-identification method in surveillance video noise environment
CN113221656A (en) * 2021-04-13 2021-08-06 电子科技大学 Cross-domain pedestrian re-identification model based on domain invariant features and method thereof
CN113095229A (en) * 2021-04-14 2021-07-09 中国矿业大学 Unsupervised domain self-adaptive pedestrian re-identification system and method
CN113095229B (en) * 2021-04-14 2024-04-12 中国矿业大学 Self-adaptive pedestrian re-identification system and method for unsupervised domain
CN113095263B (en) * 2021-04-21 2024-02-20 中国矿业大学 Training method and device for pedestrian re-recognition model under shielding and pedestrian re-recognition method and device under shielding
CN113095263A (en) * 2021-04-21 2021-07-09 中国矿业大学 Method and device for training heavy identification model of pedestrian under shielding and method and device for heavy identification of pedestrian under shielding
CN113065516B (en) * 2021-04-22 2023-12-01 中国矿业大学 Sample separation-based unsupervised pedestrian re-identification system and method
CN113065516A (en) * 2021-04-22 2021-07-02 中国矿业大学 Unsupervised pedestrian re-identification system and method based on sample separation
CN113128466B (en) * 2021-05-11 2023-12-05 深圳大学 Pedestrian re-recognition method, system, electronic device and storage medium
CN113128466A (en) * 2021-05-11 2021-07-16 深圳大学 Pedestrian re-identification method, system, electronic device and storage medium
CN113139536A (en) * 2021-05-12 2021-07-20 哈尔滨工业大学(威海) Text verification code identification method and equipment based on cross-domain meta learning and storage medium
CN113255823A (en) * 2021-06-15 2021-08-13 中国人民解放军国防科技大学 Unsupervised domain adaptation method and unsupervised domain adaptation device
CN113936301B (en) * 2021-07-02 2024-03-12 西北工业大学 Target re-identification method based on center point prediction loss function
CN113936301A (en) * 2021-07-02 2022-01-14 西北工业大学 Target re-identification method based on central point prediction loss function
CN113537403A (en) * 2021-08-14 2021-10-22 北京达佳互联信息技术有限公司 Training method and device and prediction method and device of image processing model
CN113723345B (en) * 2021-09-09 2023-11-14 河北工业大学 Domain self-adaptive pedestrian re-identification method based on style conversion and joint learning network
CN113723345A (en) * 2021-09-09 2021-11-30 河北工业大学 Domain-adaptive pedestrian re-identification method based on style conversion and joint learning network
CN113837256A (en) * 2021-09-15 2021-12-24 深圳市商汤科技有限公司 Object recognition method, network training method and device, equipment and medium
CN114220124A (en) * 2021-12-16 2022-03-22 华南农业大学 Near-infrared-visible light cross-modal double-flow pedestrian re-identification method and system
CN115100591A (en) * 2022-06-17 2022-09-23 哈尔滨工业大学 Multi-target tracking and target re-identification system and method based on joint learning
CN115147774A (en) * 2022-07-05 2022-10-04 中国科学技术大学 Pedestrian re-identification method in degradation environment based on feature alignment
CN115147774B (en) * 2022-07-05 2024-04-02 中国科学技术大学 Pedestrian re-identification method based on characteristic alignment in degradation environment
CN115345259A (en) * 2022-10-14 2022-11-15 北京睿企信息科技有限公司 Optimization method, equipment and storage medium for training named entity recognition model

Also Published As

Publication number Publication date
CN111126360B (en) 2023-03-24

Similar Documents

Publication Publication Date Title
CN111126360B (en) Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model
Wu et al. Deep learning-based methods for person re-identification: A comprehensive review
CN110414368B (en) Unsupervised pedestrian re-identification method based on knowledge distillation
CN111259786B (en) Pedestrian re-identification method based on synchronous enhancement of appearance and motion information of video
CN109961051B (en) Pedestrian re-identification method based on clustering and block feature extraction
Lin et al. RSCM: Region selection and concurrency model for multi-class weather recognition
Tao et al. Person re-identification by regularized smoothing kiss metric learning
CN111709311B (en) Pedestrian re-identification method based on multi-scale convolution feature fusion
CN111783576B (en) Pedestrian re-identification method based on improved YOLOv3 network and feature fusion
CN110717411A (en) Pedestrian re-identification method based on deep layer feature fusion
US20180260415A1 (en) Instance-level image retrieval with a region proposal network
CN109711366B (en) Pedestrian re-identification method based on group information loss function
CN112818931A (en) Multi-scale pedestrian re-identification method based on multi-granularity depth feature fusion
CN108596010B (en) Implementation method of pedestrian re-identification system
CN110598543A (en) Model training method based on attribute mining and reasoning and pedestrian re-identification method
Halima et al. Bag of words based surveillance system using support vector machines
Verma et al. Wild animal detection from highly cluttered images using deep convolutional neural network
CN110516533A (en) A kind of pedestrian based on depth measure discrimination method again
CN113920472A (en) Unsupervised target re-identification method and system based on attention mechanism
Wang et al. Action recognition using linear dynamic systems
Dutra et al. Re-identifying people based on indexing structure and manifold appearance modeling
Al-Azzawy Eigenface and SIFT for gender classification
Ko et al. View-invariant, partially occluded human detection in still images using part bases and random forest
Mohapatra et al. Comparative study of corner and feature extractors for real-time object recognition in image processing
CN114565979B (en) Pedestrian re-identification method, system, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant