CN111126360A - Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model - Google Patents
Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model Download PDFInfo
- Publication number
- CN111126360A CN111126360A CN202010143811.2A CN202010143811A CN111126360A CN 111126360 A CN111126360 A CN 111126360A CN 202010143811 A CN202010143811 A CN 202010143811A CN 111126360 A CN111126360 A CN 111126360A
- Authority
- CN
- China
- Prior art keywords
- pictures
- loss function
- picture
- sample
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a cross-domain pedestrian re-identification method based on an unsupervised combined multi-loss model, and mainly solves the problem that the existing unsupervised cross-domain pedestrian re-identification method is low in identification rate. The scheme is that 1) a data set is obtained and divided into a training set and a testing set; 2) carrying out various preprocessing and expansion on the training set; 3) selecting a residual error network as a reference network model, initializing network parameters, and adjusting a network structure; 4) constructing a target domain loss function; 5) fusing a target domain loss function, a source domain loss function and a triple loss function to obtain a total loss function; 6) training the residual error network by using the total loss function to obtain a trained network model; 7) and inputting the test set into the trained network model, and outputting a recognition result. The invention improves the recognition rate of unsupervised cross-domain pedestrian re-recognition, effectively avoids the occurrence of over-fitting condition, and can be used for the intelligent security of suspect target searching and cross-camera tracking of pedestrians.
Description
Technical Field
The invention belongs to the field of computer vision and deep learning, and particularly relates to a cross-domain pedestrian re-identification method which can be used for the intelligent security field of suspect target searching and pedestrian cross-camera tracking.
Background
With the construction of digital cities, surveillance videos are more and more widely applied to various places where people live, the massive surveillance videos play an important role in public safety fields such as city public security and crime tracking, and how to better process the data is a great challenge in the future.
Pedestrian re-recognition, also known as pedestrian re-recognition, is a technique that uses computer vision techniques to determine whether a particular pedestrian is present in an image or video sequence, and is widely recognized as a sub-problem in image retrieval. Given a monitored pedestrian image, the pedestrian image is retrieved across the device.
In real life, under the influence of factors such as the visual angle of a camera, the posture of a pedestrian, shielding and illumination, images of the same person under different cameras are greatly different, and images of different pedestrians are likely to be very similar. This makes pedestrian re-identification a very challenging hot topic in the field of computer vision. The traditional pedestrian re-identification method is mainly researched from two aspects, namely, a pedestrian feature representation method is researched, and robust pedestrian features are designed to enable the pedestrian feature representation method to have invariance to factors such as visual angle change of a camera, pedestrian posture change, illumination change and background interference. And secondly, measurement learning, namely, the distance between different images of the same pedestrian is smaller and the distance between different pedestrians is larger by learning a distance measurement function.
In recent years, deep learning is rapidly developed, more and more researchers use deep learning technology to research pedestrian re-identification, and due to the good performance of deep learning in image high-level feature extraction and the appearance of large-scale data sets, the performance of the pedestrian re-identification method based on deep learning on each data set is better than that of the traditional pedestrian re-identification method. The existing pedestrian re-identification method based on deep learning comprises supervised learning and unsupervised learning, in the supervised learning, all data sets used for training need to be provided with labels, however, the labeling of data needs to consume a large amount of manpower and material resources, in real life, massive data are not provided with labels, and in the case of huge data, the manual labeling of the data is almost impossible. The pedestrian re-identification model learned through supervised learning has very poor performance on a label-free test set, the generalization capability of the model is very weak, and the expandability and the practicability are lacked in practical application. However, the pure utilization of unsupervised learning has poor performance of the learned network model due to no label of data, and cannot be applied in actual situations. By combining the two methods, how to obtain a better recognition result on a network model trained by a source domain with a label in a target domain without the label is a hotspot of pedestrian re-recognition research.
Hehe Fan, Liang Zheng et al, published in "Unsupervised person re-identification," Clustering and fining ". The thesis learns a deep network model as an initial pedestrian feature extractor using a labeled source domain data set, and then improves the network model on a target domain data set through unsupervised clustering. But this method cannot fully utilize the data information of the tagged source domain.
The paper Image-Image domain adaptation with predicted self-similarity and domain-similarity for person re-identification published by Wei jian Deng et al mainly studies how to reduce the difference of two cross-domain datasets at the pixel level. Although the method has good identification effect, the pedestrian re-identification error rate is high under the crossed cameras due to the fact that the intra-domain data characteristic distribution of the target domain is ignored.
Disclosure of Invention
The invention aims to provide a cross-domain pedestrian re-identification method based on an unsupervised joint multi-loss model, aiming at solving the problems in the prior art, so that the error rate of pedestrian re-identification on a label-free target domain data set is reduced, and the identification precision is improved.
In order to achieve the purpose, the technical scheme of the invention comprises the following steps:
1) acquiring a target domain data set without a label and a source domain data set with a label from an open website, randomly selecting a part of pictures in the target domain data set as a test set, dividing the test set into a query set and a candidate set, and taking the rest pictures in the target domain data set and the pictures in the source domain data set as training sets;
2) sequentially turning over, cutting and randomly erasing the pictures in the training set, preprocessing the data, and expanding the preprocessed pictures in the training set through cross-camera style migration; unifying the pictures of the training set to the same size;
3) selecting a residual error network as a reference network model, initializing network parameters, and adjusting the reference network structure;
4) constructing a target domain loss function:
4a) measuring the characteristics of each sample picture in the target domain data set by adopting cosine distance, sorting the distance from large to small, selecting the first k pictures as sample neighbors, and giving weight w to the neighbors of each picturei,j:
Wherein i is the sample number, j is the sample neighbor number, k is the neighbor number, xt,iIs a target domain sampleThis picture, M (x)t,iK) is a set of numbers of k neighbors of the sample picture;
4b) according to the weight wi,jConstructing a target domain loss function Ltgt:
Wherein the content of the first and second substances,is a randomly sampled picture in the training set, ntIs the number of pictures of the batch training target domain,is a pictureProbability of belonging to neighbor j;
5) fusing the target domain loss function with the existing source domain loss function and the triple loss function to obtain a total loss function L:
L=λ1Lsrc+λ2Ltgt+λ3LT
wherein L issrcAs a source domain loss function, LTIs a triplet loss function, λ1Is a weighted value, λ, of the proportion of the source domain loss function2Is the proportional weighted value of the loss function of the target domain, lambda3The weight value of the proportion of the triple loss function is calculated;
6) inputting the training set picture into a residual error network, training the residual error network by using a total loss function L and optimizing a training process to obtain a trained residual error network model;
7) inputting the pictures in the query set and the candidate set into a trained residual error network model for feature extraction, respectively calculating Euclidean distances between the features of the target pictures in the query set and the features of the pictures in the candidate set, sequencing the calculated Euclidean distances, and selecting the picture closest to the target in the query set in the candidate set as a recognition result.
The invention has the following advantages:
firstly, the invention effectively avoids the occurrence of the over-fitting condition and improves the generalization capability of the model because the data is subjected to various pretreatments of overturning, cutting and random erasing;
secondly, the data set is subjected to cross-camera style migration, so that the error rate of pedestrian re-identification under the view angle of the crossed cameras is reduced while data is expanded;
thirdly, the method greatly excavates the characteristic distribution characteristics of the label-free target domain data set due to the construction of the target domain loss function, fuses the source domain loss function, the target domain loss function and the triple loss function, and obviously improves the recognition rate of the unsupervised cross-domain pedestrian re-recognition.
Drawings
FIG. 1 is a flow chart of an implementation of the present invention;
FIG. 2 is a diagram of a residual network architecture in accordance with the present invention;
fig. 3 is a schematic diagram of a residual error network training process in the present invention.
Detailed Description
The invention is described in detail below with reference to the figures and specific embodiments. It is to be understood that the embodiments described are only a few embodiments of the present invention, and not all embodiments. All other embodiments that can be obtained by a person skilled in the art based on the embodiments of the present invention without any inventive step shall fall within the scope of the present invention.
Referring to fig. 1, the specific implementation of this example is as follows:
the method comprises the steps of firstly, acquiring a target domain data set and a source domain data set, and organizing the acquired data sets into a training set and a testing set.
Downloading a Market-1501 data set and a DukeMTMC-reiD data set from a public website, taking the Market-1501 data set as a target domain data set and taking the DukeMTMC-reiD data set as a source domain data set;
using DukeMTMC-reiD data set as the first training set R1;
The Market-1501 data set is divided into 6: 4 into a second training set R2And the test set T is further processed by the following steps of: 2 into a candidate set G and a query set Q;
the first training set R1And a second training set R2Forming a total training set R.
Step two, preprocessing the pictures in the training set R and expanding the second training set R2And unifying the picture size.
2.1) carrying out the following preprocessing on the training set R:
randomly extracting a plurality of pictures on the training set R for turning, including horizontal turning and vertical turning;
randomly extracting a plurality of pictures on the training set R for cutting, wherein the pictures are 2/3 of the length and the width of the original picture;
randomly extracting a plurality of pictures on the training set R to erase random areas in random sizes;
2.2) using the cyclic generation to generate a countermeasure network to carry out cross-camera style migration, and using a second training set R2Pictures shot by different cameras are input into the cyclic generation countermeasure network to obtain pictures with converted camera visual angles, the converted pictures are stored in a folder, and the pictures in the folder and the pictures before conversion are added into a second training set R2To the second training set R2The expansion is performed and all pictures are unified to 256 x 128 in size.
Referring to fig. 2, the structure of the residual error network selected in this example is described in further detail:
and step three, selecting a reference network model, initializing network parameters and adjusting the reference network structure.
3.1) selecting a residual error network ResNet-50 as a reference network model, and initializing the residual error network ResNet-50 by using model parameters pre-trained on ImageNet by utilizing transfer learning;
the residual error network ResNet-50 comprises five convolution layers, five pooling layers and a full connection layer, and the structural relationship is as follows in sequence: the first convolution layer, the first normalization processing layer, the first pooling layer, the second convolution layer, the second normalization processing layer, the second pooling layer, the third convolution layer, the third normalization processing layer, the third pooling layer, the fourth convolution layer, the fourth normalization processing layer, the fourth pooling layer, the fifth convolution layer, the fifth normalization processing layer, the fifth pooling layer and the 1000-dimensional full-connection layer;
3.2) adjusting the structure of ResNet-50:
3a) all layers before the fifth pooling layer of ResNet-50 are reserved, the last 1000-dimensional full-link layer is removed, and a 4096-dimensional output full-link layer is added;
3b) sequentially adding a normalization processing layer, a linear rectification function layer, a random discarding layer and two components after the 4096-dimensional output full connection layer, wherein the first component is a classification module of source domain data and comprises an M-dimensional full connection layer and an activation function layer; the second component is a sample memory module of the target domain, which is a key value pair structure with a key slot for storing picture characteristics and a value slot for storing the input number of the picture, and the key has uniqueness.
Referring to fig. 3, the training process of the residual error network in this example is described in further detail:
and step four, constructing a target domain loss function.
Based on the following three considerations, the objective domain loss function needs to be constructed:
regarding each sample of the target domain as an independent class, wherein each sample belongs to the class of the sample and is far away from other samples as far as possible;
the image after the cross-camera style migration and the image before the migration represent the same identity of the pedestrian, and the distance between the two images needs to be shortened;
for each sample picture, a sample with the same identity as the pedestrian exists in the neighbor of the sample picture, and the distance between the sample picture and the neighbor with the same identity is required to be shortened;
the objective loss function is constructed as follows:
4a) measuring the characteristics of each sample picture in the target domain data set by adopting cosine distanceThe cosine distance is the product of multiplying two eigenvectors and dividing the product by the modulus of the two eigenvectors, the larger the cosine distance is, the smaller the included angle between the two eigenvectors is, the closer the two eigenvectors are, otherwise, the farther the two eigenvectors are; according to the sorting of the distances from big to small, the first k pictures are selected as sample neighbors, and the neighbors of each picture are given a weight wi,j:
Wherein i is the sample number, j is the sample neighbor number, k is the neighbor number, xt,iIs a target domain sample picture, M (x)t,iK) is a sample picture xt,iThe number of k neighbors, k being 6 in this example;
4b) according to the weight wi,jTo obtain a target domain loss function Ltgt:
Wherein the content of the first and second substances,is a randomly sampled picture in the training set, ntIs the number of pictures of the batch training target domain,is a pictureProbability of belonging to neighbor j;
and step five, fusing the target domain loss function with the existing source domain loss function and the triple loss function.
5.1) Source Domain loss function:
the source domain loss function is a cross entropy loss function, and the formula is as follows:
in the formula, i is the sample number, m is the number of source domain pictures in the training batch, and xs,iFor source domain sample pictures, ys,iIs xs,iLabel information carried by the specimen, p (y)s,i|xs,i) Is xs,iBelong to ys,iThe probability of (d);
5.2) triple loss function:
because the source domain data set and the target domain data set do not have the same pedestrian identity, a picture before the target domain data set is subjected to style migration is randomly selected as an anchor sample, the picture after the picture is subjected to style migration is a positive sample, the picture of the source domain data set is taken as a negative sample, and the anchor sample, the positive sample and the negative sample form a three-tuple object.
Constructing a triple loss function by using the triple objects, wherein the formula is as follows:
wherein D is an edge parameter representing the minimum distance between the anchor point and the positive sample and the distance between the anchor point and the negative sample, D (.) is the Euclidean distance, and x istFor the picture before the target domain data set format migration,for pictures after style migration, xsThe source domain picture is taken, X is the size of the batch training picture set, and X is taken as 128 in this example;
5.3) fusing the target domain loss function, the source domain loss function and the triple loss function to obtain a total loss function L:
L=λ1Lsrc+λ2Ltgt+λ3LT
wherein L issrcAs a source domain loss function, LTIs a triplet loss function, λ1Is a weighted value, λ, of the proportion of the source domain loss function2Is the proportional weighted value of the loss function of the target domain, lambda3For the proportion weighted value of the triple loss function, lambda is taken in the example1=0.5,λ2=0.3,λ3=0.2。
And step six, inputting the training set R into the residual error network ResNet-50, and training the residual error network ResNet-50 to obtain a trained residual error network model.
The training process is as follows:
6a) all the features stored in the sample memory are initialized to 0, and the sample memory is updated by the following formula:
K[i]←αK[i]+(1-α)f(xt,i)
wherein α is a hyperparameter, α epsilon [0,1 ∈ ]]Controlling the update frequency, K [ i ]]For sample picture xt,iFeatures stored in the sample memory, f (x)t,i) Is xt,iThe feature vector of the sample, which is α in this example, is 0.01 as the initial value;
6b) setting the total iteration times of training as 60 times, the number of batch pictures as 128, selecting the optimizer as a random gradient optimizer, setting the learning rate as 0.1 in the first 40 iterations, and dividing the learning rate by 10 in each iteration in the last 20 iterations;
6c) dividing a training set R into a plurality of small-batch data sets by taking 128 as a base number by adopting a small-batch iteration method, sequentially sending each small-batch data set into a residual error network ResNet-50 to extract a characteristic vector, inputting the obtained characteristic vector into a total loss function L to calculate loss, and calculating the gradient of the loss obtained by calculation;
6d) inputting the obtained gradient into a random gradient optimizer to update the weight and the threshold of the network;
6e) and repeating the steps 6c) to 6d), and stopping training when the total iteration times reach 60 times to obtain a trained model.
And step seven, testing the trained model and outputting a recognition result.
7.1) inputting the query set Q and the candidate set G into a trained residual error network model for feature extraction;
7.2) respectively calculating the Euclidean distance of the characteristics of the target picture in the query set Q and the characteristics of each picture in the candidate set G, wherein the formula is as follows:
wherein x is the target picture, yiN is the number of pictures in the candidate set G for inquiring the pictures;
7.3) ordering the calculated Euclidean distances from small to large, and selecting the picture closest to the target picture in the query set Q from the candidate set G as an identification result.
The effects of the present invention can be further explained by the evaluation results of the following experiments.
1. Experimental methods and evaluation criteria
The method used in the experiment was the present invention and the existing HHL method, an unsupervised cross-domain pedestrian re-identification method, HHL for short, proposed in the paper "general aperson regenerative model-and homogeneous", published by Zhuin Zhong, et al 2018 at the ECCV (European Conference on Computer Vision) Conference.
Evaluation criteria, using three criteria mAP, rank-1, rank-5 in the area of pedestrian re-identification, wherein:
mAP is average precision mean value, which is to sum and average the average precision in the multi-classification task;
rank-1 is the probability that the result of the 1 st graph in the search results is correct;
rank-5 is the probability that the results of the first 5 pictures in the search results are correct;
2. contents and results of the experiments
The results of the tests on the Market-1501 data set using the method of the present invention and the prior HHL method are shown in table 1.
TABLE 1
Method of producing a composite material | mAP | rank-1 | rank-5 |
The invention | 43.1 | 75.2 | 87.9 |
HHL | 31.4 | 66.4 | 78.8 |
As can be seen from Table 1, compared with the prior art, the average accuracy mean mAP of the method is improved by 11.7, the probability rank-1 that the result of the 1 st graph in the search result is correct is improved by 8.8, and the probability rank-5 that the result of the first 5 graphs in the search result is correct is improved by 9.1.
Claims (8)
1. A cross-domain pedestrian re-identification method based on an unsupervised combined multi-loss model is characterized by comprising the following steps:
1) acquiring a target domain data set without a label and a source domain data set with a label from an open website, randomly selecting a part of pictures in the target domain data set as a test set, dividing the test set into a query set and a candidate set, and taking the rest pictures in the target domain data set and the pictures in the source domain data set as training sets;
2) sequentially turning over, cutting and randomly erasing the pictures in the training set, preprocessing the data, and expanding the preprocessed pictures in the training set through cross-camera style migration; unifying the pictures of the training set to the same size;
3) selecting a residual error network as a reference network model, initializing network parameters, and adjusting the reference network structure;
4) constructing a target domain loss function:
4a) measuring the characteristics of each sample picture in the target domain data set by adopting cosine distance, sorting the distance from large to small, selecting the first k pictures as sample neighbors, and giving weight w to the neighbors of each picturei,j:
Wherein i is the sample number, j is the sample neighbor number, k is the neighbor number, xt,iIs a target domain sample picture, M (x)t,iK) is a set of numbers of k neighbors of the sample picture;
4b) according to the weight wi,jConstructing a target domain loss function Ltgt:
Wherein the content of the first and second substances,is a randomly sampled picture in the training set, ntIs the number of pictures of the batch training target domain,is a pictureProbability of belonging to neighbor j;
5) fusing the target domain loss function with the existing source domain loss function and the triple loss function to obtain a total loss function L:
L=λ1Lsrc+λ2Ltgt+λ3LT
wherein L issrcAs a source domain loss function, LTIs a triplet loss function, λ1Is a weighted value, λ, of the proportion of the source domain loss function2Occupied for target domain loss functionProportional weight value, λ3The weight value of the proportion of the triple loss function is calculated;
6) inputting the training set picture into a residual error network, training the residual error network by using a total loss function L and optimizing a training process to obtain a trained residual error network model;
7) inputting the pictures in the query set and the candidate set into a trained residual error network model for feature extraction, respectively calculating Euclidean distances between the features of the target pictures in the query set and the features of the pictures in the candidate set, sequencing the calculated Euclidean distances, and selecting the picture closest to the target in the query set in the candidate set as a recognition result.
2. The method of claim 1, wherein 2) the cyclic generation countermeasure network is utilized to perform the cross-camera style migration, the pictures taken by different cameras in the target domain data set are input into the cyclic generation countermeasure network, the pictures with converted camera view angles are obtained, and the converted pictures and the pictures before conversion are added into the target domain data set together to expand the data set.
3. The method of claim 1, wherein in 3), migration learning is used, model parameters pre-trained on ImageNet are used for initializing the residual error network, and the structure of the residual error network is adjusted, so that the following is achieved:
3a) removing the last full connection layer of the residual error network and adding a 4096-dimensional output full connection layer;
3b) sequentially adding a normalization processing layer, a linear rectification function layer, a random discarding layer and two components after the 4096-dimensional output full connection layer, wherein the first component is a classification module of source domain data and comprises an M-dimensional full connection layer and an activation function layer; the second component is a sample memory module of the target domain, which is a key value pair structure with a key slot for storing picture characteristics and a value slot for storing the input number of the picture, the key having uniqueness.
4. The method of claim 1, wherein the cosine distance in 4) is a product of multiplying two eigenvectors and dividing the product by a modulus of the two eigenvectors, and wherein the larger the cosine distance, the smaller the included angle between the two eigenvectors, the closer the two eigenvectors are, and vice versa, the farther the distance is.
5. The method of claim 1, wherein the source domain loss function in 5) is a cross-entropy loss function, and the formula is as follows:
in the formula, i is the sample number, m is the number of source domain pictures in the training batch, and xs,iFor source domain sample pictures, ys,iIs xs,iLabel information carried by the specimen, p (y)s,i|xs,i) Is xs,iBelong to ys,iThe probability of (c).
6. The method according to claim 1, wherein 5) a picture before the target domain data set style migration is randomly selected as an anchor sample, the picture after the picture style migration is a positive sample, the picture of the source domain data set is a negative sample, and a triplet object is formed by the anchor sample, the positive sample and the negative sample;
constructing a triple loss function by using the triple objects, wherein the formula is as follows:
wherein D is an edge parameter representing the minimum distance between the anchor point and the positive sample and the distance between the anchor point and the negative sample, D (.) is the Euclidean distance, and x istFor the picture before the target domain data set format migration,for pictures after style migration, xsIs the source domain picture, X is the batch training picture set size.
7. The method of claim 1, wherein 6) the residual network is trained using the total loss function L, and the method is implemented as follows:
6a) all the features stored in the sample memory are initialized to 0, and the sample memory is updated by the following formula:
K[i]←αK[i]+(1-α)f(xt,i)
wherein α is a hyperparameter, α epsilon [0,1 ∈ ]]Controlling the update frequency, K [ i ]]For sample picture xt,iFeatures stored in the sample memory, f (x)t,i) Is xt,iA feature vector of the sample;
6b) setting the total iteration times, the learning rate and the number of batch pictures of training, and selecting an optimizer;
6c) dividing a data set into a plurality of small-batch data sets by using a small-batch iteration method based on the number of batch pictures, sequentially sending each small-batch data set into a residual error network to extract a characteristic vector, and inputting the obtained characteristic vector into a total loss function L to calculate loss;
6d) solving the gradient of the calculated loss, and inputting the obtained gradient into the selected optimizer to update the network parameters;
6e) and repeating the steps 6c) to 6d), and stopping training when the total iteration times are reached to obtain a trained model.
8. The method according to claim 1, wherein 7) the euclidean distances are calculated for the features of the target pictures in the query set and the picture features in the candidate set respectively by the following formula:
wherein Q is a query set, G is a candidate set, x is a target picture, and yiFor querying pictures, n is the number of pictures.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2019111175727 | 2019-11-15 | ||
CN201911117572 | 2019-11-15 |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111126360A true CN111126360A (en) | 2020-05-08 |
CN111126360B CN111126360B (en) | 2023-03-24 |
Family
ID=70493521
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010143811.2A Active CN111126360B (en) | 2019-11-15 | 2020-03-04 | Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111126360B (en) |
Cited By (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111598004A (en) * | 2020-05-18 | 2020-08-28 | 北京星闪世图科技有限公司 | Progressive-enhancement self-learning unsupervised cross-domain pedestrian re-identification method |
CN111783576A (en) * | 2020-06-18 | 2020-10-16 | 西安电子科技大学 | Pedestrian re-identification method based on improved YOLOv3 network and feature fusion |
CN111814655A (en) * | 2020-07-03 | 2020-10-23 | 浙江大华技术股份有限公司 | Target re-identification method, network training method thereof and related device |
CN111814854A (en) * | 2020-06-28 | 2020-10-23 | 北京交通大学 | Target re-identification method adaptive to unsupervised domain |
CN111881714A (en) * | 2020-05-22 | 2020-11-03 | 北京交通大学 | Unsupervised cross-domain pedestrian re-identification method |
CN111931637A (en) * | 2020-08-07 | 2020-11-13 | 华南理工大学 | Cross-modal pedestrian re-identification method and system based on double-current convolutional neural network |
CN111950411A (en) * | 2020-07-31 | 2020-11-17 | 上海商汤智能科技有限公司 | Model determination method and related device |
CN112069929A (en) * | 2020-08-20 | 2020-12-11 | 之江实验室 | Unsupervised pedestrian re-identification method and device, electronic equipment and storage medium |
CN112084895A (en) * | 2020-08-25 | 2020-12-15 | 南京邮电大学 | Pedestrian re-identification method based on deep learning |
CN112215280A (en) * | 2020-10-12 | 2021-01-12 | 西安交通大学 | Small sample image classification method based on meta-backbone network |
CN112287089A (en) * | 2020-11-23 | 2021-01-29 | 腾讯科技(深圳)有限公司 | Classification model training and automatic question-answering method and device for automatic question-answering system |
CN112347995A (en) * | 2020-11-30 | 2021-02-09 | 中国科学院自动化研究所 | Unsupervised pedestrian re-identification method based on fusion of pixel and feature transfer |
CN112381056A (en) * | 2020-12-02 | 2021-02-19 | 山西大学 | Cross-domain pedestrian re-identification method and system fusing multiple source domains |
CN112434599A (en) * | 2020-11-23 | 2021-03-02 | 同济大学 | Pedestrian re-identification method based on random shielding recovery of noise channel |
CN112560925A (en) * | 2020-12-10 | 2021-03-26 | 中国科学院深圳先进技术研究院 | Complex scene target detection data set construction method and system |
CN112633071A (en) * | 2020-11-30 | 2021-04-09 | 之江实验室 | Pedestrian re-identification data domain adaptation method based on data style decoupling content migration |
CN112669827A (en) * | 2020-12-28 | 2021-04-16 | 清华大学 | Joint optimization method and system for automatic speech recognizer |
CN112733695A (en) * | 2021-01-04 | 2021-04-30 | 电子科技大学 | Unsupervised key frame selection method in pedestrian re-identification field |
CN112818175A (en) * | 2021-02-07 | 2021-05-18 | 中国矿业大学 | Factory worker searching method and training method of worker recognition model |
CN112861705A (en) * | 2021-02-04 | 2021-05-28 | 东北林业大学 | Cross-domain pedestrian re-identification method based on hybrid learning |
CN113065516A (en) * | 2021-04-22 | 2021-07-02 | 中国矿业大学 | Unsupervised pedestrian re-identification system and method based on sample separation |
CN113095263A (en) * | 2021-04-21 | 2021-07-09 | 中国矿业大学 | Method and device for training heavy identification model of pedestrian under shielding and method and device for heavy identification of pedestrian under shielding |
CN113095229A (en) * | 2021-04-14 | 2021-07-09 | 中国矿业大学 | Unsupervised domain self-adaptive pedestrian re-identification system and method |
CN113128466A (en) * | 2021-05-11 | 2021-07-16 | 深圳大学 | Pedestrian re-identification method, system, electronic device and storage medium |
CN113139536A (en) * | 2021-05-12 | 2021-07-20 | 哈尔滨工业大学(威海) | Text verification code identification method and equipment based on cross-domain meta learning and storage medium |
CN113221656A (en) * | 2021-04-13 | 2021-08-06 | 电子科技大学 | Cross-domain pedestrian re-identification model based on domain invariant features and method thereof |
CN113255823A (en) * | 2021-06-15 | 2021-08-13 | 中国人民解放军国防科技大学 | Unsupervised domain adaptation method and unsupervised domain adaptation device |
CN113378620A (en) * | 2021-03-31 | 2021-09-10 | 中交第二公路勘察设计研究院有限公司 | Cross-camera pedestrian re-identification method in surveillance video noise environment |
CN113537403A (en) * | 2021-08-14 | 2021-10-22 | 北京达佳互联信息技术有限公司 | Training method and device and prediction method and device of image processing model |
CN113723345A (en) * | 2021-09-09 | 2021-11-30 | 河北工业大学 | Domain-adaptive pedestrian re-identification method based on style conversion and joint learning network |
CN113837256A (en) * | 2021-09-15 | 2021-12-24 | 深圳市商汤科技有限公司 | Object recognition method, network training method and device, equipment and medium |
CN113936301A (en) * | 2021-07-02 | 2022-01-14 | 西北工业大学 | Target re-identification method based on central point prediction loss function |
CN114220124A (en) * | 2021-12-16 | 2022-03-22 | 华南农业大学 | Near-infrared-visible light cross-modal double-flow pedestrian re-identification method and system |
CN115100591A (en) * | 2022-06-17 | 2022-09-23 | 哈尔滨工业大学 | Multi-target tracking and target re-identification system and method based on joint learning |
CN115147774A (en) * | 2022-07-05 | 2022-10-04 | 中国科学技术大学 | Pedestrian re-identification method in degradation environment based on feature alignment |
CN115345259A (en) * | 2022-10-14 | 2022-11-15 | 北京睿企信息科技有限公司 | Optimization method, equipment and storage medium for training named entity recognition model |
EP4136580A4 (en) * | 2020-06-29 | 2024-01-17 | Zhejiang Dahua Technology Co | Target re-identification method, network training method thereof, and related device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109948561A (en) * | 2019-03-25 | 2019-06-28 | 广东石油化工学院 | The method and system that unsupervised image/video pedestrian based on migration network identifies again |
WO2019128367A1 (en) * | 2017-12-26 | 2019-07-04 | 广州广电运通金融电子股份有限公司 | Face verification method and apparatus based on triplet loss, and computer device and storage medium |
CN110008842A (en) * | 2019-03-09 | 2019-07-12 | 同济大学 | A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth |
CN110321813A (en) * | 2019-06-18 | 2019-10-11 | 南京信息工程大学 | Cross-domain pedestrian recognition methods again based on pedestrian's segmentation |
CN110414462A (en) * | 2019-08-02 | 2019-11-05 | 中科人工智能创新技术研究院(青岛)有限公司 | A kind of unsupervised cross-domain pedestrian recognition methods and system again |
-
2020
- 2020-03-04 CN CN202010143811.2A patent/CN111126360B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019128367A1 (en) * | 2017-12-26 | 2019-07-04 | 广州广电运通金融电子股份有限公司 | Face verification method and apparatus based on triplet loss, and computer device and storage medium |
CN110008842A (en) * | 2019-03-09 | 2019-07-12 | 同济大学 | A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth |
CN109948561A (en) * | 2019-03-25 | 2019-06-28 | 广东石油化工学院 | The method and system that unsupervised image/video pedestrian based on migration network identifies again |
CN110321813A (en) * | 2019-06-18 | 2019-10-11 | 南京信息工程大学 | Cross-domain pedestrian recognition methods again based on pedestrian's segmentation |
CN110414462A (en) * | 2019-08-02 | 2019-11-05 | 中科人工智能创新技术研究院(青岛)有限公司 | A kind of unsupervised cross-domain pedestrian recognition methods and system again |
Non-Patent Citations (1)
Title |
---|
王兴柱等: "三元组在非监督行人再识别中的应用", 《仪表技术》 * |
Cited By (60)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111598004A (en) * | 2020-05-18 | 2020-08-28 | 北京星闪世图科技有限公司 | Progressive-enhancement self-learning unsupervised cross-domain pedestrian re-identification method |
CN111598004B (en) * | 2020-05-18 | 2023-12-08 | 江苏星闪世图科技(集团)有限公司 | Progressive reinforcement self-learning unsupervised cross-domain pedestrian re-identification method |
CN111881714A (en) * | 2020-05-22 | 2020-11-03 | 北京交通大学 | Unsupervised cross-domain pedestrian re-identification method |
CN111881714B (en) * | 2020-05-22 | 2023-11-21 | 北京交通大学 | Unsupervised cross-domain pedestrian re-identification method |
CN111783576A (en) * | 2020-06-18 | 2020-10-16 | 西安电子科技大学 | Pedestrian re-identification method based on improved YOLOv3 network and feature fusion |
CN111783576B (en) * | 2020-06-18 | 2023-08-18 | 西安电子科技大学 | Pedestrian re-identification method based on improved YOLOv3 network and feature fusion |
CN111814854B (en) * | 2020-06-28 | 2023-07-28 | 北京交通大学 | Target re-identification method without supervision domain adaptation |
CN111814854A (en) * | 2020-06-28 | 2020-10-23 | 北京交通大学 | Target re-identification method adaptive to unsupervised domain |
WO2022001489A1 (en) * | 2020-06-28 | 2022-01-06 | 北京交通大学 | Unsupervised domain adaptation target re-identification method |
EP4136580A4 (en) * | 2020-06-29 | 2024-01-17 | Zhejiang Dahua Technology Co | Target re-identification method, network training method thereof, and related device |
CN111814655B (en) * | 2020-07-03 | 2023-09-01 | 浙江大华技术股份有限公司 | Target re-identification method, network training method thereof and related device |
CN111814655A (en) * | 2020-07-03 | 2020-10-23 | 浙江大华技术股份有限公司 | Target re-identification method, network training method thereof and related device |
CN111950411A (en) * | 2020-07-31 | 2020-11-17 | 上海商汤智能科技有限公司 | Model determination method and related device |
CN111931637B (en) * | 2020-08-07 | 2023-09-15 | 华南理工大学 | Cross-modal pedestrian re-identification method and system based on double-flow convolutional neural network |
CN111931637A (en) * | 2020-08-07 | 2020-11-13 | 华南理工大学 | Cross-modal pedestrian re-identification method and system based on double-current convolutional neural network |
CN112069929A (en) * | 2020-08-20 | 2020-12-11 | 之江实验室 | Unsupervised pedestrian re-identification method and device, electronic equipment and storage medium |
CN112069929B (en) * | 2020-08-20 | 2024-01-05 | 之江实验室 | Unsupervised pedestrian re-identification method and device, electronic equipment and storage medium |
CN112084895A (en) * | 2020-08-25 | 2020-12-15 | 南京邮电大学 | Pedestrian re-identification method based on deep learning |
CN112084895B (en) * | 2020-08-25 | 2022-07-29 | 南京邮电大学 | Pedestrian re-identification method based on deep learning |
CN112215280A (en) * | 2020-10-12 | 2021-01-12 | 西安交通大学 | Small sample image classification method based on meta-backbone network |
CN112287089A (en) * | 2020-11-23 | 2021-01-29 | 腾讯科技(深圳)有限公司 | Classification model training and automatic question-answering method and device for automatic question-answering system |
CN112434599A (en) * | 2020-11-23 | 2021-03-02 | 同济大学 | Pedestrian re-identification method based on random shielding recovery of noise channel |
CN112287089B (en) * | 2020-11-23 | 2022-09-20 | 腾讯科技(深圳)有限公司 | Classification model training and automatic question-answering method and device for automatic question-answering system |
CN112347995A (en) * | 2020-11-30 | 2021-02-09 | 中国科学院自动化研究所 | Unsupervised pedestrian re-identification method based on fusion of pixel and feature transfer |
CN112633071A (en) * | 2020-11-30 | 2021-04-09 | 之江实验室 | Pedestrian re-identification data domain adaptation method based on data style decoupling content migration |
CN112347995B (en) * | 2020-11-30 | 2022-09-23 | 中国科学院自动化研究所 | Unsupervised pedestrian re-identification method based on fusion of pixel and feature transfer |
CN112381056A (en) * | 2020-12-02 | 2021-02-19 | 山西大学 | Cross-domain pedestrian re-identification method and system fusing multiple source domains |
CN112381056B (en) * | 2020-12-02 | 2022-04-01 | 山西大学 | Cross-domain pedestrian re-identification method and system fusing multiple source domains |
CN112560925A (en) * | 2020-12-10 | 2021-03-26 | 中国科学院深圳先进技术研究院 | Complex scene target detection data set construction method and system |
CN112669827B (en) * | 2020-12-28 | 2022-08-02 | 清华大学 | Joint optimization method and system for automatic speech recognizer |
CN112669827A (en) * | 2020-12-28 | 2021-04-16 | 清华大学 | Joint optimization method and system for automatic speech recognizer |
CN112733695A (en) * | 2021-01-04 | 2021-04-30 | 电子科技大学 | Unsupervised key frame selection method in pedestrian re-identification field |
CN112861705A (en) * | 2021-02-04 | 2021-05-28 | 东北林业大学 | Cross-domain pedestrian re-identification method based on hybrid learning |
CN112861705B (en) * | 2021-02-04 | 2022-07-05 | 东北林业大学 | Cross-domain pedestrian re-identification method based on hybrid learning |
CN112818175A (en) * | 2021-02-07 | 2021-05-18 | 中国矿业大学 | Factory worker searching method and training method of worker recognition model |
CN112818175B (en) * | 2021-02-07 | 2023-09-01 | 中国矿业大学 | Factory staff searching method and training method of staff identification model |
CN113378620B (en) * | 2021-03-31 | 2023-04-07 | 中交第二公路勘察设计研究院有限公司 | Cross-camera pedestrian re-identification method in surveillance video noise environment |
CN113378620A (en) * | 2021-03-31 | 2021-09-10 | 中交第二公路勘察设计研究院有限公司 | Cross-camera pedestrian re-identification method in surveillance video noise environment |
CN113221656A (en) * | 2021-04-13 | 2021-08-06 | 电子科技大学 | Cross-domain pedestrian re-identification model based on domain invariant features and method thereof |
CN113095229A (en) * | 2021-04-14 | 2021-07-09 | 中国矿业大学 | Unsupervised domain self-adaptive pedestrian re-identification system and method |
CN113095229B (en) * | 2021-04-14 | 2024-04-12 | 中国矿业大学 | Self-adaptive pedestrian re-identification system and method for unsupervised domain |
CN113095263B (en) * | 2021-04-21 | 2024-02-20 | 中国矿业大学 | Training method and device for pedestrian re-recognition model under shielding and pedestrian re-recognition method and device under shielding |
CN113095263A (en) * | 2021-04-21 | 2021-07-09 | 中国矿业大学 | Method and device for training heavy identification model of pedestrian under shielding and method and device for heavy identification of pedestrian under shielding |
CN113065516B (en) * | 2021-04-22 | 2023-12-01 | 中国矿业大学 | Sample separation-based unsupervised pedestrian re-identification system and method |
CN113065516A (en) * | 2021-04-22 | 2021-07-02 | 中国矿业大学 | Unsupervised pedestrian re-identification system and method based on sample separation |
CN113128466B (en) * | 2021-05-11 | 2023-12-05 | 深圳大学 | Pedestrian re-recognition method, system, electronic device and storage medium |
CN113128466A (en) * | 2021-05-11 | 2021-07-16 | 深圳大学 | Pedestrian re-identification method, system, electronic device and storage medium |
CN113139536A (en) * | 2021-05-12 | 2021-07-20 | 哈尔滨工业大学(威海) | Text verification code identification method and equipment based on cross-domain meta learning and storage medium |
CN113255823A (en) * | 2021-06-15 | 2021-08-13 | 中国人民解放军国防科技大学 | Unsupervised domain adaptation method and unsupervised domain adaptation device |
CN113936301B (en) * | 2021-07-02 | 2024-03-12 | 西北工业大学 | Target re-identification method based on center point prediction loss function |
CN113936301A (en) * | 2021-07-02 | 2022-01-14 | 西北工业大学 | Target re-identification method based on central point prediction loss function |
CN113537403A (en) * | 2021-08-14 | 2021-10-22 | 北京达佳互联信息技术有限公司 | Training method and device and prediction method and device of image processing model |
CN113723345B (en) * | 2021-09-09 | 2023-11-14 | 河北工业大学 | Domain self-adaptive pedestrian re-identification method based on style conversion and joint learning network |
CN113723345A (en) * | 2021-09-09 | 2021-11-30 | 河北工业大学 | Domain-adaptive pedestrian re-identification method based on style conversion and joint learning network |
CN113837256A (en) * | 2021-09-15 | 2021-12-24 | 深圳市商汤科技有限公司 | Object recognition method, network training method and device, equipment and medium |
CN114220124A (en) * | 2021-12-16 | 2022-03-22 | 华南农业大学 | Near-infrared-visible light cross-modal double-flow pedestrian re-identification method and system |
CN115100591A (en) * | 2022-06-17 | 2022-09-23 | 哈尔滨工业大学 | Multi-target tracking and target re-identification system and method based on joint learning |
CN115147774A (en) * | 2022-07-05 | 2022-10-04 | 中国科学技术大学 | Pedestrian re-identification method in degradation environment based on feature alignment |
CN115147774B (en) * | 2022-07-05 | 2024-04-02 | 中国科学技术大学 | Pedestrian re-identification method based on characteristic alignment in degradation environment |
CN115345259A (en) * | 2022-10-14 | 2022-11-15 | 北京睿企信息科技有限公司 | Optimization method, equipment and storage medium for training named entity recognition model |
Also Published As
Publication number | Publication date |
---|---|
CN111126360B (en) | 2023-03-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111126360B (en) | Cross-domain pedestrian re-identification method based on unsupervised combined multi-loss model | |
Wu et al. | Deep learning-based methods for person re-identification: A comprehensive review | |
CN110414368B (en) | Unsupervised pedestrian re-identification method based on knowledge distillation | |
CN111259786B (en) | Pedestrian re-identification method based on synchronous enhancement of appearance and motion information of video | |
CN109961051B (en) | Pedestrian re-identification method based on clustering and block feature extraction | |
Lin et al. | RSCM: Region selection and concurrency model for multi-class weather recognition | |
Tao et al. | Person re-identification by regularized smoothing kiss metric learning | |
CN111709311B (en) | Pedestrian re-identification method based on multi-scale convolution feature fusion | |
CN111783576B (en) | Pedestrian re-identification method based on improved YOLOv3 network and feature fusion | |
CN110717411A (en) | Pedestrian re-identification method based on deep layer feature fusion | |
US20180260415A1 (en) | Instance-level image retrieval with a region proposal network | |
CN109711366B (en) | Pedestrian re-identification method based on group information loss function | |
CN112818931A (en) | Multi-scale pedestrian re-identification method based on multi-granularity depth feature fusion | |
CN108596010B (en) | Implementation method of pedestrian re-identification system | |
CN110598543A (en) | Model training method based on attribute mining and reasoning and pedestrian re-identification method | |
Halima et al. | Bag of words based surveillance system using support vector machines | |
Verma et al. | Wild animal detection from highly cluttered images using deep convolutional neural network | |
CN110516533A (en) | A kind of pedestrian based on depth measure discrimination method again | |
CN113920472A (en) | Unsupervised target re-identification method and system based on attention mechanism | |
Wang et al. | Action recognition using linear dynamic systems | |
Dutra et al. | Re-identifying people based on indexing structure and manifold appearance modeling | |
Al-Azzawy | Eigenface and SIFT for gender classification | |
Ko et al. | View-invariant, partially occluded human detection in still images using part bases and random forest | |
Mohapatra et al. | Comparative study of corner and feature extractors for real-time object recognition in image processing | |
CN114565979B (en) | Pedestrian re-identification method, system, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |