CN110163110A - A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic - Google Patents

A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic Download PDF

Info

Publication number
CN110163110A
CN110163110A CN201910329733.2A CN201910329733A CN110163110A CN 110163110 A CN110163110 A CN 110163110A CN 201910329733 A CN201910329733 A CN 201910329733A CN 110163110 A CN110163110 A CN 110163110A
Authority
CN
China
Prior art keywords
pedestrian
image
training
depth
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910329733.2A
Other languages
Chinese (zh)
Other versions
CN110163110B (en
Inventor
丁剑飞
王进
阚丹会
闫盈盈
曹扬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Division Big Data Research Institute Co Ltd
Original Assignee
Division Big Data Research Institute Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Division Big Data Research Institute Co Ltd filed Critical Division Big Data Research Institute Co Ltd
Priority to CN201910329733.2A priority Critical patent/CN110163110B/en
Publication of CN110163110A publication Critical patent/CN110163110A/en
Application granted granted Critical
Publication of CN110163110B publication Critical patent/CN110163110B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/2163Partitioning the feature space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The present invention provides a kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic, comprising the following steps: pre-training-human body attitude correction and segmentation-feature vector-depth characteristic fusion-training pattern-test model-recognition result.The present invention is by extracting the global and local feature of pedestrian using depth convolutional neural networks, depth integration is carried out to two kinds of features and obtains final pedestrian's characteristic present, then in depth convolutional neural networks training process, by the way of transfer learning and then the acquisition better pedestrian of effect identifies network model again, the feature for finally making pedestrian identify that network model extracts again has stronger resolution capability, to achieve the purpose that promote pedestrian's weight recognition accuracy.

Description

A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic
Technical field
The present invention relates to a kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic, belong to depth It practises, transfer learning technical field.
Background technique
Pedestrian identifies again, pedestrian's matching task that main purpose carries out under the multi-cam network of non-overlap visual angle domain, i.e., Find out the target pedestrian that the camera of different location is shot in different moments.
With the development of artificial intelligence technology, pedestrian's weight identification technology in the application scenarios such as public security protection, image retrieval It is widely studied domain concern.But the pedestrian's weight traditional biologicals such as identification technology and recognition of face, gesture identification identification technology phase Than, because of the complicated reasons such as uncontrollable of the environment of monitor video, and then face low image resolution ratio, visual angle change, attitudes vibration, Light changes and blocks etc. the problem that factors cause accuracy of identification low.Therefore, pedestrian's weight identification technology is in practical application field Biggish challenge is faced under scape.
In order to promote the robustness of accuracy rate that pedestrian identifies and enhancing system again, numerous scholars by long-term research, Propose different pedestrians recognition methods again.It is spacious to identify that field obtains huge progress again in pedestrian depending on science and technology Facce++, it should A kind of new method is proposed in the paper AlignedReID that team delivers, is passed through dynamic alignment (Dynamic Alignment) With Cooperative Study (Mutual Learing), then at rearrangement (Re-Ranking), which is found through experiments that, Test phase extracts pedestrian's global characteristics and the recognition accuracy of fusion pedestrian overall situation and partial situation feature is very nearly the same;The bases such as Yi A kind of depth measure learning method is proposed in twin convolutional neural networks, is achieved good results;Liu et al. is based on neighborhood Constituent analysis and depth confidence network propose a kind of depth nonlinear metric learning method, and the effect of neighborhood transform analysis is logical Crossing data transformation maximizes the recognizable number of samples of every class data in training data, in order to extend in neighborhood transform analysis Data transformation, learnt using depth confidence network nonlinear characteristic transformation.But it finds under study for action, above-mentioned pedestrian knows again Other method is most of to extract global characteristics vector, while some sides as input based on pedestrian's global image in the training process Although method is extracted local feature, pedestrian's local feature is not made full use of to carry out depth integration, and obtaining has distinction Characterization image.And simply finely tuned being expert on personal data library using pre-training model, there is no in view of source domain with Data distribution difference between aiming field data set.And then it is undesirable to network migration effect.
Summary of the invention
In order to solve the above technical problems, the present invention provides a kind of pedestrians merged based on transfer learning and depth characteristic Recognition methods again, being somebody's turn to do the pedestrian based on transfer learning and depth characteristic fusion, recognition methods can be directed to pedestrian's global characteristics again Depth integration is not carried out with local feature, and the difference of data distribution is not fully taken into account in network trim process.
The present invention is achieved by the following technical programs.
A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic provided by the invention, including it is following Step:
1. pre-training again: the pre-training model based on ImageNet being identified in pedestrian and carries out pre-training in data, is obtained Pedestrian identifies pre-training network model again;
2. human body attitude correction and segmentation: choosing difficulty on personal data of being expert at collection and divide sample pair, and be input to skeleton pass In 14 key points that key point detection network detects, and human body attitude correction and pedestrian part ROI segmentation are carried out, obtained The enhanced difficulty of data divide sample to, correction after global and local image;
3. feature vector: dividing the global and local image after correction, the enhanced difficulty of data to sample pair, be input to row People identifies pre-training network model again, obtains pedestrian part and global characteristics vector;
4. depth characteristic merges: pedestrian part and global characteristics vector being carried out depth characteristic fusion, obtain final pedestrian Feature vector;
5. training pattern: by the way of transfer learning and step 4. in final pedestrian's feature vector, pedestrian is identified again Pre-training network model is finely adjusted, and is identified again in pedestrian and added adaptation layer in pre-training network model altogether, and pedestrian is obtained Network model is identified again;
6. test model: input inquiry pedestrian and target pedestrian image identify that network model extracts two using only pedestrian again A pedestrian's global characteristics vector with distinction;
7. recognition result: based on pedestrian's global characteristics vector in step 6., calculating inquiry pedestrian and target line personal data The similarity between any one image is concentrated, it is considered as identical pedestrian that wherein similarity is highest.
Pedestrian in training stage identifies the input of network model again, using triple pedestrian image.
1. the step is divided into following steps:
(1.1) the preparatory trained depth convolutional network model on ImageNet data set is obtained, and by it in pedestrian It identifies and is trained in data again;
(1.2) when pedestrian identifies pre-training depth convolutional neural networks model in data again, merely with sample mark Information is finely adjusted depth convolutional network model.
The step (1.2) is divided into following steps:
The ResNet50 network model of the pre-training on ImageNet data set is removed the full connection of top layer by (1.2.1) Layer, adds two layers of full articulamentum and one layer softmax layers after maximum pond layer;
The label information that (1.2.2) is marked using pedestrian image is finely adjusted the depth convolutional neural networks of building, The three first layers of constant depth convolutional neural networks in trim process;
(1.2.3) obtains the prediction probability of pedestrian's global image according to depth convolutional neural networks;
(1.2.4) defines the loss function in depth convolutional neural networks according to prediction probability.
The step 2. in global and local image after the correction that obtains, for difficult point of positive negative sample triad row people figure Pedestrian's global image and local ROI image after picture, human body attitude correction.
2. the step is divided into following steps:
(2.1) each is trained into batch, selects the pedestrian of P ID at random, each pedestrian select at random K open it is different Image, each batch contain P × K pedestrian images;
(2.2) each is trained image in batch is anchor sample Hn, select a positive sample being most difficult toWith one The negative sample being most difficult toAnd HnA triple is formed, the difficult requirement for dividing sample pair of selection isMaximum,It is minimum;
(2.3) triple pedestrian image is input to skeleton critical point detection network, detects 14 people respectively Body bone key point, including head, four limbs, the upper part of the body, the lower part of the body, and using 14 key points as coordinate pair human body attitude into Row correction;
(2.4) according to 14 skeleton key points, pedestrian's global image is divided into head, the upper part of the body, the lower part of the body Three pedestrian part ROI images obtain pedestrian's global image and three pedestrian topographies after a correction.
In the step (2.2), using the pre-training depth convolutional neural networks model in step (1.1), with anchor sample This HnIn identical pedestrian ID image select score it is minimum pedestrian image sample composition difficulty divide positive sample pair, with anchor sample HnNo Divide negative sample pair with selecting the pedestrian image sample of highest scoring to form difficulty in pedestrian's ID image.
3. the step is divided into following steps:
(3.1) it obtains by step (1.1) pre-training depth convolutional neural networks model and after 2. step is corrected Global and local image, and the softmax layer for removing pre-training depth convolutional neural networks model top layer connects entirely with one layer Layer;
(3.2) by the enhanced difficulty of data divide sample to, correction after global and local image be separately input to depth volume Product neural network model obtains pedestrian's global characteristics vector by the depth convolutional neural networks model that step (3.1) construct With pedestrian's local feature vectors.
4. the step is divided into following steps:
(4.1) by step 3. in pedestrian part and global characteristics vector be input to one layer of full articulamentum, it is special to carry out depth Sign fusion, obtains and exports fused pedestrian's feature vector;
(4.2) it by pedestrian's local feature vectors in fused pedestrian's feature vector and step (3.2), inputs respectively One layer of square of layer, square layer measure the similarity that hardly possible is divided between sample pair using squared euclidean distance.
6. the step is divided into following steps:
(6.1) inquiry is input to human body key point with target pedestrian image and posture correction network carries out human body attitude Correction;
(6.2) the pedestrian image input pedestrian after correcting human body attitude identifies network model again, and it is global special to obtain pedestrian Levy vector.
The beneficial effects of the present invention are: by extracting the global and local feature of pedestrian using depth convolutional neural networks, Depth integration is carried out to two kinds of features and obtains final pedestrian's characteristic present, then in depth convolutional neural networks training process In, by the way of transfer learning and then the acquisition better pedestrian of effect identifies network model again, finally pedestrian is identified again The feature that network model extracts has stronger resolution capability, to achieve the purpose that promote pedestrian's weight recognition accuracy.
Detailed description of the invention
Fig. 1 is flow chart of the invention;
Fig. 2 is the network structure of global characteristics of the embodiment of the present invention Yu local feature depth integration;
Fig. 3 is depth characteristic fusion and local feature learning mould of the embodiment of the present invention based on depth convolutional neural networks The network structure of type.
Specific embodiment
Be described further below technical solution of the present invention, but claimed range be not limited to it is described.
As shown in Figure 1, a kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic, including following step It is rapid:
1. pre-training again: the pre-training model based on ImageNet being identified in pedestrian and carries out pre-training in data, is obtained Pedestrian identifies pre-training network model again;
2. human body attitude correction and segmentation: choosing difficulty on personal data of being expert at collection and divide sample pair, and be input to skeleton pass In 14 key points that key point detection network detects, and human body attitude correction and pedestrian part ROI segmentation are carried out, obtained The enhanced difficulty of data divide sample to, correction after global and local image;
3. feature vector: dividing the global and local image after correction, the enhanced difficulty of data to sample pair, be input to row People identifies pre-training network model again, obtains pedestrian part and global characteristics vector;
4. depth characteristic merges: pedestrian part and global characteristics vector being carried out depth characteristic fusion, obtain final pedestrian Feature vector;
5. training pattern: by the way of transfer learning and step 4. in final pedestrian's feature vector, pedestrian is identified again Pre-training network model is finely adjusted, and is identified again in pedestrian and added adaptation layer in pre-training network model altogether, and pedestrian is obtained Network model is identified again;
6. test model: input inquiry pedestrian and target pedestrian image identify that network model extracts two using only pedestrian again A pedestrian's global characteristics vector with distinction;
7. recognition result: based on pedestrian's global characteristics vector in step 6., calculating inquiry pedestrian and target line personal data The similarity between any one image is concentrated, it is considered as identical pedestrian that wherein similarity is highest.
Pedestrian in training stage identifies the input of network model again, using triple pedestrian image.
1. the step is divided into following steps:
(1.1) the preparatory trained depth convolutional network model on ImageNet data set is obtained, and by it in pedestrian It identifies and is trained in data again;
(1.2) when pedestrian identifies pre-training depth convolutional neural networks model in data again, merely with sample mark Information is finely adjusted depth convolutional network model.
The step (1.2) is divided into following steps:
The ResNet50 network model of the pre-training on ImageNet data set is removed the full connection of top layer by (1.2.1) Layer, adds two layers of full articulamentum and one layer softmax layers after maximum pond layer;
The label information that (1.2.2) is marked using pedestrian image is finely adjusted the depth convolutional neural networks of building, The three first layers of constant depth convolutional neural networks in trim process;
(1.2.3) obtains the prediction probability of pedestrian's global image according to depth convolutional neural networks;
(1.2.4) defines the loss function in depth convolutional neural networks according to prediction probability.
The step 2. in global and local image after the correction that obtains, for difficult point of positive negative sample triad row people figure Pedestrian's global image and local ROI image after picture, human body attitude correction.
2. the step is divided into following steps:
(2.1) each is trained into batch, selects the pedestrian of P ID at random, each pedestrian select at random K open it is different Image, each batch contain P × K pedestrian images;
(2.2) each is trained image in batch is anchor sample Hn, select a positive sample being most difficult toWith one The negative sample being most difficult toAnd HnA triple is formed, the difficult requirement for dividing sample pair of selection isMaximum,It is minimum;
(2.3) triple pedestrian image is input to skeleton critical point detection network, detects 14 people respectively Body bone key point, including head, four limbs, the upper part of the body, the lower part of the body, and using 14 key points as coordinate pair human body attitude into Row correction;
(2.4) according to 14 skeleton key points, pedestrian's global image is divided into head, the upper part of the body, the lower part of the body Three pedestrian part ROI images obtain pedestrian's global image and three pedestrian topographies after a correction.
In the step (2.2), using the pre-training depth convolutional neural networks model in step (1.1), with anchor sample This HnIn identical pedestrian ID image select score it is minimum pedestrian image sample composition difficulty divide positive sample pair, with anchor sample HnNo Divide negative sample pair with selecting the pedestrian image sample of highest scoring to form difficulty in pedestrian's ID image.
3. the step is divided into following steps:
(3.1) it obtains by step (1.1) pre-training depth convolutional neural networks model and after 2. step is corrected Global and local image, and the softmax layer for removing pre-training depth convolutional neural networks model top layer connects entirely with one layer Layer;
(3.2) by the enhanced difficulty of data divide sample to, correction after global and local image be separately input to depth volume Product neural network model obtains pedestrian's global characteristics vector by the depth convolutional neural networks model that step (3.1) construct With pedestrian's local feature vectors.
4. the step is divided into following steps:
(4.1) by step 3. in pedestrian part and global characteristics vector be input to one layer of full articulamentum, it is special to carry out depth Sign fusion, obtains and exports fused pedestrian's feature vector;
(4.2) it by pedestrian's local feature vectors in fused pedestrian's feature vector and step (3.2), inputs respectively One layer of square of layer, square layer measure the similarity that hardly possible is divided between sample pair using squared euclidean distance.
6. the step is divided into following steps:
(6.1) inquiry is input to human body key point with target pedestrian image and posture correction network carries out human body attitude Correction;
(6.2) the pedestrian image input pedestrian after correcting human body attitude identifies network model again, and it is global special to obtain pedestrian Levy vector.
In conclusion advantage of the present invention using transfer learning and the adaptive learning of deep learning, using fusion pedestrian Image local feature and global feature obtain the network model that can pay close attention to pedestrian's local feature, improve what pedestrian identified again Accuracy rate.
Embodiment 1
As described above, a kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic, including following step It is rapid:
1. pre-training again: the pre-training model based on ImageNet being identified in pedestrian and carries out pre-training in data, is obtained Pedestrian identifies pre-training network model again;It is specifically divided into following steps:
(1.1) the preparatory trained depth convolutional network model on ImageNet data set is obtained, and by it in pedestrian It identifies and is trained in data again;
(1.2) when pedestrian identifies pre-training depth convolutional neural networks model in data again, merely with sample mark Information is finely adjusted depth convolutional network model;
The ResNet50 network model of the pre-training on ImageNet data set is removed the full connection of top layer by (1.2.1) Layer, adds two layers of full articulamentum and one layer softmax layers after maximum pond layer;
The label information that (1.2.2) is marked using pedestrian image is finely adjusted the depth convolutional neural networks of building, The three first layers of constant depth convolutional neural networks in trim process, because of the spy that the three first layers of depth convolutional neural networks are extracted Sign is usually texture, edge etc., and feature has certain versatility;
(1.2.3) obtains the prediction probability y of pedestrian's global image according to depth convolutional neural networksi, indicate are as follows:
Wherein, yiIndicate that sample x belongs to the probability of i-th of classification,It is normalization item, C is classification sum;
(1.2.4) defines the loss function L in depth convolutional neural networks according to prediction probabilityI, indicate are as follows:
Wherein, qjIndicate that label probability, C are classification sum.
2. human body attitude correction and segmentation: choosing difficulty on personal data of being expert at collection and divide sample pair, and be input to skeleton pass In 14 key points that key point detection network detects, and human body attitude correction and pedestrian part ROI segmentation are carried out, obtained The enhanced difficulty of data divide sample to, correction after global and local image;Wherein correct after global and local image, be Pedestrian's global image and local ROI image after the positive negative sample triad row people image of hardly possible point, human body attitude correction;Specific point For following steps:
(2.1) each is trained into batch, selects the pedestrian of P ID at random, each pedestrian select at random K open it is different Image, each batch contain P × K pedestrian images;
(2.2) each is trained image in batch is anchor sample Hn, select a positive sample being most difficult toWith one The negative sample being most difficult toAnd HnA triple is formed, the difficult requirement for dividing sample pair of selection isMaximum,It is minimum;
Specifically, using the pre-training depth convolutional neural networks model in step (1.1), with anchor sample HnIt is identical In pedestrian's ID image select score it is minimum pedestrian image sample composition difficulty divide positive sample pair, with anchor sample HnDifferent pedestrians The pedestrian image sample composition difficulty of highest scoring is selected to divide negative sample pair in ID image;
(2.3) triple pedestrian image is input to skeleton critical point detection network, detects 14 people respectively Body bone key point, including head, four limbs, the upper part of the body, the lower part of the body, and using 14 key points as coordinate pair human body attitude into Row correction;
(2.4) according to 14 skeleton key points, pedestrian's global image is divided into head, the upper part of the body, the lower part of the body Three pedestrian part ROI images obtain pedestrian's global image and three pedestrian topographies after a correction.
3. feature vector: dividing the global and local image after correction, the enhanced difficulty of data to sample pair, be input to row People identifies pre-training network model again, obtains pedestrian part and global characteristics vector;It is specifically divided into following steps:
(3.1) it obtains by step (1.1) pre-training depth convolutional neural networks model and after 2. step is corrected Global and local image, and the softmax layer for removing pre-training depth convolutional neural networks model top layer connects entirely with one layer Layer;
(3.2) by the enhanced difficulty of data divide sample to, correction after global and local image be separately input to depth volume Product neural network model obtains pedestrian's global characteristics vector A by the depth convolutional neural networks model that step (3.1) construct With pedestrian's local feature vectors B1、 B2、B3, wherein B1For head zone feature vector, B2For upper part of the body provincial characteristics vector, B3 For lower part of the body feature vector.
Further, it is each when the enhanced difficulty of data divides sample to depth convolutional neural networks are input to parallel respectively Depth convolutional neural networks model is propagated simultaneously, while shared weight.
4. depth characteristic merges: pedestrian part and global characteristics vector being carried out depth characteristic fusion, obtain final pedestrian Feature vector;It is specifically divided into following steps:
(4.1) by step 3. in pedestrian part and global characteristics vector be input to one layer of full articulamentum, it is special to carry out depth Sign fusion, obtains and exports fused pedestrian's feature vector C;
(4.2) by pedestrian's local feature vectors B in fused pedestrian's feature vector C and step (3.2)1、B2、B3, One layer of square of layer is inputted respectively, and square layer measures the similarity that hardly possible is divided between sample pair using squared euclidean distance, indicates are as follows:
Wherein, a is anchor sample, and p is to be most difficult to positive sample, and n is to be most difficult to negative sample, da,pFor hardly possible divide between positive sample pair away from From da,nFor hardly possible point negative sample to the distance between.
Preferably, it in order to enable depth convolutional neural networks to extract pedestrian's feature with higher distinguishing, fills simultaneously Point using pedestrian sample markup information, in training process, intersection entropy loss and triple has been used to lose, wherein fusion is complete The depth convolutional neural networks of office's feature and local feature use two kinds of loss functions in the training process, extract head, upper half Body, the lower part of the body depth convolutional neural networks only used triple loss function;
Further, the depth convolutional neural networks of amalgamation of global characteristics and local feature used intersection entropy loss and TriHard loss, indicates are as follows:
Wherein, the sample set with anchor sample a with identical ID is A, and the sample set for being left different ID is B.LthFor TriHard loss, LITo intersect entropy loss, LthIn α be think setting threshold parameter, qjIndicate label probability, C is class Not sum;
Further, extract the head of pedestrian, the upper part of the body, the lower part of the body depth convolutional neural networks used TriHard Loss function enables the depth convolutional neural networks for extracting pedestrian's global characteristics more to pay close attention to by sharing weight parameter Local feature with distinction, wherein loss function LthAre as follows:
Wherein, the sample set with anchor sample a with identical ID is A, and the sample set for being left different ID is B.LthIn α be Think the threshold parameter of setting;
The loss that depth integration feature and local feature will finally be extracted forms Total according to corresponding Weight Loss carries out backpropagation to overall network and updates network parameter;
5. training pattern: by the way of transfer learning and step 4. in final pedestrian's feature vector, pedestrian is identified again Pre-training network model is finely adjusted, and is identified again in pedestrian and added adaptation layer in pre-training network model altogether, and pedestrian is obtained Network model is identified again;In order to obtain better transfer learning effect, adaptation layer is added, so that the number of source domain and aiming field It is closer according to being distributed, so that pedestrian identifies that the effect of network model is more preferable again;
Specifically, the parameter learning that multicore MMD is measured adds the purpose degree in the training of depth convolutional neural networks The difference of source domain and aiming field is measured, wherein the multicore of multicore MMD measurement indicates are as follows:
Distribution distance between source domain and aiming field indicates are as follows:
Wherein, φ () is mapping, for by former variable mappings to reproducing kernel Hilbert space, H to indicate degree of a representation span From be mapped the data by φ () regeneration Hilbert space (RKHS) in measured;
The optimization aim of adaptation layer is made of loss function and adaptive loss, is indicated are as follows:
Wherein, Θ indicates all weights and offset parameter of network, is the target component of study, l1To l2It is Network adaptation Beginning and end layer, front without adaptation,naIndicate the set of all labeled data in source domain and aiming field, J () is loss function;
Specifically, selecting a row after the pre-training depth convolutional neural networks of acquisition are removed top layer softmax layers People's image inputs and uses the score of the several layers of convolutional layers in trained classifier calculated convolutional neural networks top, fixed score Network before highest one layer, one layer of highest scoring and network layer later are finely adjusted;
6. test model: identifying network model, input inquiry pedestrian and target pedestrian image again using only pedestrian, obtain Two have pedestrian's feature vectors of distinction, and it is complete to extract pedestrian respectively from two pedestrian's feature vectors with distinction Office's feature vector;It is specifically divided into following steps:
(6.1) inquiry is input to human body key point with target pedestrian image and posture correction network carries out human body attitude Correction;
(6.2) the pedestrian image input pedestrian after correcting human body attitude identifies network model again, and it is global special to obtain pedestrian Levy vector.
7. recognition result: based on pedestrian's global characteristics vector in step 6., calculating inquiry pedestrian and target line personal data The similarity between any one image is concentrated, it is considered as identical pedestrian that wherein similarity is highest.
Further, the pedestrian in the training stage identifies the input of network model again, using triple (triplet) row People's image.
Embodiment 2
As described above, a kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic, including following step It is rapid:
Pre-training model based on ImageNet is identified again in pedestrian and carries out pre-training in data, gone by step S1 People identifies pre-training network model again;
Step S11, obtain on ImageNet data set preparatory trained depth convolutional network model, and by its Pedestrian identifies again to be trained in data;
Step S12 only used sample when pedestrian identifies pre-training depth convolutional neural networks model in data again Markup information is finely adjusted network model;
The ResNet50 network model of the pre-training on ImageNet data set is removed connecting entirely for top layer by step S121 Layer is connect, two layers of full articulamentum and one layer softmax layers are added after maximum pond layer;
Further, two layers of addition full connection layer parameter is respectively 1 × 1 × 2048,1 × 1 × 751, input picture 224 × 224, when carrying out pre-training to ResNet50, optimization is iterated using gradient descent method, the number of iterations is set as 75 times, learning rate is initialized as 0.1, and weight pad value is set as 0.001 in optimization process, and each batch inputs 64 pedestrians Sample;
Step S122 is finely adjusted the depth convolutional neural networks of building using the label information that pedestrian image marks, The three first layers of fixed network in trim process.Because the feature that the three first layers of convolutional neural networks are extracted is usually texture, side Edge etc., feature have certain versatility;
Step S123 obtains the prediction probability of pedestrian's global image according to the convolutional neural networks, indicates are as follows:
Wherein, yiIndicate that sample x belongs to the probability of i-th of classification,It is normalization item, C is classification sum;
Preferably, when Market-1501 database is trained with testing, C=751.
Step S124 sets L for the loss function in the convolutional neural networks according to the prediction probabilityI, table It is shown as:
Wherein, qjIndicate label probability, C 751;
The input of step S2, training stage, network model use triple (triplet) pedestrian image.Firstly, being expert at Difficulty is chosen on personal data collection and divides sample pair, and is input to skeleton critical point detection network and detects 14 key points, so Human body attitude correction and pedestrian part ROI segmentation, global image and topography after being corrected are carried out afterwards;
Each is trained batch by step S21, selects the pedestrian of P ID at random, and each pedestrian selects K not at random Same image, each batch contain P × K pedestrian images;
Specifically, choosing the pedestrian of 6 ID in the present embodiment, each ID pedestrian selects 16 different images at random, Each batch contains 64 pedestrian images;
Step S22, it is anchor sample H that each, which is trained image in batch,n, select a positive sample being most difficult toWith one A negative sample being most difficult toAnd HnA triple is formed, the difficult requirement for dividing sample pair of selection isMaximum,It is minimum;
Further, using the convolutional neural networks model of step S1 pre-training, with anchor sample HnIdentical pedestrian ID figure As in selection score it is minimum pedestrian image sample composition difficulty divide positive sample pair, with anchor sample HnIn different pedestrian ID images The pedestrian image sample composition difficulty of selection highest scoring divides negative sample pair;
Triple pedestrian image is input to skeleton critical point detection network, detects 14 respectively by step S23 A skeleton key point, including head, four limbs, the upper part of the body, the lower part of the body, and using 14 key points as coordinate pair human body appearance State is corrected;
Step S24, according to 14 skeleton key points, by pedestrian's global image be divided into head, the upper part of the body, under Three pedestrian part ROI images of half body, and then obtain pedestrian's global image and three pedestrian topographies after a correction;
Step S3, will by human body attitude correction and the enhanced difficulty of data divide sample pair, be input to pre-training network into And obtain pedestrian part and global characteristics vector;
Step S31 obtains the depth convolutional neural networks model Jing Guo step S1 pre-training and obtains base by step S2 Divide pedestrian's global image and the topography of sample pair in hardly possible, and removes pre-training depth convolutional neural networks top layer Softmax layers and one layer of full articulamentum;
In the present embodiment, the full connection layer parameter of addition is respectively 1 × 1 × 751, input picture 224 × 224, right When ResNet50 carries out pre-training, optimization is iterated using gradient descent method, the number of iterations is set as 60 times, and first 20 times repeatedly Be initialized as 0.01 for learning rate, behind the learning rates of 40 iteration be 0.001, weight pad value is set as in optimization process 0.0001, each batch inputs 64 pedestrian samples;
Step S32, the difficulty that will acquire divide sample to being separately input to depth convolutional neural networks, including pedestrian's overall situation figure Picture and topography obtain pedestrian's global characteristics vector A and pedestrian office by the depth convolutional neural networks that step S31 is constructed Portion feature vector B1,B2,B3, wherein B1For head zone feature vector, B2For upper part of the body provincial characteristics vector, B3For the lower part of the body Feature vector;
Step S33, when difficulty divides sample to depth convolutional neural networks are input to parallel respectively, each depth convolutional Neural Network model is propagated simultaneously, while shared weight;
Pedestrian's local feature vectors of acquisition are carried out depth characteristic with global characteristics vector and merged, obtained most by step S4 Whole pedestrian's feature vector;
Step S41 obtains pedestrian global characteristics vector A and pedestrian's local feature vectors B by step S31,B2,B3, so Pedestrian's global characteristics vector sum local feature vectors are input to one layer of full articulamentum afterwards and carry out depth characteristic fusion, are carried out defeated Fused pedestrian's feature vector C out, as shown in Figure 2;
Step S42, by fused pedestrian's feature vector C and local feature vectors B1,B2,B3, one layer is inputted respectively puts down Square layer, square layer measure the similarity that hardly possible is divided between sample pair using squared euclidean distance, indicate are as follows:
Wherein, a is anchor sample, and p is to be most difficult to positive sample, and n is to be most difficult to negative sample, da,pFor hardly possible divide between positive sample pair away from From da,nFor hardly possible point negative sample to the distance between;
Step S43, in order to enable depth convolutional neural networks to extract pedestrian's feature with higher distinguishing, simultaneously The markup information of pedestrian sample is made full use of, has used intersection entropy loss and triple to lose during training network, wherein melting The depth convolutional neural networks for closing global characteristics and local feature use two kinds of loss functions in the training process, extract head, Above the waist, the depth convolutional neural networks of lower part of the body feature only used triple loss function;
The depth convolutional neural networks of step S431, amalgamation of global characteristics and local feature used intersection entropy loss and TriHard loss, indicates are as follows:
Wherein, the sample set with anchor sample a with identical ID is A, and the sample set for being left different ID is B.LthFor TriHard loss, LITo intersect entropy loss, LthIn α be think setting threshold parameter, qjIndicate that label probability, C are 751;
Step S432, extract the head of pedestrian, the upper part of the body, the lower part of the body depth convolutional neural networks used TriHard Loss function enables the depth convolutional neural networks for extracting pedestrian's global characteristics more to pay close attention to by sharing weight parameter Local feature with distinction, wherein loss function LthAre as follows:
Wherein, the sample set with anchor sample a with identical ID is A, and the sample set for being left different ID is B.LthIn α be Think the threshold parameter of setting;
Step S433 will finally extract the loss of depth integration feature and local feature, according to corresponding Weight, Total loss is formed, backpropagation is carried out to overall network and updates network parameter, as shown in Figure 3;
Further, each network losses according to corresponding Weight combination are as follows:
Wherein, pcFor the intersection entropy loss for extracting depth integration feature, ptThe depth respectively extracted The TriHard loss of fusion feature, head feature, upper part of the body feature, lower part of the body feature, weight factor α1、α2、α3、α4、α5Point It is not set as 0.2,0.2,0.2,0.2,0.2;
Step S5, during the entire process of pedestrian identifies network model training again, to pre- instruction by the way of transfer learning Practice network to be finely adjusted, by adding adaptation layer in a network;
Step S51 after the pre-training convolutional neural networks of acquisition are removed top layer softmax layers, selects pedestrian's figure It is fixed as being input in network and using the score of the several layers of convolutional layers in trained classifier calculated convolutional neural networks top Network before one layer of highest scoring, one layer of highest scoring and network layer later are finely adjusted;
Step S52 adds adaptation layer, purpose to obtain better transfer learning effect in the network layer of fine tuning So that source domain and the data distribution of aiming field are closer, so that pedestrian identifies that network obtains better effect again.By multicore MMD The parameter learning of measurement adds in the training of depth convolutional neural networks, and purpose measures the difference of source domain and aiming field, wherein The multicore of multicore MMD measurement indicates are as follows:
Distribution distance between source domain and aiming field indicates are as follows:
Wherein, φ () is mapping, for by former variable mappings to reproducing kernel Hilbert space, H to indicate degree of a representation span From be mapped the data by φ () regeneration Hilbert space (RKHS) in measured;
The optimization aim of step S53, adaptation layer are made of loss function and adaptive loss, are indicated are as follows:
Wherein, Θ indicates all weights and offset parameter of network, is the target component of study, l1To l2It is Network adaptation Beginning and end layer, front without adaptation,naIndicate the set of all labeled data in source domain and aiming field, J () is loss function;
Step S6 is used only trained network model and extracts pedestrian's global characteristics vector in test phase.Based on upper The trained network model of step, input inquiry pedestrian and target pedestrian image are stated, two rows with higher distinction are obtained People's feature vector;
Step S61, pedestrian's global characteristics that the depth convolutional neural networks after above-mentioned steps training extract are to measurer There is higher distinction, therefore the test phase of model only extracts pedestrian's global characteristics vector;
Inquiry is input to human body key point with target pedestrian image and posture correction network carries out human body appearance by step S62 State correction;
Step S7 is based on pedestrian's global characteristics vector, calculates in inquiry pedestrian image and target pedestrian image data set and appoints The similarity anticipated between an image, it is considered as identical pedestrian that wherein similarity is highest, carries out obtaining pedestrian identifying again As a result.
Specifically, the embodiment of the present invention is using Market-1501 pedestrian's database as training set and test set, proposition The rank-1 of recognition methods reaches 85%, mAp and reaches 60% again by a kind of pedestrian merged based on transfer learning and depth characteristic. Pedestrian of the present invention again recognition methods using transfer learning and in training using depth integration pedestrian global characteristics and local feature Method greatly improves the accuracy rate that pedestrian identifies again, it can be seen that the validity of the method for the present invention.
In conclusion the net of the present invention training depth integration pedestrian part and global characteristics by the way of transfer learning Network model divides negative sample pair to difficulty with difficulty point positive sample is chosen on pedestrian's data set, and be input to human body in the training stage Bone critical point detection network detects 14 key points, is corrected and is divided to pedestrian's posture with 14 key points It is cut into three pedestrian image subregions;It will be defeated to dividing pedestrian's training image of negative sample pair to distinguish with hardly possible comprising difficult point positive sample Enter to pre-training network, wherein inputting every sample is extended to pedestrian's general image and three pedestrian's sub-district area images, obtains Obtain pedestrian part and global characteristics vector;Three pedestrian's local feature vectors are input to one layer of full articulamentum and pedestrian is global Feature vector is merged, and pedestrian's feature vector of depth characteristic fusion is obtained;Identify that network training process uses again in pedestrian The mode of transfer learning is finely adjusted pre-training network, pre-training network top add adaptation layer come complete source domain and Aiming field it is adaptive so that source domain and the data distribution of aiming field are more nearly, so that pedestrian identifies the effect of network more again It is good;Global characteristics network model, input inquiry pedestrian image and target pedestrian image is used only to global characteristics in test phase Network model is extracted, two global characteristics vectors is obtained, and then calculate the similarity of inquiry pedestrian and target pedestrian, is known Other result.

Claims (10)

1. a kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic, it is characterised in that: the following steps are included:
1. pre-training again: the pre-training model based on ImageNet being identified in pedestrian and carries out pre-training in data, obtains pedestrian Pre-training network model is identified again;
2. human body attitude correction and segmentation: choosing difficulty on personal data of being expert at collection and divide sample pair, and be input to skeleton key point In 14 key points that detection network detects, and human body attitude correction and pedestrian part ROI segmentation are carried out, obtains data and increase Difficulty after strong divides sample to the global and local image after, correction;
3. feature vector again: dividing the global and local image after correction, the enhanced difficulty of data to sample pair, be input to pedestrian and know Other pre-training network model obtains pedestrian part and global characteristics vector;
4. depth characteristic merges: pedestrian part and global characteristics vector being carried out depth characteristic fusion, obtain final pedestrian's feature Vector;
5. training pattern: by the way of transfer learning and step 4. in final pedestrian's feature vector, pedestrian identify again and is instructed in advance Practice network model to be finely adjusted, and identified again in pedestrian and add adaptation layer in pre-training network model altogether, obtains pedestrian and know again Other network model;
6. test model: input inquiry pedestrian and target pedestrian image identify that network model extracts two tools using only pedestrian again There is pedestrian's global characteristics vector of distinction;
7. recognition result: based on pedestrian's global characteristics vector in step 6., calculating inquiry pedestrian and target line personal data is concentrated Similarity between any one image, it is considered as identical pedestrian that wherein similarity is highest.
2. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, it is characterised in that: Pedestrian in training stage identifies the input of network model again, using triple pedestrian image.
3. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, it is characterised in that: 1. the step is divided into following steps:
(1.1) the preparatory trained depth convolutional network model on ImageNet data set is obtained, and it is known again in pedestrian It is trained in other data;
(1.2) when pedestrian identifies pre-training depth convolutional neural networks model in data again, merely with sample markup information Depth convolutional network model is finely adjusted.
4. the pedestrian's recognition methods again merged as claimed in claim 3 based on transfer learning and depth characteristic, it is characterised in that: The step (1.2) is divided into following steps:
The ResNet50 network model of the pre-training on ImageNet data set is removed the full articulamentum of top layer by (1.2.1), Two layers of full articulamentum and one layer softmax layers are added after maximum pond layer;
The label information that (1.2.2) is marked using pedestrian image is finely adjusted the depth convolutional neural networks of building, is finely tuning The three first layers of constant depth convolutional neural networks in the process;
(1.2.3) obtains the prediction probability of pedestrian's global image according to depth convolutional neural networks;
(1.2.4) defines the loss function in depth convolutional neural networks according to prediction probability.
5. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, it is characterised in that: The step 2. in global and local image after the correction that obtains, for difficult point of positive negative sample triad row people image, human body appearance Pedestrian's global image and local ROI image after state correction.
6. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, it is characterised in that: 2. the step is divided into following steps:
(2.1) each being trained into batch, selects the pedestrian of P ID at random, each pedestrian selects K different images at random, Each batch contains P × K pedestrian images;
(2.2) each is trained image in batch is anchor sample Hn, select the positive sample being most difficult to pointIt is most difficult to one The negative sample dividedAnd HnA triple is formed, the difficult requirement for dividing sample pair of selection isMaximum, It is minimum;
(2.3) triple pedestrian image is input to skeleton critical point detection network, detects 14 human body bones respectively Bone key point, including head, four limbs, the upper part of the body, the lower part of the body, and rectified using 14 key points as coordinate pair human body attitude Just;
(2.4) according to 14 skeleton key points, pedestrian's global image is divided into head, the upper part of the body, the lower part of the body three Pedestrian part ROI image obtains pedestrian's global image and three pedestrian topographies after a correction.
7. the pedestrian's recognition methods again merged as claimed in claim 6 based on transfer learning and depth characteristic, it is characterised in that: In the step (2.2), using the pre-training depth convolutional neural networks model in step (1.1), with anchor sample HnIt is identical In pedestrian's ID image select score it is minimum pedestrian image sample composition difficulty divide positive sample pair, with anchor sample HnDifferent pedestrian ID The pedestrian image sample composition difficulty of highest scoring is selected to divide negative sample pair in image.
8. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, it is characterised in that: 3. the step is divided into following steps:
(3.1) it obtains by step (1.1) pre-training depth convolutional neural networks model and the overall situation after 2. step is corrected And topography, and remove the softmax layer and one layer of full articulamentum of pre-training depth convolutional neural networks model top layer;
(3.2) sample is divided to be separately input to depth convolution mind to the global and local image after, correction the enhanced difficulty of data Pedestrian's global characteristics vector sum pedestrian is obtained by the depth convolutional neural networks model that step (3.1) construct through network model Local feature vectors.
9. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, it is characterised in that: 4. the step is divided into following steps:
(4.1) by step 3. in pedestrian part and global characteristics vector be input to one layer of full articulamentum, carry out depth characteristic and melt It closes, obtains and export fused pedestrian's feature vector;
(4.2) by pedestrian's local feature vectors in fused pedestrian's feature vector and step (3.2), one layer is inputted respectively and is put down Square layer, square layer measure the similarity that hardly possible is divided between sample pair using squared euclidean distance.
10. the pedestrian's recognition methods again merged as described in claim 1 based on transfer learning and depth characteristic, feature are existed In: 6. the step is divided into following steps:
(6.1) inquiry is input to human body key point with target pedestrian image and posture correction network carries out human body attitude correction;
(6.2) the pedestrian image input pedestrian after correcting human body attitude identifies network model again, obtain pedestrian's global characteristics to Amount.
CN201910329733.2A 2019-04-23 2019-04-23 Pedestrian re-recognition method based on transfer learning and depth feature fusion Active CN110163110B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910329733.2A CN110163110B (en) 2019-04-23 2019-04-23 Pedestrian re-recognition method based on transfer learning and depth feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910329733.2A CN110163110B (en) 2019-04-23 2019-04-23 Pedestrian re-recognition method based on transfer learning and depth feature fusion

Publications (2)

Publication Number Publication Date
CN110163110A true CN110163110A (en) 2019-08-23
CN110163110B CN110163110B (en) 2023-06-06

Family

ID=67639792

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910329733.2A Active CN110163110B (en) 2019-04-23 2019-04-23 Pedestrian re-recognition method based on transfer learning and depth feature fusion

Country Status (1)

Country Link
CN (1) CN110163110B (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110533184A (en) * 2019-08-31 2019-12-03 南京人工智能高等研究院有限公司 A kind of training method and device of network model
CN110555420A (en) * 2019-09-09 2019-12-10 电子科技大学 fusion model network and method based on pedestrian regional feature extraction and re-identification
CN110569779A (en) * 2019-08-28 2019-12-13 西北工业大学 Pedestrian attribute identification method based on pedestrian local and overall attribute joint learning
CN110674881A (en) * 2019-09-27 2020-01-10 长城计算机软件与系统有限公司 Trademark image retrieval model training method, system, storage medium and computer equipment
CN110688976A (en) * 2019-10-09 2020-01-14 创新奇智(北京)科技有限公司 Store comparison method based on image identification
CN110705499A (en) * 2019-10-12 2020-01-17 成都考拉悠然科技有限公司 Crowd counting method based on transfer learning
CN110795580A (en) * 2019-10-23 2020-02-14 武汉理工大学 Vehicle weight recognition method based on space-time constraint model optimization
CN111126599A (en) * 2019-12-20 2020-05-08 复旦大学 Neural network weight initialization method based on transfer learning
CN111126275A (en) * 2019-12-24 2020-05-08 广东省智能制造研究所 Pedestrian re-identification method and device based on multi-granularity feature fusion
CN111126198A (en) * 2019-12-11 2020-05-08 中山大学 Pedestrian re-identification method based on deep representation learning and dynamic matching
CN111160295A (en) * 2019-12-31 2020-05-15 广州视声智能科技有限公司 Video pedestrian re-identification method based on region guidance and space-time attention
CN111274922A (en) * 2020-01-17 2020-06-12 山东师范大学 Pedestrian re-identification method and system based on multi-level deep learning network
CN111401265A (en) * 2020-03-19 2020-07-10 重庆紫光华山智安科技有限公司 Pedestrian re-identification method and device, electronic equipment and computer-readable storage medium
CN111428675A (en) * 2020-04-02 2020-07-17 南开大学 Pedestrian re-recognition method integrated with pedestrian posture features
CN111428650A (en) * 2020-03-26 2020-07-17 北京工业大学 Pedestrian re-identification method based on SP-PGGAN style migration
CN111539257A (en) * 2020-03-31 2020-08-14 苏州科达科技股份有限公司 Personnel re-identification method, device and storage medium
CN111582154A (en) * 2020-05-07 2020-08-25 浙江工商大学 Pedestrian re-identification method based on multitask skeleton posture division component
CN111696056A (en) * 2020-05-25 2020-09-22 五邑大学 Digital archive image correction method based on multi-task transfer learning
CN111881842A (en) * 2020-07-30 2020-11-03 深圳力维智联技术有限公司 Pedestrian re-identification method and device, electronic equipment and storage medium
CN111967389A (en) * 2020-08-18 2020-11-20 厦门理工学院 Face attribute recognition method and system based on deep double-path learning network
CN112132200A (en) * 2020-09-17 2020-12-25 山东大学 Lithology identification method and system based on multi-dimensional rock image deep learning
CN112183438A (en) * 2020-10-13 2021-01-05 深圳龙岗智能视听研究院 Image identification method for illegal behaviors based on small sample learning neural network
CN112784630A (en) * 2019-11-06 2021-05-11 广东毓秀科技有限公司 Method for re-identifying pedestrians based on local features of physical segmentation
CN112990424A (en) * 2019-12-17 2021-06-18 杭州海康威视数字技术股份有限公司 Method and device for training neural network model
CN112989911A (en) * 2020-12-10 2021-06-18 奥比中光科技集团股份有限公司 Pedestrian re-identification method and system
CN112990144A (en) * 2021-04-30 2021-06-18 德鲁动力科技(成都)有限公司 Data enhancement method and system for pedestrian re-identification
CN113221770A (en) * 2021-05-18 2021-08-06 青岛根尖智能科技有限公司 Cross-domain pedestrian re-identification method and system based on multi-feature hybrid learning
CN113379627A (en) * 2021-06-07 2021-09-10 北京百度网讯科技有限公司 Training method of image enhancement model and method for enhancing image
CN113378729A (en) * 2021-06-16 2021-09-10 西安理工大学 Pose embedding-based multi-scale convolution feature fusion pedestrian re-identification method
CN113377991A (en) * 2021-06-10 2021-09-10 电子科技大学 Image retrieval method based on most difficult positive and negative samples
CN113591864A (en) * 2021-07-28 2021-11-02 北京百度网讯科技有限公司 Training method, device and system for text recognition model framework
CN114863138A (en) * 2022-07-08 2022-08-05 腾讯科技(深圳)有限公司 Image processing method, image processing apparatus, storage medium, and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107766791A (en) * 2017-09-06 2018-03-06 北京大学 A kind of pedestrian based on global characteristics and coarseness local feature recognition methods and device again
CN107832672A (en) * 2017-10-12 2018-03-23 北京航空航天大学 A kind of pedestrian's recognition methods again that more loss functions are designed using attitude information
US20180114055A1 (en) * 2016-10-25 2018-04-26 VMAXX. Inc. Point to Set Similarity Comparison and Deep Feature Learning for Visual Recognition
CN108334849A (en) * 2018-01-31 2018-07-27 中山大学 A kind of recognition methods again of the pedestrian based on Riemann manifold
CN109002761A (en) * 2018-06-13 2018-12-14 中山大学新华学院 A kind of pedestrian's weight identification monitoring system based on depth convolutional neural networks
CN109271870A (en) * 2018-08-21 2019-01-25 平安科技(深圳)有限公司 Pedestrian recognition methods, device, computer equipment and storage medium again
CN109446898A (en) * 2018-09-20 2019-03-08 暨南大学 A kind of recognition methods again of the pedestrian based on transfer learning and Fusion Features

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180114055A1 (en) * 2016-10-25 2018-04-26 VMAXX. Inc. Point to Set Similarity Comparison and Deep Feature Learning for Visual Recognition
CN107766791A (en) * 2017-09-06 2018-03-06 北京大学 A kind of pedestrian based on global characteristics and coarseness local feature recognition methods and device again
CN107832672A (en) * 2017-10-12 2018-03-23 北京航空航天大学 A kind of pedestrian's recognition methods again that more loss functions are designed using attitude information
CN108334849A (en) * 2018-01-31 2018-07-27 中山大学 A kind of recognition methods again of the pedestrian based on Riemann manifold
CN109002761A (en) * 2018-06-13 2018-12-14 中山大学新华学院 A kind of pedestrian's weight identification monitoring system based on depth convolutional neural networks
CN109271870A (en) * 2018-08-21 2019-01-25 平安科技(深圳)有限公司 Pedestrian recognition methods, device, computer equipment and storage medium again
CN109446898A (en) * 2018-09-20 2019-03-08 暨南大学 A kind of recognition methods again of the pedestrian based on transfer learning and Fusion Features

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LONGHUI WEI: ""GLAD: Global-Local-Alignment Descriptor for Scalable Person Re-Identification"", 《IEEE TRANSACTIONS ON MULTIMEDIA》 *
徐梦洋: ""基于深度学习的行人再识别研究综述"", 《计算机科学》 *

Cited By (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110569779A (en) * 2019-08-28 2019-12-13 西北工业大学 Pedestrian attribute identification method based on pedestrian local and overall attribute joint learning
CN110569779B (en) * 2019-08-28 2022-10-04 西北工业大学 Pedestrian attribute identification method based on pedestrian local and overall attribute joint learning
CN110533184A (en) * 2019-08-31 2019-12-03 南京人工智能高等研究院有限公司 A kind of training method and device of network model
CN110555420A (en) * 2019-09-09 2019-12-10 电子科技大学 fusion model network and method based on pedestrian regional feature extraction and re-identification
CN110555420B (en) * 2019-09-09 2022-04-12 电子科技大学 Fusion model network and method based on pedestrian regional feature extraction and re-identification
CN110674881A (en) * 2019-09-27 2020-01-10 长城计算机软件与系统有限公司 Trademark image retrieval model training method, system, storage medium and computer equipment
CN110674881B (en) * 2019-09-27 2022-02-11 长城计算机软件与系统有限公司 Trademark image retrieval model training method, system, storage medium and computer equipment
CN110688976A (en) * 2019-10-09 2020-01-14 创新奇智(北京)科技有限公司 Store comparison method based on image identification
CN110705499A (en) * 2019-10-12 2020-01-17 成都考拉悠然科技有限公司 Crowd counting method based on transfer learning
CN110705499B (en) * 2019-10-12 2020-06-02 成都考拉悠然科技有限公司 Crowd counting method based on transfer learning
CN110795580A (en) * 2019-10-23 2020-02-14 武汉理工大学 Vehicle weight recognition method based on space-time constraint model optimization
CN110795580B (en) * 2019-10-23 2023-12-08 武汉理工大学 Vehicle weight identification method based on space-time constraint model optimization
CN112784630A (en) * 2019-11-06 2021-05-11 广东毓秀科技有限公司 Method for re-identifying pedestrians based on local features of physical segmentation
CN111126198A (en) * 2019-12-11 2020-05-08 中山大学 Pedestrian re-identification method based on deep representation learning and dynamic matching
CN111126198B (en) * 2019-12-11 2023-05-09 中山大学 Pedestrian re-identification method based on deep representation learning and dynamic matching
CN112990424A (en) * 2019-12-17 2021-06-18 杭州海康威视数字技术股份有限公司 Method and device for training neural network model
CN111126599B (en) * 2019-12-20 2023-09-05 复旦大学 Neural network weight initialization method based on transfer learning
CN111126599A (en) * 2019-12-20 2020-05-08 复旦大学 Neural network weight initialization method based on transfer learning
CN111126275B (en) * 2019-12-24 2023-05-05 广东省智能制造研究所 Pedestrian re-identification method and device based on multi-granularity feature fusion
CN111126275A (en) * 2019-12-24 2020-05-08 广东省智能制造研究所 Pedestrian re-identification method and device based on multi-granularity feature fusion
CN111160295A (en) * 2019-12-31 2020-05-15 广州视声智能科技有限公司 Video pedestrian re-identification method based on region guidance and space-time attention
CN111160295B (en) * 2019-12-31 2023-05-12 广州视声智能科技有限公司 Video pedestrian re-recognition method based on region guidance and space-time attention
CN111274922A (en) * 2020-01-17 2020-06-12 山东师范大学 Pedestrian re-identification method and system based on multi-level deep learning network
CN111274922B (en) * 2020-01-17 2022-11-29 山东师范大学 Pedestrian re-identification method and system based on multi-level deep learning network
CN111401265A (en) * 2020-03-19 2020-07-10 重庆紫光华山智安科技有限公司 Pedestrian re-identification method and device, electronic equipment and computer-readable storage medium
CN111428650A (en) * 2020-03-26 2020-07-17 北京工业大学 Pedestrian re-identification method based on SP-PGGAN style migration
CN111428650B (en) * 2020-03-26 2024-04-02 北京工业大学 Pedestrian re-recognition method based on SP-PGGAN style migration
CN111539257A (en) * 2020-03-31 2020-08-14 苏州科达科技股份有限公司 Personnel re-identification method, device and storage medium
CN111539257B (en) * 2020-03-31 2022-07-26 苏州科达科技股份有限公司 Person re-identification method, device and storage medium
CN111428675A (en) * 2020-04-02 2020-07-17 南开大学 Pedestrian re-recognition method integrated with pedestrian posture features
CN111582154A (en) * 2020-05-07 2020-08-25 浙江工商大学 Pedestrian re-identification method based on multitask skeleton posture division component
CN111696056B (en) * 2020-05-25 2023-05-02 五邑大学 Digital archive image correction method based on multitasking transfer learning
CN111696056A (en) * 2020-05-25 2020-09-22 五邑大学 Digital archive image correction method based on multi-task transfer learning
CN111881842A (en) * 2020-07-30 2020-11-03 深圳力维智联技术有限公司 Pedestrian re-identification method and device, electronic equipment and storage medium
CN111967389B (en) * 2020-08-18 2022-02-18 厦门理工学院 Face attribute recognition method and system based on deep double-path learning network
CN111967389A (en) * 2020-08-18 2020-11-20 厦门理工学院 Face attribute recognition method and system based on deep double-path learning network
CN112132200A (en) * 2020-09-17 2020-12-25 山东大学 Lithology identification method and system based on multi-dimensional rock image deep learning
CN112183438A (en) * 2020-10-13 2021-01-05 深圳龙岗智能视听研究院 Image identification method for illegal behaviors based on small sample learning neural network
CN112183438B (en) * 2020-10-13 2022-11-04 深圳龙岗智能视听研究院 Image identification method for illegal behaviors based on small sample learning neural network
CN112989911A (en) * 2020-12-10 2021-06-18 奥比中光科技集团股份有限公司 Pedestrian re-identification method and system
CN112990144A (en) * 2021-04-30 2021-06-18 德鲁动力科技(成都)有限公司 Data enhancement method and system for pedestrian re-identification
CN113221770A (en) * 2021-05-18 2021-08-06 青岛根尖智能科技有限公司 Cross-domain pedestrian re-identification method and system based on multi-feature hybrid learning
CN113379627B (en) * 2021-06-07 2023-06-27 北京百度网讯科技有限公司 Training method of image enhancement model and method for enhancing image
CN113379627A (en) * 2021-06-07 2021-09-10 北京百度网讯科技有限公司 Training method of image enhancement model and method for enhancing image
CN113377991A (en) * 2021-06-10 2021-09-10 电子科技大学 Image retrieval method based on most difficult positive and negative samples
CN113377991B (en) * 2021-06-10 2022-04-15 电子科技大学 Image retrieval method based on most difficult positive and negative samples
CN113378729A (en) * 2021-06-16 2021-09-10 西安理工大学 Pose embedding-based multi-scale convolution feature fusion pedestrian re-identification method
WO2023005253A1 (en) * 2021-07-28 2023-02-02 北京百度网讯科技有限公司 Method, apparatus and system for training text recognition model framework
CN113591864A (en) * 2021-07-28 2021-11-02 北京百度网讯科技有限公司 Training method, device and system for text recognition model framework
CN114863138B (en) * 2022-07-08 2022-09-06 腾讯科技(深圳)有限公司 Image processing method, device, storage medium and equipment
CN114863138A (en) * 2022-07-08 2022-08-05 腾讯科技(深圳)有限公司 Image processing method, image processing apparatus, storage medium, and device

Also Published As

Publication number Publication date
CN110163110B (en) 2023-06-06

Similar Documents

Publication Publication Date Title
CN110163110A (en) A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic
CN107145842B (en) Face recognition method combining LBP characteristic graph and convolutional neural network
CN106326886B (en) Finger vein image quality appraisal procedure based on convolutional neural networks
CN108256421A (en) A kind of dynamic gesture sequence real-time identification method, system and device
CN107463920A (en) A kind of face identification method for eliminating partial occlusion thing and influenceing
CN108229444A (en) A kind of pedestrian's recognition methods again based on whole and local depth characteristic fusion
CN104850865B (en) A kind of Real Time Compression tracking of multiple features transfer learning
CN109101865A (en) A kind of recognition methods again of the pedestrian based on deep learning
CN109447115A (en) Zero sample classification method of fine granularity based on multilayer semanteme supervised attention model
CN106529499A (en) Fourier descriptor and gait energy image fusion feature-based gait identification method
CN104021559B (en) Image registration method based on mutual information and Harris corner point detection
CN109598268A (en) A kind of RGB-D well-marked target detection method based on single flow depth degree network
CN105869178A (en) Method for unsupervised segmentation of complex targets from dynamic scene based on multi-scale combination feature convex optimization
CN103440510A (en) Method for positioning characteristic points in facial image
CN105536205A (en) Upper limb training system based on monocular video human body action sensing
CN102054170B (en) Visual tracking method based on minimized upper bound error
CN108596211A (en) It is a kind of that pedestrian's recognition methods again is blocked based on focusing study and depth e-learning
CN110033007A (en) Attribute recognition approach is worn clothes based on the pedestrian of depth attitude prediction and multiple features fusion
CN104392223A (en) Method for recognizing human postures in two-dimensional video images
CN108537143B (en) A kind of face identification method and system based on key area aspect ratio pair
CN104881852B (en) Image partition method based on immune clone and fuzzy kernel clustering
CN106599785A (en) Method and device for building human body 3D feature identity information database
CN104751111A (en) Method and system for recognizing human action in video
CN109902585A (en) A kind of three modality fusion recognition methods of finger based on graph model
CN105069745A (en) face-changing system based on common image sensor and enhanced augmented reality technology and method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant