CN107729805A - The neutral net identified again for pedestrian and the pedestrian based on deep learning recognizer again - Google Patents

The neutral net identified again for pedestrian and the pedestrian based on deep learning recognizer again Download PDF

Info

Publication number
CN107729805A
CN107729805A CN201710780179.0A CN201710780179A CN107729805A CN 107729805 A CN107729805 A CN 107729805A CN 201710780179 A CN201710780179 A CN 201710780179A CN 107729805 A CN107729805 A CN 107729805A
Authority
CN
China
Prior art keywords
pedestrian
neural network
identification
image
cnn
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710780179.0A
Other languages
Chinese (zh)
Other versions
CN107729805B (en
Inventor
张史梁
田奇
高文
李佳宁
苏驰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Peking University
Original Assignee
Peking University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Peking University filed Critical Peking University
Priority to CN201710780179.0A priority Critical patent/CN107729805B/en
Publication of CN107729805A publication Critical patent/CN107729805A/en
Application granted granted Critical
Publication of CN107729805B publication Critical patent/CN107729805B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/14Transformations for image registration, e.g. adjusting or mapping for alignment of images
    • G06T3/147Transformations for image registration, e.g. adjusting or mapping for alignment of images using affine transformations

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses the neutral net identified again for pedestrian and the recognizer again of the pedestrian based on deep learning.The neutral net includes:Inputted using the original whole body images of pedestrian as first and export the first nerves network of the first identification feature;Affine transformation image using the human body image extracted from the original whole body images of pedestrian inputs as second and exports the nervus opticus network of the second identification feature, wherein, human body comprises at least head, trunk and four limbs, and the first identification feature and the second identification feature are combined as total identification feature.Pedestrian's characteristic matching ability with more robust, so as to improve correct recognition rata and/or reduce misclassification rate.

Description

Neural network for pedestrian re-recognition and pedestrian re-recognition algorithm based on deep learning
Technical Field
The present invention relates to the field of convolutional neural networks and image recognition. More particularly, the present invention relates to a neural network for pedestrian re-recognition and a pedestrian re-recognition algorithm based on deep learning.
Background
With the popularization of video monitoring technology, the role of pedestrian re-identification technology becomes increasingly important, and the pedestrian re-identification technology can help people to automatically complete the task of searching for specific personnel from massive image or video data.
Feature extraction and feature matching using convolutional neural networks are two important components of pedestrian re-identification technology. However, there are usually large posture changes and complicated viewing angle changes between different images of the pedestrian acquired by different cameras, which both greatly increase the difficulty of matching the pedestrian with the pedestrian re-recognition algorithm.
At least in order to solve the above technical problems, a new pedestrian re-identification algorithm capable of adapting to the posture change needs to be provided, so that the algorithm can better adapt to the posture change of the pedestrian, extract more robust features, and thus improve the final correct identification rate and/or reduce the false identification rate.
Disclosure of Invention
The purpose of the invention is realized by the following technical scheme.
A neural network for pedestrian re-identification, comprising:
a first neural network that uses an original whole-body image of a pedestrian as a first input and outputs a first recognition feature;
a second neural network that uses an affine transformation image of a human body part image extracted from an original whole-body image of a pedestrian as a second input and outputs a second recognition feature,
the human body part at least comprises a head, a trunk and four limbs, and the first identification feature and the second identification feature are combined into a total identification feature.
The neural network for pedestrian re-recognition according to the present invention further comprises:
a feature embedding sub-neural network (FEN) includes a Pose Transformation Network (PTN) for applying an affine transformation to each body part in a body part map to obtain a more robust affine transformation image of the body part map image.
The neural network for pedestrian re-recognition according to the present invention further comprises:
a feature weighting sub-neural network (FWN) for weighting and biasing the second identifying features output by the second neural network to combine the first identifying features and the second identifying features into a total identifying feature.
According to the neural network for pedestrian re-identification of the present invention, the first neural network includes a first Convolutional Neural Network (CNN) and a second CNN, and the second neural network includes a third CNN and a fourth CNN, wherein the first CNN and the third CNN share a weight.
According to the neural network for pedestrian re-recognition of the present invention, the second CNN and the fourth CNN adopt independent weights.
The neural network for pedestrian re-identification according to the invention is characterized in that the first neural network and the second neural network each further comprise a Convolution (CONV) layer and a Global Average Pooling (GAP) layer at the respective output.
The pedestrian re-identification algorithm based on deep learning comprises the following steps:
constructing a neural network for pedestrian re-identification according to the above;
training the neural network for pedestrian re-recognition using a pedestrian re-recognition dataset;
pedestrian re-recognition is performed using a trained neural network for pedestrian re-recognition.
The invention has the advantages that: the pedestrian feature matching method has more robust pedestrian feature matching capability, so that the correct recognition rate can be improved and/or the false recognition rate can be reduced.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the specific embodiments. The drawings are only for purposes of illustrating the particular embodiments and are not to be construed as limiting the invention. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
fig. 1 shows a conceptual schematic diagram of a first neural network for pedestrian re-identification according to an embodiment of the present invention.
Fig. 2 shows an overall schematic diagram of a second type of neural network for pedestrian re-identification according to an embodiment of the invention.
Fig. 3 shows a schematic block diagram of an FEN comprised by a second neural network for pedestrian re-identification according to an embodiment of the present invention.
Fig. 4 shows a partial processing result of the FEN included in the second neural network for pedestrian re-recognition according to the embodiment of the present invention.
Fig. 5 shows a schematic block diagram of a PTN in an FEN included in a second neural network for pedestrian re-recognition according to an embodiment of the present invention.
Fig. 6 shows a detailed view of a third neural network for pedestrian re-identification according to an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
Fig. 1 shows a conceptual schematic diagram of a first neural network 100 for pedestrian re-identification according to an embodiment of the present invention.
As shown in fig. 1, a first type of neural network 100 for pedestrian re-identification includes a first neural network 101 and a second neural network 103.
As shown in fig. 1, the first neural network 101 uses an original whole-body image of a pedestrian as a first input and outputs a first recognition feature Fglobal
The second neural network 103 uses an affine transformation image of the human body part image extracted from the original whole-body image of the pedestrian as a second input and outputs a second recognition feature Fpart
Wherein the human body part at least comprises a head, a trunk and four limbs, and a first identification characteristic FglobalAnd a second identifying feature FpartCombined into a total recognition feature Ffusion
For example, as known to those skilled in the art, the original whole-body image of the pedestrian may be selected from pedestrian re-recognition data sets such as CUHK 03, Market1501, and VIPeR for training the weight of the first neural network or testing the correct recognition rate.
For example, as known to those skilled in the art, the body part image may be manually extracted from the original whole-body image of the pedestrian by an artificial manner, and the affine transformation parameters are manually adjusted to obtain an affine transformation image, which is used as the second input for training the weight of the second neural network or testing the correct recognition rate.
The extraction of the body region image can also be performed by a pose estimation algorithm commonly used in the art (e.g., a pose estimation algorithm based on a Full Convolution Network (FCN)), and specific technical solutions refer to j.long, e.shell, and article "full volumetric networks for segmentation", published by t.darrell in 2015 on CVPR.
Fig. 2 shows an overall schematic block diagram of a second type of neural network 200 for pedestrian re-identification according to an embodiment of the present invention.
As shown in fig. 2, in order to realize a function of automatically acquiring an affine transformation image of a human body part image from an original whole body image of a pedestrian, a second neural network 200 for pedestrian re-recognition includes a FEN 201 in addition to a first neural network 101 and a second neural network 103.
The FEN 201 is configured to automatically acquire a human body part image from an original whole-body image of a pedestrian, then perform affine transformation on the human body part image using PTN, which takes the original whole-body image of the pedestrian as an input, and output the image of the human body part image after arithmetic processing, which has undergone PTN affine transformation, to the second neural network 103.
Fig. 3 shows a schematic block diagram of an FEN 201 comprised by a second neural network 200 for pedestrian re-identification according to an embodiment of the present invention.
As shown in fig. 3, FEN 201 includes an FCN-based pose estimation module 301 using existing techniques (e.g., as disclosed by j. long et al and which may be separately trained and incorporated into FEN 201), a human body position image extraction module 303 using techniques customary in the art, and first, second, third, fourth, and fifth PTNs 305, 307, 309, 311, 313. The first PTN 305, the second PTN 307, the third PTN 309, the fourth PTN 311, and the fifth PTN 313 are respectively configured to perform affine transformation on the images of the left arm, the right arm, the torso, the left leg, and the right leg of the pedestrian extracted by the human body part image extraction module 303, and finally obtain affine transformation images of human body parts other than the head. Finally, the FEN 201 combines the head original image and the affine transformation image of each human body part other than the head, and outputs the combined image as a PTN affine-transformed image.
That is, the FEN 201 includes PTN for applying affine transformation to each human body part in the human body parts map to obtain an affine transformation image of a more robust human body parts map image.
Fig. 4 shows a partial processing result of the FEN 201 included in the second neural network 200 for pedestrian re-recognition according to the embodiment of the present invention.
As shown in fig. 4, its subgraph (a) is a pair of original whole-body images of the same pedestrian (different view angles are adopted for camera shooting). Sub-diagram (b) is a diagram of the human body posture estimation result obtained by the human body posture estimation module 301 using FCN, for example. Sub-diagram (c) is a pair of joint indication diagrams labeled 14 human joints, for example, acquired by pose estimation module 301 using FCN. Sub-diagram (d) is a pixel region corresponding to each of 6 human body parts including a head, a torso, and limbs, which are acquired by the human posture estimation module 301 using FCN, for example. Sub-diagram (e) is, for example, an image including only the above 6 human body parts acquired by using the human body part image extraction module 303. The sub-image (f) is, for example, an affine transformation image obtained by processing the image in the sub-image (e) using PTN. Optionally, the image in (e) may be rotated and scaled, and the image in (f) may be normalized.
Fig. 5 shows a schematic block diagram of a PTN in the FEN 201 included in the second neural network for pedestrian re-recognition according to the embodiment of the present invention.
As shown in fig. 5, the PTN network (i.e., the first PTN 305, the second PTN 307, the third PTN 309, the fourth PTN 311, and the fifth PTN 313 shown in fig. 3) is used to acquire parameters a for performing affine transformation of an image of limbs (left arm, right arm, left leg, and right leg) and an image of a trunk by trainingTheta-limbs or trunk. The PTN network was modified based on the Spatial Transform Network (STN) disclosed in the article Spatial transform networks 2015 published by m.jaderberg, k.simonyan, a.zisserman et al on NIPS.
As shown in formula (1), the affine transformation parameter A for the human body part image obtained by the PTN networkθIs 6-dimensional, of which theta1、θ2、θ4And theta5Is the scaling and rotation parameters, and theta3And theta6Is a translation parameter. (x)s,ys) Is the pixel coordinate in the original image of the body part, and (x)t,yt) Is the pixel coordinates in the affine transformation image of the human body part.
Optionally, as shown in fig. 2, the second neural network 200 for pedestrian re-identification further comprises FWN 203.
FWN 203 is used to weight and bias the second identifying features output by the second neural network 103 to combine the first identifying features and the second identifying features into a total identifying feature. The weighting and biasing are implemented by equation (2).
tanh(Fpart⊙W+B) (2)
Wherein,w and B are weight and bias vector, respectively, which are associated with the second identifying feature FpartIs the same dimension, ⊙ denotes a Hadamard product operation.
Also, to implement error back-propagation, the following gradient update formula may be defined for FWN 203 for training FWN 203.
Wherein f isi∈Ffusion(i=1,2…m+n),gj∈|Fglobal(j=1,2…m),pk∈Fpart(k=1,2…n),wkI ∈ W (k ═ 1, 2 … n), B ∈ B (k ═ 1, 2 … n), m and n are F, respectivelyglobalAnd FpartDimension (d) of (a).
Fig. 6 shows a detailed schematic diagram of a third neural network 300 for pedestrian re-identification according to an embodiment of the present invention.
As shown in FIG. 6, the first neural network used by the third neural network 300 includes a first CNN and a second CNN (CNN in FIG. 6)g) The second neural network includes a third CNN and a fourth CNN (CNN in FIG. 6)p) And the first CNN and the third CNN share the weight value.
Alternatively, as shown in fig. 6, the second CNN and the fourth CNN used by the third neural network 300 use independent weights.
Optionally, as shown in fig. 6, the first neural network and the second neural network used by the third neural network 300 each further include a CONV layer and a GAP layer at respective output ends to map output features to appropriate dimensions for subsequent processing and may support input images of different sizes.
The invention also provides a pedestrian re-identification algorithm based on deep learning and combined with the neural network for pedestrian re-identification. The pedestrian re-identification algorithm based on deep learning comprises the following steps:
a neural network for pedestrian re-identification according to the above is constructed.
Training the neural network for pedestrian re-recognition using a pedestrian re-recognition dataset.
Pedestrian re-recognition is performed using a trained neural network for pedestrian re-recognition.
The technical scheme of the invention has better capability of feature extraction and feature matching, thereby improving the final correct recognition rate and/or reducing the false recognition rate.
The above description is only an exemplary embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the appended claims.

Claims (7)

1. A neural network for pedestrian re-identification, comprising:
a first neural network that uses an original whole-body image of a pedestrian as a first input and outputs a first recognition feature;
a second neural network that uses an affine transformation image of a human body part image extracted from an original whole-body image of a pedestrian as a second input and outputs a second recognition feature,
the human body part at least comprises a head, a trunk and four limbs, and the first identification feature and the second identification feature are combined into a total identification feature.
2. The neural network for pedestrian re-identification according to claim 1, further comprising:
the characteristic embedding sub-neural network FEN comprises a posture transformation neural network PTN, and the PTN is used for applying affine transformation to each human body part in the human body part map to obtain a more robust affine transformation image of the human body part map image.
3. The neural network for pedestrian re-identification according to claim 1 or 2, further comprising:
the feature weighting sub-neural network FWN is used to weight and bias the second identifying features output by the second neural network, thereby combining the first identifying features and the second identifying features into a total identifying feature.
4. The neural network for pedestrian re-identification according to claim 3, wherein the first neural network comprises a first Convolutional Neural Network (CNN) and a second CNN, and the second neural network comprises a third CNN and a fourth CNN, wherein the first CNN and the third CNN share weight values.
5. The neural network for pedestrian re-identification, as claimed in claim 4, wherein the second CNN and the fourth CNN use independent weights.
6. The neural network for pedestrian re-identification according to claim 1 or 2, characterized in that the first and second neural networks each further comprise a convolutional CONV layer and a global average pooling GAP layer at the respective output.
7. A pedestrian re-identification algorithm based on deep learning, comprising:
constructing a neural network for pedestrian re-identification according to any one of claims 1-6;
training the neural network for pedestrian re-recognition using a pedestrian re-recognition dataset;
pedestrian re-recognition is performed using a trained neural network for pedestrian re-recognition.
CN201710780179.0A 2017-09-01 2017-09-01 The neural network identified again for pedestrian and the pedestrian based on deep learning recognizer again Active CN107729805B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710780179.0A CN107729805B (en) 2017-09-01 2017-09-01 The neural network identified again for pedestrian and the pedestrian based on deep learning recognizer again

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710780179.0A CN107729805B (en) 2017-09-01 2017-09-01 The neural network identified again for pedestrian and the pedestrian based on deep learning recognizer again

Publications (2)

Publication Number Publication Date
CN107729805A true CN107729805A (en) 2018-02-23
CN107729805B CN107729805B (en) 2019-09-13

Family

ID=61205446

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710780179.0A Active CN107729805B (en) 2017-09-01 2017-09-01 The neural network identified again for pedestrian and the pedestrian based on deep learning recognizer again

Country Status (1)

Country Link
CN (1) CN107729805B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109101901A (en) * 2018-07-23 2018-12-28 北京旷视科技有限公司 Human action identification and its neural network generation method, device and electronic equipment
CN109886113A (en) * 2019-01-17 2019-06-14 桂林远望智能通信科技有限公司 A kind of spacious view pedestrian recognition methods again based on region candidate network
CN109934081A (en) * 2018-08-29 2019-06-25 厦门安胜网络科技有限公司 A kind of pedestrian's attribute recognition approach, device and storage medium based on deep neural network
CN109977837A (en) * 2019-03-20 2019-07-05 常熟理工学院 Pedestrian based on human body attitude invariant features recognition methods again
CN110543817A (en) * 2019-07-25 2019-12-06 北京大学 Pedestrian re-identification method based on posture guidance feature learning
CN110688873A (en) * 2018-07-04 2020-01-14 上海智臻智能网络科技股份有限公司 Multi-target tracking method and face recognition method
CN110956576A (en) * 2018-09-27 2020-04-03 北京小米移动软件有限公司 Image processing method, device, equipment and storage medium
US10783393B2 (en) 2017-06-20 2020-09-22 Nvidia Corporation Semi-supervised learning for landmark localization
CN111920436A (en) * 2020-07-08 2020-11-13 浙江大学 Dual-tracer PET (positron emission tomography) separation method based on multi-task learning three-dimensional convolutional coding and decoding network
CN112990144A (en) * 2021-04-30 2021-06-18 德鲁动力科技(成都)有限公司 Data enhancement method and system for pedestrian re-identification
CN113255615A (en) * 2021-07-06 2021-08-13 南京视察者智能科技有限公司 Pedestrian retrieval method and device for self-supervision learning
TWI740624B (en) * 2019-10-28 2021-09-21 中國商深圳市商湯科技有限公司 Image processing method, device and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140177946A1 (en) * 2012-12-21 2014-06-26 Electronics and Telecommunicatidons Research Institute Human detection apparatus and method
CN104915643A (en) * 2015-05-26 2015-09-16 中山大学 Deep-learning-based pedestrian re-identification method
CN106022392A (en) * 2016-06-02 2016-10-12 华南理工大学 Deep neural network sample automatic accepting and rejecting training method
CN106529442A (en) * 2016-10-26 2017-03-22 清华大学 Pedestrian identification method and apparatus
CN106845415A (en) * 2017-01-23 2017-06-13 中国石油大学(华东) A kind of pedestrian based on deep learning becomes more meticulous recognition methods and device
CN106951872A (en) * 2017-03-24 2017-07-14 江苏大学 A kind of recognition methods again of the pedestrian based on unsupervised depth model and hierarchy attributes

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140177946A1 (en) * 2012-12-21 2014-06-26 Electronics and Telecommunicatidons Research Institute Human detection apparatus and method
CN104915643A (en) * 2015-05-26 2015-09-16 中山大学 Deep-learning-based pedestrian re-identification method
CN106022392A (en) * 2016-06-02 2016-10-12 华南理工大学 Deep neural network sample automatic accepting and rejecting training method
CN106529442A (en) * 2016-10-26 2017-03-22 清华大学 Pedestrian identification method and apparatus
CN106845415A (en) * 2017-01-23 2017-06-13 中国石油大学(华东) A kind of pedestrian based on deep learning becomes more meticulous recognition methods and device
CN106951872A (en) * 2017-03-24 2017-07-14 江苏大学 A kind of recognition methods again of the pedestrian based on unsupervised depth model and hierarchy attributes

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10783394B2 (en) 2017-06-20 2020-09-22 Nvidia Corporation Equivariant landmark transformation for landmark localization
US10783393B2 (en) 2017-06-20 2020-09-22 Nvidia Corporation Semi-supervised learning for landmark localization
CN110688873A (en) * 2018-07-04 2020-01-14 上海智臻智能网络科技股份有限公司 Multi-target tracking method and face recognition method
CN109101901A (en) * 2018-07-23 2018-12-28 北京旷视科技有限公司 Human action identification and its neural network generation method, device and electronic equipment
CN109101901B (en) * 2018-07-23 2020-10-27 北京旷视科技有限公司 Human body action recognition method and device, neural network generation method and device and electronic equipment
CN109934081A (en) * 2018-08-29 2019-06-25 厦门安胜网络科技有限公司 A kind of pedestrian's attribute recognition approach, device and storage medium based on deep neural network
CN110956576A (en) * 2018-09-27 2020-04-03 北京小米移动软件有限公司 Image processing method, device, equipment and storage medium
CN110956576B (en) * 2018-09-27 2024-03-22 北京小米移动软件有限公司 Image processing method, device, equipment and storage medium
CN109886113A (en) * 2019-01-17 2019-06-14 桂林远望智能通信科技有限公司 A kind of spacious view pedestrian recognition methods again based on region candidate network
CN109977837A (en) * 2019-03-20 2019-07-05 常熟理工学院 Pedestrian based on human body attitude invariant features recognition methods again
CN110543817A (en) * 2019-07-25 2019-12-06 北京大学 Pedestrian re-identification method based on posture guidance feature learning
TWI740624B (en) * 2019-10-28 2021-09-21 中國商深圳市商湯科技有限公司 Image processing method, device and storage medium
CN111920436A (en) * 2020-07-08 2020-11-13 浙江大学 Dual-tracer PET (positron emission tomography) separation method based on multi-task learning three-dimensional convolutional coding and decoding network
CN112990144A (en) * 2021-04-30 2021-06-18 德鲁动力科技(成都)有限公司 Data enhancement method and system for pedestrian re-identification
CN113255615A (en) * 2021-07-06 2021-08-13 南京视察者智能科技有限公司 Pedestrian retrieval method and device for self-supervision learning

Also Published As

Publication number Publication date
CN107729805B (en) 2019-09-13

Similar Documents

Publication Publication Date Title
CN107729805B (en) The neural network identified again for pedestrian and the pedestrian based on deep learning recognizer again
CN109684924B (en) Face living body detection method and device
Wang et al. 360sd-net: 360 stereo depth estimation with learnable cost volume
CN111862296B (en) Three-dimensional reconstruction method, three-dimensional reconstruction device, three-dimensional reconstruction system, model training method and storage medium
GB2608975A (en) Person identification across multiple captured images
CN103607554A (en) Fully-automatic face seamless synthesis-based video synthesis method
CN107767339B (en) Binocular stereo image splicing method
CN112528902B (en) Video monitoring dynamic face recognition method and device based on 3D face model
CN111260543A (en) Underwater image splicing method based on multi-scale image fusion and SIFT features
CN108171735B (en) Billion pixel video alignment method and system based on deep learning
Wang et al. A segmentation based robust deep learning framework for multimodal retinal image registration
CN110263605A (en) Pedestrian's dress ornament color identification method and device based on two-dimension human body guise estimation
Charco et al. Deep learning based camera pose estimation in multi-view environment
CN111553939A (en) Image registration algorithm of multi-view camera
CN111898571A (en) Action recognition system and method
Ruan et al. Image stitching algorithm based on SURF and wavelet transform
Zheng et al. Online subspace learning from gradient orientations for robust image alignment
CN114708617A (en) Pedestrian re-identification method and device and electronic equipment
CN111582036A (en) Cross-view-angle person identification method based on shape and posture under wearable device
CN111126250A (en) Pedestrian re-identification method and device based on PTGAN
Mo et al. A Robust Infrared and Visible Image Registration Method for Dual Sensor UAV System
CN110321452A (en) A kind of image search method based on direction selection mechanism
CN111047513B (en) Robust image alignment method and device for cylindrical panorama stitching
CN114066954A (en) Feature extraction and registration method for multi-modal images
Alzohairy et al. Image mosaicing based on neural networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant