CN117710903B - Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models - Google Patents

Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models Download PDF

Info

Publication number
CN117710903B
CN117710903B CN202410161198.5A CN202410161198A CN117710903B CN 117710903 B CN117710903 B CN 117710903B CN 202410161198 A CN202410161198 A CN 202410161198A CN 117710903 B CN117710903 B CN 117710903B
Authority
CN
China
Prior art keywords
picture
loss
model
distance
extracting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202410161198.5A
Other languages
Chinese (zh)
Other versions
CN117710903A (en
Inventor
庄建军
万超
庄宇辰
万理
陈永强
王楠
庞明义
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN202410161198.5A priority Critical patent/CN117710903B/en
Publication of CN117710903A publication Critical patent/CN117710903A/en
Application granted granted Critical
Publication of CN117710903B publication Critical patent/CN117710903B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a visual specific pedestrian tracking method and system based on ReID and Yolov double models, belonging to the technical field of image recognition, wherein the method comprises the following steps: (1) Pedestrian data are obtained, and the data are preprocessed and stored; (2) Extracting the preprocessed data frame by frame, extracting person pictures in each frame of pictures by YOLOv, and storing the person pictures; (3) Acquiring a mark 1501 data set, randomly dividing a test set and a training set, and training Resnet a model to acquire a final model; (4) Inputting the trained model into the picture obtained in the step (2) through a transfer learning method, and extracting the characteristics; (5) Combining the extracted pictures into a tensor, transmitting the tensor to a ReID model, and carrying out feature extraction and normalization; (6) determining the magnitude of the minimum average distance and the distance threshold; (7) saving the final video; according to the invention, by means of YOLOv target detection of a specific class and ReID search of a specific ID, the function of tracking offender in collaborative search is realized.

Description

Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models
Technical Field
The invention relates to the technical field of image recognition, in particular to a visual specific pedestrian tracking method and system based on ReID and Yolov5 double models.
Background
In recent years, with the great increase of the storage amount of non-motor vehicles in campuses and the improvement of the opening degree of universities, more and more non-motor vehicles enter the campuses and illegal parking occurs. The safety problem of campus traffic is increasingly outstanding, and the safety problem of campus traffic is deeply concerned by school leaders, teachers and students. Therefore, how to solve the safety problem of campus traffic is urgent.
Disclosure of Invention
The invention aims to: the invention aims to provide a visual specific pedestrian tracking method and system based on ReID and Yolov double models, which realize pedestrian re-identification of photo-video under the double models by means of model transformation through a self-training model of a mark 1501 dataset and a strong base line ReID and a Yolov5s.pt weight file initialization model, and search specific IDs frame by frame to solve the security problem of a campus.
The technical scheme is as follows: the invention discloses a visual specific pedestrian tracking method based on ReID and Yolov double models, which comprises the following steps:
(1) Acquiring pedestrian data by using a plurality of 1280 x 1080 high-definition cameras and SD cameras, preprocessing the data and storing the data;
(2) Extracting the preprocessed data frame by frame, extracting a person picture in each frame of picture by YOLOv, and storing the picture by extracting the person picture coordinates;
(3) Acquiring a mark 1501 data set, randomly dividing a test set and a training set, and training Resnet a model to acquire a final model;
(4) Inputting the trained model into the picture obtained in the step (2) through a transfer learning method, and extracting the characteristics;
(5) Combining the extracted pictures into a tensor, transmitting the tensor to a trained ReID model, extracting and normalizing the characteristics, and finding out the index of the minimum average distance by calculating the mean value of Euclidean distances between the query picture and each stored picture;
(6) Setting corresponding distance thresholds through different scenes, judging the size of the minimum average distance and the distance threshold, and outputting a final target boundary frame when the minimum average distance is smaller than the distance threshold, wherein the final target boundary frame comprises position coordinates, confidence coefficient and category information of the boundary frame;
(7) And saving the final video.
Further, in the step (1), the pretreatment is specifically as follows: labeling the data, including manual labeling and labeling with a DPM detector; fixing the image size to 256×128; each image is segmented with a probability level of 0.5, each image is decoded into 32-bit anchor point original pixel values, and the RGB channels are normalized.
Further, in the step (1), the preprocessed data is stored, and naming criteria are as follows: 000N_cNsN_000ABC_0N.jpeg or 000N_cNsN_000ABC_00.jpeg; wherein 000N represents the tag number of each person; cN represents the nth camera; sN represents an nth video clip; 000ABC represents the picture of the 000 th ABC frame of cNsN; 0N represents the Nth detection box on cNsN _000ABC frames.
Further, the step (3) specifically includes the following steps: the dimension of the Resnet model full-connection layer is modified to N; optimizing the model by adopting an Adam method; selecting identity loss, center loss and enhanced triplet loss as total losses in back propagation; the formula is as follows:
Wherein, For identity loss,/>For enhanced triplet loss,/>Is the center loss.
Further, the identity loss formula is as follows:
wherein n represents the number of training samples in each batch, provided with a given label Input image/>ThenRepresentation/>Identified as category/>Is used for predicting the probability of (1);
The enhanced triplet loss is as follows:
wherein (i, j, k) represents each anchor sample Triplet in each training batch,/>Representing the distance between positive sample pairs,/>Representing the distance between the negative pair of samples; for anchor samples/>And/>Is the corresponding positive sample; /(I)Expressed as anchor sample/>And its corresponding positive sample/>Weight value of Euclidean distance of/>Is a weight value, N represents the positive sample pair number, p represents the environment of the positive sample pair, and N represents the environment of the negative sample pair;
the euclidean distance between two samples is:
Wherein, And/>Representation/>And/>Corresponding feature vectors;
the definition formula of k is:
the square error is expressed as:
The center loss formula is as follows:
Wherein, Representing identity/>Is a class center of (c).
The invention discloses a visual specific pedestrian tracking system based on ReID and Yolov double models, which comprises the following components:
and a pretreatment module: the method comprises the steps that pedestrian data are obtained by using a plurality of 1280 x 1080 high-definition cameras and SD cameras, and are preprocessed and stored;
And a picture extraction module: the method comprises the steps of extracting preprocessed data frame by frame, extracting a person picture in each frame of picture by YOLOv, and storing the picture by extracting a person picture coordinate;
Model training module: the method comprises the steps of obtaining a mark 1501 data set, randomly dividing a test set and a training set, training Resnet models, and obtaining a final model;
And the feature extraction module is used for: the method comprises the steps of inputting a trained model into a picture obtained by a picture extraction module through a transfer learning method, and extracting features;
And an index module: the method comprises the steps of merging extracted pictures into a tensor, transmitting the tensor to a trained ReID model, extracting and normalizing characteristics, and finding out an index of the minimum average distance by calculating the average value of Euclidean distances between a query picture and each stored picture;
And an output module: the method comprises the steps of setting corresponding distance thresholds through different scenes, judging the size of the minimum average distance and the distance threshold, and outputting a final target boundary frame when the minimum average distance is smaller than the distance threshold, wherein the final target boundary frame comprises position coordinates, confidence and category information of the boundary frame;
And a storage module: for saving the final video.
Further, in the preprocessing module, the preprocessing is specifically as follows: labeling the data, including manual labeling and labeling with a DPM detector; fixing the image size to 256×128; each image is segmented with a probability level of 0.5, each image is decoded into 32-bit anchor point original pixel values, and the RGB channels are normalized.
Further, in the preprocessing module, the preprocessed data is stored, and naming criteria are as follows: 000N_cNsN_000ABC_0N.jpeg or 000N_cNsN_000ABC_00.jpeg; wherein 000N represents the tag number of each person; cN represents the nth camera; sN represents an nth video clip; 000ABC represents the picture of the 000 th ABC frame of cNsN; 0N represents the Nth detection box on cNsN _000ABC frames.
Further, in the model training module, the specific steps are as follows: the dimension of the Resnet model full-connection layer is modified to N; optimizing the model by adopting an Adam method; selecting identity loss, center loss and enhanced triplet loss as total losses in back propagation; the formula is as follows:
Wherein, For identity loss,/>For enhanced triplet loss,/>Is the center loss.
Further, in the model training module, the identity loss formula is as follows:
wherein n represents the number of training samples in each batch, provided with a given label Input image/>ThenRepresentation/>Identified as category/>Is used for predicting the probability of (1);
The enhanced triplet loss is as follows:
wherein (i, j, k) represents each anchor sample Triplet in each training batch,/>Representing the distance between positive sample pairs,/>Representing the distance between the negative pair of samples; for anchor samples/>And/>Is the corresponding positive sample; /(I)Expressed as anchor sample/>And its corresponding positive sample/>Weight value of Euclidean distance of/>Is a weight value, N represents the positive sample pair number, p represents the environment of the positive sample pair, and N represents the environment of the negative sample pair;
the euclidean distance between two samples is:
Wherein, And/>Representation/>And/>Corresponding feature vectors;
the definition formula of k is:
the square error is expressed as:
The center loss formula is as follows:
Wherein, Representing identity/>Is a class center of (c).
The beneficial effects are that: compared with the prior art, the invention has the following remarkable advantages: the function of cooperatively searching and tracking the offender under multiple cameras is realized by YOLOv for target detection of a specific class (person) and ReID for searching of a specific ID.
Drawings
Fig. 1 is a schematic diagram of the principle of the present invention.
Detailed Description
The technical scheme of the invention is further described below with reference to the accompanying drawings.
As shown in fig. 1, an embodiment of the present invention provides a method for tracking a specific pedestrian based on ReID and Yolov5 double models, which includes the following steps:
(1) Acquiring pedestrian data by using a plurality of 1280 x 1080 high-definition cameras and SD cameras, preprocessing the data and storing the data; the pretreatment is specifically as follows: labeling the data, including manual labeling and labeling with a DPM detector; fixing the image size to 256×128; dividing each image with 0.5 probability level, decoding each image into 32-bit locating point original pixel values, and carrying out normalization processing on RGB channels; storing the preprocessed data, wherein naming criteria are as follows: 000N_cNsN_000ABC_0N.jpeg or 000N_cNsN_000ABC_00.jpeg; wherein 000N represents the tag number of each person; cN represents the nth camera; sN represents an nth video clip; 000ABC represents the picture of the 000 th ABC frame of cNsN; 0N represents the Nth detection box on cNsN _000ABC frame
(2) Extracting the preprocessed data frame by frame, extracting a person picture in each frame of picture by YOLOv, and storing the picture by extracting the person picture coordinates;
(3) Acquiring a mark 1501 data set, randomly dividing a test set and a training set, and training Resnet a model to acquire a final model; the method comprises the following steps: the dimension of the Resnet model full-connection layer is modified to N; optimizing the model by adopting an Adam method; selecting identity loss, center loss and enhanced triplet loss as total losses in back propagation; the formula is as follows:
Wherein, For identity loss,/>For enhanced triplet loss,/>Is the center loss.
The identity loss formula is as follows:
wherein n represents the number of training samples in each batch, provided with a given label Input image/>ThenRepresentation/>Identified as category/>Is used for predicting the probability of (1);
The enhanced triplet loss is as follows:
wherein (i, j, k) represents each anchor sample Triplet in each training batch,/>Representing the distance between positive sample pairs,/>Representing the distance between the negative pair of samples; for anchor samples/>And/>Is the corresponding positive sample; /(I)Expressed as anchor sample/>And its corresponding positive sample/>Weight value of Euclidean distance of/>Is a weight value, N represents the positive sample pair number, p represents the environment of the positive sample pair, and N represents the environment of the negative sample pair;
the euclidean distance between two samples is:
Wherein, And/>Representation/>And/>Corresponding feature vectors;
the definition formula of k is:
the square error is expressed as:
The center loss formula is as follows:
Wherein, Representing identity/>Is a class center of (c).
(4) Inputting the trained model into the picture obtained in the step (2) through a transfer learning method, and extracting the characteristics;
(5) And combining the extracted pictures into a tensor, transmitting the tensor to a trained ReID model, extracting and normalizing the characteristics, and finding the index of the minimum average distance by calculating the mean value of Euclidean distances between the query picture and each stored picture.
(6) Setting corresponding distance thresholds through different scenes, judging the size of the minimum average distance and the distance threshold, and outputting a final target boundary frame when the minimum average distance is smaller than the distance threshold, wherein the final target boundary frame comprises position coordinates, confidence coefficient and category information of the boundary frame;
(7) And saving the final video.
By comparing various leading edge methods with the pedestrian re-recognition method of the invention, the invention has excellent performance in two data set training as shown in table 1.
TABLE 1 training comparison of this model with other leading edge models
The embodiment of the invention also provides a visual specific pedestrian tracking system based on ReID and Yolov double models, which comprises the following steps:
And a pretreatment module: the method comprises the steps that pedestrian data are obtained by using a plurality of 1280 x 1080 high-definition cameras and SD cameras, and are preprocessed and stored; the pretreatment is specifically as follows: labeling the data, including manual labeling and labeling with a DPM detector; fixing the image size to 256×128; each image is segmented with a probability level of 0.5, each image is decoded into 32-bit anchor point original pixel values, and the RGB channels are normalized. Storing the preprocessed data, wherein naming criteria are as follows: 000N_cNsN_000ABC_0N.jpeg or 000N_cNsN_000ABC_00.jpeg; wherein 000N represents the tag number of each person; cN represents the nth camera; sN represents an nth video clip; 000ABC represents the picture of the 000 th ABC frame of cNsN; 0N represents the Nth detection box on cNsN _000ABC frames.
And a picture extraction module: the method comprises the steps of extracting preprocessed data frame by frame, extracting a person picture in each frame of picture by YOLOv, and storing the picture by extracting a person picture coordinate;
Model training module: the method comprises the steps of obtaining a mark 1501 data set, randomly dividing a test set and a training set, training Resnet models, and obtaining a final model; the method comprises the following steps: the dimension of the Resnet model full-connection layer is modified to N; optimizing the model by adopting an Adam method; selecting identity loss, center loss and enhanced triplet loss as total losses in back propagation; the formula is as follows:
Wherein, For identity loss,/>For enhanced triplet loss,/>Is the center loss.
The identity loss formula is as follows:
wherein n represents the number of training samples in each batch, provided with a given label Input image/>ThenRepresentation/>Identified as category/>Is used for predicting the probability of (1);
The enhanced triplet loss is as follows:
wherein (i, j, k) represents each anchor sample Triplet in each training batch,/>Representing the distance between positive sample pairs,/>Representing the distance between the negative pair of samples; for anchor samples/>And/>Is the corresponding positive sample; /(I)Expressed as anchor sample/>And its corresponding positive sample/>Weight value of Euclidean distance of/>Is a weight value, N represents the positive sample pair number, p represents the environment of the positive sample pair, and N represents the environment of the negative sample pair;
the euclidean distance between two samples is:
Wherein, And/>Representation/>And/>Corresponding feature vectors;
the definition formula of k is:
the square error is expressed as:
The center loss formula is as follows:
Wherein, Representing identity/>Is a class center of (c).
And the feature extraction module is used for: the method comprises the steps of inputting a trained model into a picture obtained by a picture extraction module through a transfer learning method, and extracting features;
And an index module: and the method is used for merging the extracted pictures into a tensor, transmitting the tensor to a trained ReID model, extracting and normalizing the characteristics, and finding the index of the minimum average distance by calculating the mean value of Euclidean distances between the query picture and each stored picture.
And an output module: the method comprises the steps of setting corresponding distance thresholds through different scenes, judging the size of the minimum average distance and the distance threshold, and outputting a final target boundary frame when the minimum average distance is smaller than the distance threshold, wherein the final target boundary frame comprises position coordinates, confidence and category information of the boundary frame;
And a storage module: for saving the final video.

Claims (6)

1. A method for visual specific pedestrian tracking based on ReID and Yolov5 double models, comprising the steps of:
(1) Acquiring pedestrian data by using a plurality of 1280 x 1080 high-definition cameras and SD cameras, preprocessing the data and storing the data;
(2) Extracting the preprocessed data frame by frame, extracting a person picture in each frame of picture by YOLOv, and storing the picture by extracting the person picture coordinates;
(3) Acquiring a mark 1501 data set, randomly dividing a test set and a training set, and training Resnet a model to acquire a final model; the method comprises the following steps: the dimension of the Resnet model full-connection layer is modified to N; optimizing the model by adopting an Adam method; selecting identity loss, center loss and enhanced triplet loss as total losses in back propagation; the formula is as follows:
Wherein, For identity loss,/>For enhanced triplet loss,/>Is the center loss; the identity loss formula is as follows:
wherein n represents the number of training samples in each batch, provided with a given label Input image/>Then/>Representation/>Identified as category/>Is used for predicting the probability of (1);
The enhanced triplet loss is as follows:
wherein (i, j, k) represents each anchor sample Triplet in each training batch,/>Representing the distance between positive sample pairs,/>Representing the distance between the negative pair of samples; for anchor samples/>And/>Is the corresponding positive sample; /(I)Expressed as anchor sample/>And its corresponding positive sample/>Weight value of Euclidean distance of/>Is a weight value, N represents the positive sample pair number, p represents the environment of the positive sample pair, and N represents the environment of the negative sample pair;
the euclidean distance between two samples is:
Wherein, And/>Representation/>And/>Corresponding feature vectors;
the definition formula of k is:
the square error is expressed as:
The center loss formula is as follows:
Wherein, Representing identity/>Is a class center of (2);
(4) Inputting the trained model into the picture obtained in the step (2) through a transfer learning method, and extracting the characteristics;
(5) Combining the extracted pictures into a tensor, transmitting the tensor to a trained ReID model, extracting and normalizing the characteristics, and finding out the index of the minimum average distance by calculating the mean value of Euclidean distances between the query picture and each stored picture;
(6) Setting corresponding distance thresholds through different scenes, judging the size of the minimum average distance and the distance threshold, and outputting a final target boundary frame when the minimum average distance is smaller than the distance threshold, wherein the final target boundary frame comprises position coordinates, confidence coefficient and category information of the boundary frame;
(7) And saving the final video.
2. The method for visual specific pedestrian tracking based on ReID and Yolov double models according to claim 1, wherein in the step (1), the preprocessing is specifically as follows: labeling the data, including manual labeling and labeling with a DPM detector; fixing the image size to 256×128; each image is segmented with a probability level of 0.5, each image is decoded into 32-bit anchor point original pixel values, and the RGB channels are normalized.
3. The method for visual specific pedestrian tracking based on ReID and Yolov double models according to claim 1, wherein in the step (1), the preprocessed data is stored, and naming criteria are as follows: 000N_cNsN_000ABC_0N.jpeg or 000N_cNsN_000ABC_00.jpeg; wherein 000N represents the tag number of each person; cN represents the nth camera; sN represents an nth video clip; 000ABC represents the picture of the 000 th ABC frame of cNsN; 0N represents the Nth detection box on cNsN _000ABC frames.
4. A visual specific pedestrian tracking system based on ReID and Yolov double models, comprising:
and a pretreatment module: the method comprises the steps that pedestrian data are obtained by using a plurality of 1280 x 1080 high-definition cameras and SD cameras, and are preprocessed and stored;
And a picture extraction module: the method comprises the steps of extracting preprocessed data frame by frame, extracting a person picture in each frame of picture by YOLOv, and storing the picture by extracting a person picture coordinate;
Model training module: the method comprises the steps of obtaining a mark 1501 data set, randomly dividing a test set and a training set, training Resnet models, and obtaining a final model; the method comprises the following steps: the dimension of the Resnet model full-connection layer is modified to N; optimizing the model by adopting an Adam method; selecting identity loss, center loss and enhanced triplet loss as total losses in back propagation; the formula is as follows:
Wherein, For identity loss,/>For enhanced triplet loss,/>Is the center loss; the identity loss formula is as follows:
wherein n represents the number of training samples in each batch, provided with a given label Input image/>Then/>Representation/>Identified as category/>Is used for predicting the probability of (1);
The enhanced triplet loss is as follows:
wherein (i, j, k) represents each anchor sample Triplet in each training batch,/>Representing the distance between positive sample pairs,/>Representing the distance between the negative pair of samples; for anchor samples/>And/>Is the corresponding positive sample; /(I)Expressed as anchor sample/>And its corresponding positive sample/>Weight value of Euclidean distance of/>Is a weight value, N represents the positive sample pair number, p represents the environment of the positive sample pair, and N represents the environment of the negative sample pair;
the euclidean distance between two samples is:
Wherein, And/>Representation/>And/>Corresponding feature vectors;
the definition formula of k is:
the square error is expressed as:
The center loss formula is as follows:
Wherein, Representing identity/>Is a class center of (2);
And the feature extraction module is used for: the method comprises the steps of inputting a trained model into a picture obtained by a picture extraction module through a transfer learning method, and extracting features;
And an index module: the method comprises the steps of merging extracted pictures into a tensor, transmitting the tensor to a trained ReID model, extracting and normalizing characteristics, and finding out an index of the minimum average distance by calculating the average value of Euclidean distances between a query picture and each stored picture;
And an output module: the method comprises the steps of setting corresponding distance thresholds through different scenes, judging the size of the minimum average distance and the distance threshold, and outputting a final target boundary frame when the minimum average distance is smaller than the distance threshold, wherein the final target boundary frame comprises position coordinates, confidence and category information of the boundary frame;
And a storage module: for saving the final video.
5. The visual specific pedestrian tracking system based on ReID and Yolov double models according to claim 4, wherein the preprocessing module is configured to: labeling the data, including manual labeling and labeling with a DPM detector; fixing the image size to 256×128; each image is segmented with a probability level of 0.5, each image is decoded into 32-bit anchor point original pixel values, and the RGB channels are normalized.
6. The visual specific pedestrian tracking system based on ReID and Yolov double models of claim 4, wherein the preprocessing module stores the preprocessed data, and naming criteria are as follows: 000N_cNsN_000ABC_0N.jpeg or 000N_cNsN_000ABC_00.jpeg; wherein 000N represents the tag number of each person; cN represents the nth camera; sN represents an nth video clip; 000ABC represents the picture of the 000 th ABC frame of cNsN; 0N represents the Nth detection box on cNsN _000ABC frames.
CN202410161198.5A 2024-02-05 2024-02-05 Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models Active CN117710903B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410161198.5A CN117710903B (en) 2024-02-05 2024-02-05 Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410161198.5A CN117710903B (en) 2024-02-05 2024-02-05 Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models

Publications (2)

Publication Number Publication Date
CN117710903A CN117710903A (en) 2024-03-15
CN117710903B true CN117710903B (en) 2024-05-03

Family

ID=90153828

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410161198.5A Active CN117710903B (en) 2024-02-05 2024-02-05 Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models

Country Status (1)

Country Link
CN (1) CN117710903B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111881777A (en) * 2020-07-08 2020-11-03 泰康保险集团股份有限公司 Video processing method and device
CN113158891A (en) * 2021-04-20 2021-07-23 杭州像素元科技有限公司 Cross-camera pedestrian re-identification method based on global feature matching
CN113408356A (en) * 2021-05-21 2021-09-17 深圳市广电信义科技有限公司 Pedestrian re-identification method, device and equipment based on deep learning and storage medium
CN115497056A (en) * 2022-11-21 2022-12-20 南京华苏科技有限公司 Method for detecting lost articles in region based on deep learning
CN115841649A (en) * 2022-11-21 2023-03-24 哈尔滨工程大学 Multi-scale people counting method for urban complex scene
WO2023093241A1 (en) * 2021-11-29 2023-06-01 中兴通讯股份有限公司 Pedestrian re-identification method and apparatus, and storage medium
CN117079309A (en) * 2023-07-26 2023-11-17 上海云从企业发展有限公司 ReID model training method, reID pedestrian recognition method, device and medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111881777A (en) * 2020-07-08 2020-11-03 泰康保险集团股份有限公司 Video processing method and device
CN113158891A (en) * 2021-04-20 2021-07-23 杭州像素元科技有限公司 Cross-camera pedestrian re-identification method based on global feature matching
CN113408356A (en) * 2021-05-21 2021-09-17 深圳市广电信义科技有限公司 Pedestrian re-identification method, device and equipment based on deep learning and storage medium
WO2023093241A1 (en) * 2021-11-29 2023-06-01 中兴通讯股份有限公司 Pedestrian re-identification method and apparatus, and storage medium
CN115497056A (en) * 2022-11-21 2022-12-20 南京华苏科技有限公司 Method for detecting lost articles in region based on deep learning
CN115841649A (en) * 2022-11-21 2023-03-24 哈尔滨工程大学 Multi-scale people counting method for urban complex scene
CN117079309A (en) * 2023-07-26 2023-11-17 上海云从企业发展有限公司 ReID model training method, reID pedestrian recognition method, device and medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于改进YOLOv5s的视频行人目标检索平台设计与实现;李铭伦;中国优秀硕士学位论文全文数据库 信息科技辑;20230512;全文 *

Also Published As

Publication number Publication date
CN117710903A (en) 2024-03-15

Similar Documents

Publication Publication Date Title
CN109344787B (en) Specific target tracking method based on face recognition and pedestrian re-recognition
US20220415027A1 (en) Method for re-recognizing object image based on multi-feature information capture and correlation analysis
CN111832514B (en) Unsupervised pedestrian re-identification method and unsupervised pedestrian re-identification device based on soft multiple labels
CN112004111B (en) News video information extraction method for global deep learning
CN102549603B (en) Relevance-based image selection
CN108960184B (en) Pedestrian re-identification method based on heterogeneous component deep neural network
CN110717411A (en) Pedestrian re-identification method based on deep layer feature fusion
CN111582178B (en) Vehicle weight recognition method and system based on multi-azimuth information and multi-branch neural network
CN114067444A (en) Face spoofing detection method and system based on meta-pseudo label and illumination invariant feature
CN112580657B (en) Self-learning character recognition method
CN112836675B (en) Unsupervised pedestrian re-identification method and system for generating pseudo tags based on clusters
CN112215190A (en) Illegal building detection method based on YOLOV4 model
CN109635647B (en) Multi-picture multi-face clustering method based on constraint condition
CN112464775A (en) Video target re-identification method based on multi-branch network
CN114596548A (en) Target detection method, target detection device, computer equipment and computer-readable storage medium
CN114596546A (en) Vehicle weight recognition method and device, computer and readable storage medium
CN117710903B (en) Visual specific pedestrian tracking method and system based on ReID and Yolov5 double models
CN110765940B (en) Target object statistical method and device
CN112541453A (en) Luggage weight recognition model training and luggage weight recognition method
CN113435329B (en) Unsupervised pedestrian re-identification method based on video track feature association learning
CN115937862A (en) End-to-end container number identification method and system
CN112215189A (en) Accurate detecting system for illegal building
CN110880022A (en) Labeling method, labeling device and storage medium
CN116052220B (en) Pedestrian re-identification method, device, equipment and medium
CN117690031B (en) SAM model-based small sample learning remote sensing image detection method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant