CN112200076A - Method for carrying out multi-target tracking based on head and trunk characteristics - Google Patents

Method for carrying out multi-target tracking based on head and trunk characteristics Download PDF

Info

Publication number
CN112200076A
CN112200076A CN202011076008.8A CN202011076008A CN112200076A CN 112200076 A CN112200076 A CN 112200076A CN 202011076008 A CN202011076008 A CN 202011076008A CN 112200076 A CN112200076 A CN 112200076A
Authority
CN
China
Prior art keywords
image
head
human body
target
key points
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011076008.8A
Other languages
Chinese (zh)
Other versions
CN112200076B (en
Inventor
柯逍
叶宇
李悦洲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fuzhou University
Original Assignee
Fuzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fuzhou University filed Critical Fuzhou University
Priority to CN202011076008.8A priority Critical patent/CN112200076B/en
Publication of CN112200076A publication Critical patent/CN112200076A/en
Application granted granted Critical
Publication of CN112200076B publication Critical patent/CN112200076B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Closed-Circuit Television Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a method for carrying out multi-target tracking based on head and trunk characteristics, which comprises the following steps: step S1: acquiring a pedestrian detection result in the video, screening the result, and deleting an error detection result; step S2: preprocessing the screened detection results, and inputting the preprocessed detection results into a human body key point detection network to obtain all human body key points; step S3: screening the obtained key points of each pedestrian, selecting the key points of the head and the shoulders to combine to obtain the characteristics of the head and the trunk; step S4: and inputting the obtained head and trunk characteristics of the single pedestrian into a tracker for initialization, and further tracking the target. The invention can effectively extract the head and trunk characteristics in the detection frame, so that the effective information ratio acquired by the tracker during initialization is maximized.

Description

Method for carrying out multi-target tracking based on head and trunk characteristics
Technical Field
The invention relates to the field of computer vision, in particular to a method for multi-target tracking based on head and trunk characteristics.
Background
Multi-Object Tracking (MOT). The main task is to give an image sequence, find moving objects in the image sequence, correspond moving objects in different frames one to one (Identity), and then give the motion tracks of different objects. The mainstream framework adopted by the academia in the multi-target Tracking (MOT) problem at present is TBD (Tracking-by-Detection), that is, Tracking based on Detection, and in this mainstream Tracking framework, the multi-target Tracking problem is expressed as an association matching problem: if the detection result obtained from a certain frame is matched with the detection result obtained from the previous frame, the same target is identified.
With the continuous development of single-target trackers in recent years, a large number of trackers with good tracking effect and high running speed appear. In previous work, a single-target tracker has been applied to a multi-target tracking task and achieved certain effect, but the performance of the single-target tracker in a complex scene (such as an MOT19 data set) is not ideal, because in the complex scene, a large amount of redundant features and interference information are contained in a detection frame, and the tracking effect is greatly affected by initializing the tracker with the features and the information.
Disclosure of Invention
In view of this, the present invention provides a method for performing multi-target tracking based on head and torso features, which can effectively extract head and torso features in a detection frame, so as to maximize an effective information ratio obtained by a tracker during initialization.
In order to achieve the purpose, the invention adopts the following technical scheme:
a method for multi-target tracking based on head and trunk characteristics comprises the following steps:
step S1: acquiring a pedestrian detection result in the video, screening the result, and deleting an error detection result;
step S2: preprocessing the screened detection results, and inputting the preprocessed detection results into a human body key point detection network to obtain all human body key points;
step S3: screening the obtained key points of each pedestrian, selecting the key points of the head and the shoulders to combine to obtain the characteristics of the head and the trunk;
step S4: and inputting the obtained head and trunk characteristics of the single pedestrian into a tracker for initialization, and further tracking the target.
Further, the step S1 is specifically:
step S11, after each frame of image of the video is preprocessed, the image is detected by using a target detection network;
step S12: detecting information of each frame of the video by using a pedestrian detector to obtain a detection result R, and enabling R to be { K ═ Ki,PjDet _ x, det _ y, det _ w, det _ h, det _ c, i 1,2,. M, j 1,2,. N, representing the set of all detection results in a video sequence, where M represents the number of all image frames in a video sequence, N represents the number of all detected pedestrians in a frame of image, and K represents the number of detected pedestrians in a frame of imageiRepresenting the ith frame of a video sequence, PjRepresenting the jth pedestrian in the frame image, det _ x, det _ y, det _ w and det _ h respectively represent the x coordinate and the y coordinate of the upper left corner of the detection frame of the pedestrian and the width and the height of the detection frame, and det _ c represents the confidence coefficient of the detection frame;
step S13: let the confidence threshold value of pedestrian detection be TdThe pedestrian aspect ratio threshold is TrDeleting the detection results satisfying the following conditions:
det_c<Td or det_w/det_h>Tr
further, the step S2 is specifically:
step S21, the screened detection frame resize is set to a preset size;
step S22: preprocessing the image after resize, copying the preprocessed image, horizontally turning the copied image, and inputting the original image and the turned image into a human body key point detection network;
step S23: obtaining the output result S of the human body key point detection network, and enabling S to be { J ═ JzZ, where Z represents the number of human keypoints in the image; j. the design is a squarez- { jo int _ x, jo int _ y, jo int _ c } denotes the thz key points, wherein jo int _ x represents the x coordinate of the human body key point, jo int _ y represents the y coordinate of the human body key point, and jo int _ c represents the confidence coefficient of the human body key point;
step S24: let the human body key point detection result of the original image be SsrcLet the human body key point detection result of the flip image be SflipAnd fusing the two detection results.
Further, the pretreatment specifically comprises: firstly, removing sharp noise of an image by using Gaussian filtering, and then removing fine interference details by using a USM sharpening enhancement algorithm, wherein the calculation method comprises the following steps:
Figure BDA0002716672720000031
where output represents an output image, orign _ image represents an original image, gaus _ image represents an image after gaussian filtering, and ω represents a USM coefficient.
Further, the fusion method comprises the following steps:
Figure BDA0002716672720000032
Figure BDA0002716672720000033
Figure BDA0002716672720000034
wherein c issrcRepresenting confidence, x, of key points of the body in the original imagesrcX-coordinate, y representing key points of a human body in an original imagesrcY-coordinate representing key points of the human body in the original image, where cflipRepresenting confidence, x, of key points of a human body in a flip imageflipX-coordinate, y representing key points of a human body in a flip imageflipY-coordinate representing key points of the human body in the flip image, final _ x and final _ y, respectivelyAnd x and y coordinates of the human key points of which the final fusion is finished are represented, and final _ c represents the confidence coefficient of the human key points of which the final fusion is finished.
Further, the step S3 is specifically:
step S31: screening the selected key points, and adopting a screening scheme based on confidence coefficient to enable the detection confidence coefficient threshold of the human key points to be TkpDeleting the human body key points meeting the following requirements:
jo int_c<Tkp
step S32: combining the screened key points, enabling the screened human body key point set to be Q, traversing the set Q, sequencing the whole set Q according to the sizes of x coordinates and y coordinates, finding out the human body key points at the top, the bottom, the left and the right in the set, and obtaining the minimum rectangular convex hull in the set, wherein the image content in the rectangular convex hull is the head and trunk characteristics of the target.
Further, the step S4 is specifically:
step S41: let the set of targets in the tracked state be OtrackThe set comprises all targets in a tracking state from a first frame of the video to a current frame;
step S42: traversing a set of targets O in a state being trackedtrackPerforming IOU and OKS calculation on the newly obtained head and trunk characteristics and all targets in the set to confirm whether the target is in a tracking state;
step S43: calculating the fusion metric value of the bounding box of the tracked target and the newly obtained head and trunk characteristic bounding box in the target set in the tracked state, wherein the calculation method is as follows
Figure BDA0002716672720000041
If FF is greater than 0.5, the target is considered to exist, and the target does not need to be initialized to a new target again, and the step S1 is executed; otherwise the feature is identified as belonging to a new target, proceeding to the next step;
step S44: inputting new head and trunk characteristics into a tracker for initialization, and adding a target set O in a tracked statetrackGo to step S1.
Further, the method for calculating the IOU and OKS is as follows:
Figure BDA0002716672720000042
Figure BDA0002716672720000043
wherein A represents the area of the bounding box of the first object and B represents the area of the bounding box of the second object; the calculation method of the bounding box area is that the length of the rectangle is multiplied by the width of the rectangle; viszIndicates the visibility of the z-th keypoint (greater than 0 indicating visibility), dis indicates the Euclidean distance, scale, of the existing and detected human keypoints2The square root, σ, representing the size of the area occupied by these key pointszA normalization factor representing the z-th human body key point.
Compared with the prior art, the invention has the following beneficial effects:
1. the invention does not depend on the quality of the detection result, and can also be corrected through the human body key point detection network even if the detection frame has deviation with the group Truth in the data set;
2. according to the invention, the head trunk characteristics acquired through the human body key point detection network are used for target tracking, and the angle of monitoring scene shooting is considered, so that the head trunk is not easy to be shielded, and effective characteristics can be extracted for tracking even in a monitoring scene crowded with people;
3. the method adopts the human body key point detection network to obtain effective information contained in the head and trunk characteristics, and the effective information has larger average proportion in the image, thereby being more beneficial to the initialization and the subsequent tracking of the tracker.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
Referring to fig. 1, the present invention provides a method for multi-target tracking based on head and torso features, comprising the following steps:
step S1: acquiring a pedestrian detection result in the video, screening the result, and deleting an error detection result;
step S2: inputting the screened detection result into a human body key point detection network to obtain all human body key points; the dependence degree on the quality of the detection result can be reduced, and even if the quality of the detection result is poor, the detection result can be corrected through key points of a human body;
step S3: considering the shooting angle of a monitoring scene, as the head trunk is not easy to be shielded, effective features can be extracted for tracking even in the monitoring scene crowded with people, the obtained key points of each pedestrian are screened, and then the key points of the head and the shoulders are selected for combination to obtain the head trunk features;
step S4: the obtained head and trunk characteristics of the single pedestrian are input into the tracker to be initialized, and then the target is tracked.
In this embodiment, the step S1 includes the following steps:
step S11: after each frame of image of the video is preprocessed, a target detection network is utilized
Detecting the same;
step S12: detecting information of each frame of the video by using a pedestrian detector to obtain a detection result R, and enabling R to be { K ═ Ki,PjDet _ x, det _ y, det _ w, det _ h, det _ c, i 1,2,. M, j 1,2,. N, representing the set of all detection results in a video sequence, where M represents all images in a video sequenceNumber of frames, N represents the number of all detected pedestrians in one frame image, KiRepresenting the ith frame of a video sequence, PjRepresenting the jth pedestrian in the frame image, det _ x, det _ y, det _ w and det _ h respectively represent the x coordinate and the y coordinate of the upper left corner of the detection frame of the pedestrian and the width and the height of the detection frame, and det _ c represents the confidence coefficient of the detection frame;
step S13: let the confidence threshold value of pedestrian detection be TdThe pedestrian aspect ratio threshold is TrDeleting the detection results satisfying the following conditions:
det_c<Td or det_w/det_h>Tr
in this embodiment, step S2 specifically includes the following steps:
step S21: the screened detection box resize is to a size of 4:3, using a size of 344 × 258;
step S22: preprocessing the image after resize, firstly removing sharp noise of the image by using Gaussian filtering, wherein the Gaussian filtering is used because boundary information can be better reserved; then, a USM sharpening enhancement algorithm is used for removing fine interference details, and the calculation method is as follows:
Figure BDA0002716672720000061
where output represents an output image, orign _ image represents an original image, gaus _ image represents an image after gaussian filtering, and ω represents a USM coefficient. Copying the preprocessed output image, horizontally turning the copied image, and inputting the original image and the turned image into a human body key point detection network;
step S23: obtaining the output result S of the human body key point detection network, and enabling S to be { J ═ JzZ, where Z represents the number of human keypoints in the image. J. the design is a squarezThe z-th key point is represented by { jo int _ x, jo int _ y and jo int _ c }, wherein the jo int _ x represents the x coordinate of the human body key point, the jo int _ y represents the y coordinate of the human body key point, and the jo int _ c represents the confidence coefficient of the human body key point;
step S24: let the human body key point detection result of the original image be SsrcLet the human body key point detection result of the flip image be SflipAnd fusing the two detection results, wherein the step is to achieve more accurate human body key point coordinates, and the fusing method comprises the following steps:
Figure BDA0002716672720000071
Figure BDA0002716672720000072
Figure BDA0002716672720000073
wherein c issrcRepresenting confidence, x, of key points of the body in the original imagesrcX-coordinate, y representing key points of a human body in an original imagesrcY-coordinate representing key points of the human body in the original image, where cflipRepresenting confidence, x, of key points of a human body in a flip imageflipX-coordinate, y representing key points of a human body in a flip imageflipAnd the y coordinate represents the human body key point in the overturned image, final _ x and final _ y represent the x and y coordinates of the human body key point after final fusion is finished respectively, and final _ c represents the confidence coefficient of the human body key point after final fusion is finished.
In this embodiment, selecting human body key points with good tracking characteristics specifically includes: human body key points of the head (eyes, nose, ears) and two shoulders, the step S3 specifically includes the following steps:
step S31: screening the selected key points, and adopting a screening scheme based on confidence coefficient to enable the detection confidence coefficient threshold of the human key points to be TkpDeleting the human body key points meeting the following requirements:
jo int_c<Tkp
step S32: combining the screened key points, enabling the screened human body key point set to be Q, traversing the set Q, sorting the whole set Q according to the sizes of x coordinates and y coordinates respectively, finding out the human body key points at the top, the bottom, the left and the right in the set, and obtaining the minimum rectangular convex hull in the set. The image content within the rectangular convex hull is the head-torso feature of the target.
In this embodiment, step S4 specifically includes the following steps:
step S41: let the set of targets in the tracked state be OtrackThe set contains all the objects in the tracking state from the first frame of the video to the current frame.
Step S42: traversing a set of targets O in a state being trackedtrackAnd performing IOU (cross-over ratio) and OKS (object key point similarity) calculation on the newly obtained head and trunk features and all the objects in the set to confirm whether the object is in a tracking state or not, wherein the IOU and OKS calculation method comprises the following steps:
Figure BDA0002716672720000081
Figure BDA0002716672720000082
wherein A represents the area of the bounding box of the first object and B represents the area of the bounding box of the second object; the calculation method of the bounding box area is that the length of the rectangle is multiplied by the width of the rectangle; viszIndicates the visibility of the z-th keypoint (greater than 0 indicating visibility), dis indicates the Euclidean distance, scale, of the existing and detected human keypoints2The square root, σ, representing the size of the area occupied by these key pointszAnd (3) a normalization factor representing the z-th human body key point (the factor is obtained by calculating the standard deviation of all group Truth in the existing data set, and reflects the influence degree of the current key point on the whole body).
Step S43: calculating the fusion metric value of the bounding box of the tracked target and the newly obtained head and trunk characteristic bounding box in the target set in the tracked state, wherein the calculation method is as follows
Figure BDA0002716672720000091
If FF is greater than 0.5, the target is considered to exist, and the target does not need to be initialized to a new target again, and the step S11 is executed; otherwise the feature is deemed to be subject to a new target and the next step is taken.
Step S44: inputting new head and trunk characteristics into a tracker for initialization, and adding a target set O in a tracked statetrackGo to step S11.
The above description is only a preferred embodiment of the present invention, and all equivalent changes and modifications made in accordance with the claims of the present invention should be covered by the present invention.

Claims (8)

1. A method for multi-target tracking based on head and trunk characteristics is characterized by comprising the following steps:
step S1: acquiring a pedestrian detection result in the video, screening the result, and deleting an error detection result;
step S2: preprocessing the screened detection results, and inputting the preprocessed detection results into a human body key point detection network to obtain all human body key points;
step S3: screening the obtained key points of each pedestrian, selecting the key points of the head and the shoulders to combine to obtain the characteristics of the head and the trunk;
step S4: and inputting the obtained head and trunk characteristics of the single pedestrian into a tracker for initialization, and further tracking the target.
2. The method for multi-target tracking based on head-torso features of claim 1, wherein the step S1 is specifically performed by:
step S11, after each frame of image of the video is preprocessed, the image is detected by using a target detection network;
step S12: detecting information of each frame of the video by using a pedestrian detector to obtain a detection result R, and enabling R to be { K ═ Ki,PjDet _ x, det _ y, det _ w, det _ h, det _ c, i 1,2,. M, j 1,2,. N, representing the set of all detection results in a video sequence, where M represents the number of all image frames in a video sequence, N represents the number of all detected pedestrians in a frame of image, and K represents the number of detected pedestrians in a frame of imageiRepresenting the ith frame of a video sequence, PjRepresenting the jth pedestrian in the frame image, det _ x, det _ y, det _ w and det _ h respectively represent the x coordinate and the y coordinate of the upper left corner of the detection frame of the pedestrian and the width and the height of the detection frame, and det _ c represents the confidence coefficient of the detection frame;
step S13: let the confidence threshold value of pedestrian detection be TdThe pedestrian aspect ratio threshold is TrDeleting the detection results satisfying the following conditions:
det_c<Td or det_w/det_h>Tr
3. the method for multi-target tracking based on head-torso features of claim 1, wherein the step S2 is specifically performed by:
step S21, the screened detection frame resize is set to a preset size;
step S22: preprocessing the image after resize, copying the preprocessed image, horizontally turning the copied image, and inputting the original image and the turned image into a human body key point detection network;
step S23: obtaining the output result S of the human body key point detection network, and enabling S to be { J ═ JzZ, where Z represents the number of human keypoints in the image; j. the design is a squarez-join _ x, join _ y, join _ c represents the z-th keypoint, where join _ x represents the x-coordinate of a human keypoint, join _ y represents the y-coordinate of a human keypoint, and join _ c represents the confidence of this human keypoint;
step S24: let the human body key point detection result of the original image be SsrcTo turn over the human body of the imageThe key point detection result is SflipAnd fusing the two detection results.
4. The method for multi-target tracking based on head and torso features of claim 3, wherein the preprocessing is specifically: firstly, removing sharp noise of an image by using Gaussian filtering, and then removing fine interference details by using a USM sharpening enhancement algorithm, wherein the calculation method comprises the following steps:
Figure FDA0002716672710000021
where output represents an output image, orign _ image represents an original image, gaus _ image represents an image after gaussian filtering, and ω represents a USM coefficient.
5. The method for multi-target tracking based on head and torso features of claim 3, wherein the fusion method is as follows:
Figure FDA0002716672710000022
Figure FDA0002716672710000023
Figure FDA0002716672710000024
wherein c issrcRepresenting confidence, x, of key points of the body in the original imagesrcX-coordinate, y representing key points of a human body in an original imagesrcY-coordinate representing key points of the human body in the original image, where cflipRepresenting confidence, x, of key points of a human body in a flip imageflipX-coordinate, y representing key points of a human body in a flip imageflipRepresenting a flipped imageAnd the y coordinate of the middle human key point, final _ x and final _ y respectively represent the x and y coordinates of the final fused human key point, and final _ c represents the confidence coefficient of the final fused human key point.
6. The method for multi-target tracking based on head-torso features of claim 1, wherein the step S3 is specifically performed by:
step S31: screening the selected key points, and adopting a screening scheme based on confidence coefficient to enable the detection confidence coefficient threshold of the human key points to be TkpDeleting the human body key points meeting the following requirements:
joint_c<Tkp
step S32: combining the screened key points, enabling the screened human body key point set to be Q, traversing the set Q, sequencing the whole set Q according to the sizes of x coordinates and y coordinates, finding out the human body key points at the top, the bottom, the left and the right in the set, and obtaining the minimum rectangular convex hull in the set, wherein the image content in the rectangular convex hull is the head and trunk characteristics of the target.
7. The method for multi-target tracking based on head-torso features of claim 1, wherein the step S4 is specifically performed by:
step S41: let the set of targets in the tracked state be OtrackThe set comprises all targets in a tracking state from a first frame of the video to a current frame;
step S42: traversing a set of targets O in a state being trackedtrackPerforming IOU and OKS calculation on the newly obtained head and trunk characteristics and all targets in the set to confirm whether the target is in a tracking state;
step S43: calculating the fusion metric value of the bounding box of the tracked target and the newly obtained head and trunk characteristic bounding box in the target set in the tracked state, wherein the calculation method is as follows
Figure FDA0002716672710000031
If FF is greater than 0.5, the target is considered to exist, and the target does not need to be initialized to a new target again, and the step S1 is executed; otherwise the feature is identified as belonging to a new target, proceeding to the next step;
step S44: inputting new head and trunk characteristics into a tracker for initialization, and adding a target set O in a tracked statetrackGo to step S1.
8. The method for multi-target tracking based on head torso features of claim 7, wherein the IOU and OKS calculations are performed as follows:
Figure FDA0002716672710000041
Figure FDA0002716672710000042
wherein A represents the area of the bounding box of the first object and B represents the area of the bounding box of the second object; the calculation method of the bounding box area is that the length of the rectangle is multiplied by the width of the rectangle; viszIndicates the visibility of the z-th keypoint (greater than 0 indicating visibility), dis indicates the Euclidean distance, scale, of the existing and detected human keypoints2The square root, σ, representing the size of the area occupied by these key pointszA normalization factor representing the z-th human body key point.
CN202011076008.8A 2020-10-10 2020-10-10 Method for carrying out multi-target tracking based on head and trunk characteristics Active CN112200076B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011076008.8A CN112200076B (en) 2020-10-10 2020-10-10 Method for carrying out multi-target tracking based on head and trunk characteristics

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011076008.8A CN112200076B (en) 2020-10-10 2020-10-10 Method for carrying out multi-target tracking based on head and trunk characteristics

Publications (2)

Publication Number Publication Date
CN112200076A true CN112200076A (en) 2021-01-08
CN112200076B CN112200076B (en) 2023-02-21

Family

ID=74013249

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011076008.8A Active CN112200076B (en) 2020-10-10 2020-10-10 Method for carrying out multi-target tracking based on head and trunk characteristics

Country Status (1)

Country Link
CN (1) CN112200076B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117953015A (en) * 2024-03-26 2024-04-30 武汉工程大学 Multi-row person tracking method, system, equipment and medium based on video super-resolution

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108416258A (en) * 2018-01-23 2018-08-17 华侨大学 A kind of multi-human body tracking method based on human body model
CN109522843A (en) * 2018-11-16 2019-03-26 北京市商汤科技开发有限公司 A kind of multi-object tracking method and device, equipment and storage medium
WO2019232894A1 (en) * 2018-06-05 2019-12-12 中国石油大学(华东) Complex scene-based human body key point detection system and method
CN110674785A (en) * 2019-10-08 2020-01-10 中兴飞流信息科技有限公司 Multi-person posture analysis method based on human body key point tracking
CN111062239A (en) * 2019-10-15 2020-04-24 平安科技(深圳)有限公司 Human body target detection method and device, computer equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108416258A (en) * 2018-01-23 2018-08-17 华侨大学 A kind of multi-human body tracking method based on human body model
WO2019232894A1 (en) * 2018-06-05 2019-12-12 中国石油大学(华东) Complex scene-based human body key point detection system and method
CN109522843A (en) * 2018-11-16 2019-03-26 北京市商汤科技开发有限公司 A kind of multi-object tracking method and device, equipment and storage medium
CN110674785A (en) * 2019-10-08 2020-01-10 中兴飞流信息科技有限公司 Multi-person posture analysis method based on human body key point tracking
CN111062239A (en) * 2019-10-15 2020-04-24 平安科技(深圳)有限公司 Human body target detection method and device, computer equipment and storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
AHMAD JALAL AT EL.: "Human Body Parts Estimation and Detection for Physical Sports Movements", 《2019 2ND INTERNATIONAL CONFERENCE ON COMMUNICATION, COMPUTING AND DIGITAL SYSTEMS (C-CODE)》 *
周亮 等: "基于跟踪—关联模块的多目标跟踪方法研究", 《西南大学学报(自然科学版)》 *
宫法明等: "基于时空双分支网络的人体动作识别研究", 《计算机技术与发展》 *
马旋等: "基于人脸检测和关键点识别的快速人体组件划分", 《计算机应用与软件》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117953015A (en) * 2024-03-26 2024-04-30 武汉工程大学 Multi-row person tracking method, system, equipment and medium based on video super-resolution

Also Published As

Publication number Publication date
CN112200076B (en) 2023-02-21

Similar Documents

Publication Publication Date Title
WO2021196294A1 (en) Cross-video person location tracking method and system, and device
WO2020042419A1 (en) Gait-based identity recognition method and apparatus, and electronic device
US8131011B2 (en) Human detection and tracking system
US6826292B1 (en) Method and apparatus for tracking moving objects in a sequence of two-dimensional images using a dynamic layered representation
CN109543641B (en) Multi-target duplicate removal method for real-time video, terminal equipment and storage medium
CN110427905A (en) Pedestrian tracting method, device and terminal
CN110060277A (en) A kind of vision SLAM method of multiple features fusion
CN106885574A (en) A kind of monocular vision robot synchronous superposition method based on weight tracking strategy
CN110986969B (en) Map fusion method and device, equipment and storage medium
CN112215155A (en) Face tracking method and system based on multi-feature fusion
CN111160291B (en) Human eye detection method based on depth information and CNN
CN108257155B (en) Extended target stable tracking point extraction method based on local and global coupling
Tian et al. Scene Text Detection in Video by Learning Locally and Globally.
CN110555867B (en) Multi-target object tracking method integrating object capturing and identifying technology
CN109146920A (en) A kind of method for tracking target that insertion type is realized
CN108564598A (en) A kind of improved online Boosting method for tracking target
CN103617631B (en) A kind of tracking based on Spot detection
CN110895683A (en) Kinect-based single-viewpoint gesture and posture recognition method
CN111709893A (en) ORB-SLAM2 improved algorithm based on information entropy and sharpening adjustment
CN112200076B (en) Method for carrying out multi-target tracking based on head and trunk characteristics
Han et al. Accurate and robust vanishing point detection method in unstructured road scenes
CN114550219B (en) Pedestrian tracking method and device
CN106558065A (en) The real-time vision tracking to target is realized based on color of image and texture analysiss
CN114373203A (en) Picture archiving method and device, terminal equipment and computer readable storage medium
CN111640058B (en) Image fusion processing method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant