CN114923491A - Three-dimensional multi-target online tracking method based on feature fusion and distance fusion - Google Patents

Three-dimensional multi-target online tracking method based on feature fusion and distance fusion Download PDF

Info

Publication number
CN114923491A
CN114923491A CN202210512577.5A CN202210512577A CN114923491A CN 114923491 A CN114923491 A CN 114923491A CN 202210512577 A CN202210512577 A CN 202210512577A CN 114923491 A CN114923491 A CN 114923491A
Authority
CN
China
Prior art keywords
target
dimensional
distance
fusion
current frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210512577.5A
Other languages
Chinese (zh)
Inventor
达飞鹏
陈汶铭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202210512577.5A priority Critical patent/CN114923491A/en
Publication of CN114923491A publication Critical patent/CN114923491A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Automation & Control Theory (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a three-dimensional multi-target online tracking method based on feature fusion and distance fusion, which comprises the following steps of: firstly, acquiring detection data of each target by using a target detection algorithm; then extracting fusion characteristics and position characteristics of the target; then calculating the position distance and the characteristic distance between the track and the detection; then calculating the characteristic distance weight and calculating a final distance matrix according to the weight; and finally, carrying out data association by using greedy matching according to a final distance matrix between the tracks and the detection to obtain the multi-target tracking track of the current frame. By extracting and fusing multi-modal characteristics of the target, the characteristics which fully express the target properties in different environments can be extracted; according to the invention, the information of a plurality of modes is fused in different layers, so that the precision of multi-target tracking is improved; and the fusion process does not need manual intervention, so that a large amount of manual experiments are avoided for searching a proper threshold value, and a large amount of time is saved while the multi-target tracking precision is improved.

Description

Three-dimensional multi-target online tracking method based on feature fusion and distance fusion
Technical Field
The invention relates to the field of machine learning and computer vision, in particular to a three-dimensional multi-target online tracking method based on feature fusion and distance fusion, which is particularly suitable for outdoor complex tracking scenes of targets.
Background
With the continuous development of computer technology, the target tracking technology gradually becomes a key development technology direction in intelligent video analysis and multimedia applications (such as intelligent monitoring, motion analysis and automatic driving). A plurality of targets are detected and tracked in a complex scene, the target tracks are analyzed and predicted in real time, and richer data information is provided for practical application.
Due to the high degree of iteration of computer hardware and deep learning techniques, image and point cloud based target detection techniques are also becoming more mature. The accurate target detection result provides stronger data support for the multi-target tracking technology. In a real scene, the precision of the multi-target tracking algorithm is often interfered by the change of an external environment. Due to complex interaction and shielding among targets and changes of external environments of illumination and rainwater, the single modal characteristics of the targets are difficult to fully express the states and properties of the targets. Moreover, the high-dimensional semantic features and the low-dimensional physical features of the target are not suitable for direct fusion on a feature plane, and if more modal features are utilized, the problem of fusion level needs to be solved.
Aiming at the problems, the characteristics of different modes of the target are extracted, the weight of characteristic channels is adjusted through the extrusion excitation module and the characteristics are fused together, so that the characteristics of all the modes can play a role, and the final tracking precision is improved; modeling the motion of the targets to obtain the motion information of the targets so as to obtain the distance between the targets, and further fusing information of more modes in a decision level; the characteristic distance weight estimation network is learned in a data-driven mode, the weight of the characteristic distance between the targets is accurately given, the time consumed by manual experiments is reduced, and the efficiency is improved.
Disclosure of Invention
The purpose of the invention is as follows: in order to overcome the defects in the existing multi-target tracking technology, the invention provides a three-dimensional multi-target online tracking method based on feature fusion and distance fusion.
The technical scheme is as follows: in order to realize the purpose, the invention adopts the technical scheme that:
a three-dimensional multi-target online tracking method based on feature fusion and distance fusion comprises the following steps:
step 1: detecting vehicles of the current image sequence and the point cloud sequence by using a target detector to obtain detection data;
step 2: extracting two-dimensional characteristics, three-dimensional characteristics and position information of a target in a current frame;
and step 3: fusing the two-dimensional characteristic and the three-dimensional characteristic of the target in the current frame to obtain a fused characteristic of the target; predicting the position of the survival track in the current frame by using Kalman filtering;
and 4, step 4: calculating the Euclidean distance between the fusion characteristic distance and the position between the current frame and all targets of the historical track;
and 5: fusing the characteristic distance and Euclidean distance between the targets;
step 6: and performing data association according to the distance between the targets.
Further, the step 1 specifically includes the following steps:
selecting a target detector DSA-PV-RCNN to perform target detection on a currently input image sequence and a point cloud sequence, wherein the training of the target detector is based on a public data set KITTI, and the detection category is a vehicle; acquiring M detection data bounding box data of a current frame:
Figure BDA0003640004010000021
where α represents the angle of the target relative to the sensor, u 1 And v 1 Representing the horizontal and vertical coordinates, u, of the upper left corner of the two-dimensional bounding box in the image 2 And v 2 The horizontal and vertical coordinates of the lower right corner of the two-dimensional surrounding frame in the image are represented, h, w and l represent the height, width and length of the three-dimensional surrounding frame, x, y and z represent the coordinates of the central point of the three-dimensional surrounding frame in the point cloud, and r represents the rotation angle of the three-dimensional surrounding frame.
Further, the step 2 specifically includes the following steps:
step 2.1: according to detected data
Figure BDA0003640004010000022
Cutting down the picture of the corresponding position in the image, and zooming to the fixed size 224 multiplied by 224; then inputting the picture into a two-dimensional feature extraction network to extract two-dimensional features of the target
Figure BDA0003640004010000023
Performing this operation on all M objects can obtain a two-dimensional feature of each object, which can be expressed as
Figure BDA0003640004010000024
Step 2.2: according to (h) in the detection data i ,w i ,l i ,x i ,y i ,z i ,r i ) Extracting the points at the corresponding positions in the point cloud, inputting the points into a three-dimensional feature extraction network, and extracting the three-dimensional features of the target
Figure BDA0003640004010000025
Performing this operation on all M objects can obtain the three-dimensional features of each object, which can be expressed as
Figure BDA0003640004010000026
Step 2.3: according to (x) in the detected data i ,y i ,z i ) And obtaining the three-dimensional position information of the target.
Further, the step 3 specifically includes the following steps:
step 3.1: using squeeze actuated modules N 1 Adjusting the weight of the target two-dimensional feature channel to obtain a new target two-dimensional feature
Figure BDA0003640004010000027
Figure BDA0003640004010000028
Step 3.2: using a squeeze actuated module N 2 Adjusting the weight of the target three-dimensional characteristic channel to obtain a new target three-dimensional characteristic
Figure BDA0003640004010000029
Figure BDA00036400040100000210
Step 3.3: splicing the new two-dimensional characteristic and the three-dimensional characteristic of the target to obtain the fusion characteristic of the target
Figure BDA0003640004010000031
Figure BDA0003640004010000032
Figure BDA0003640004010000033
Representing feature splicing;
step 3.4: based on detection data in the historical track set, a state equation of the tracked target is established, and the state equation consists of 6 variables:
Figure BDA0003640004010000034
(x, y, z) represents three-dimensional position information of the object,
Figure BDA0003640004010000035
a difference between the three-dimensional position information representing the target and the three-dimensional position information of the previous frame; then, Kalman filtering prediction is applied to obtain three-dimensional position information of the track in the current frame
Figure BDA0003640004010000036
Further, the step 4 specifically includes the following steps:
step 4.1: estimating a network N by a characteristic distance 3 Calculating the characteristic distance between the ith target of the current frame and the jth track in the historical tracks
Figure BDA0003640004010000037
Figure BDA0003640004010000038
Figure BDA0003640004010000039
Represents the fusion characteristic of the ith target of the current frame,
Figure BDA00036400040100000310
showing the fusion characteristics of the jth historical track in all the historical tracks up to the t-1 frame. Calculating the characteristic distance between the historical track and all the targets of the current frame to obtain a characteristic distance matrix
Figure BDA00036400040100000311
M represents the target number of the current frame, and N represents the target number of the adjacent previous frame;
step 4.2: according to the predicted position obtained in the step 3.4 and the detection position extracted in the step 2.3, the Euclidean distance can be calculated
Figure BDA00036400040100000312
Figure BDA00036400040100000313
(x i ,y i ,z i ) Three-dimensional position information of the ith object in the current frame is represented,
Figure BDA00036400040100000314
representing the predicted position of the jth historical track in the current frame; the Euclidean distance between the historical track and all the targets of the current frame is calculated to obtain an Euclidean distance matrix
Figure BDA00036400040100000315
M represents the target number of the current frame, and N represents the target number of the adjacent previous frame;
further, the step 5 specifically includes the following steps:
step 5.1: estimating network N by feature distance weights 4 Calculating a characteristic distance weight A:
Figure BDA00036400040100000316
Figure BDA00036400040100000317
target fusion characteristics calculated in step 3.3;
step 5.2: and according to the weight, weighting and fusing the characteristic distance and Euclidean distance between the targets:
Figure BDA0003640004010000041
where denotes multiplication of corresponding elements.
Further, the step 6 specifically includes the following steps:
obtaining a distance matrix D between the current frame target and the historical track according to the step 5 final Selecting the column with the lowest distance from the first row of the matrix, if the distance is less than 2, regarding that the row and the column represent the corresponding targets, assigning the same Identity (ID), and setting the columns to be 100 to prevent the secondary matching; if it is higher than 2, the object represented by the row is considered to have no matching history object, and a new ID is assigned.
Has the advantages that: the three-dimensional multi-target online tracking method based on feature fusion and distance fusion has the following advantages.
1) According to the accurate detection result of DSA-PV-RCNN, besides the position information of the bounding box, the extracted high-dimensional semantic features can more fully express the target features.
2) The extrusion excitation module based on the attention mechanism adjusts the channel weight of the two-dimensional characteristic and the three-dimensional characteristic of the target, so that a more important part in the original characteristic is emphasized, and a more important role is played in the subsequent calculation process.
3) The fusion characteristics after the multi-modal characteristics are fused can more fully express the state and the property of the target, and the robustness of the algorithm in a complex scene is enhanced.
4) The Euclidean distance obtained by target motion modeling is fused in a decision level, and the robustness of the algorithm in a complex scene is further enhanced by utilizing information of more modes of the target.
5) The characteristic distance weight estimation network is learned in a data-driven mode, time consumed by manually searching for proper parameters is reduced, and the final tracking effect is improved.
Drawings
FIG. 1 is an overall flow chart of a three-dimensional multi-target online tracking method based on feature fusion and distance fusion provided by the invention;
FIG. 2 is a network architecture diagram of a two-dimensional feature extraction network provided by the present invention;
FIG. 3 is a network architecture diagram of a three-dimensional feature extraction network provided by the present invention;
FIG. 4 is a network architecture diagram of a squeeze activation module for use with the present invention;
FIG. 5 is an algorithmic flow diagram of a fusion module provided by the present invention;
fig. 6 is a network structure diagram of a characteristic distance estimation network provided by the present invention.
Fig. 7 is a diagram of a feature distance weight estimation network architecture provided by the present invention.
FIG. 8 is a flow chart of a data correlation algorithm used by the present invention.
FIG. 9 is a data set tracking result of the three-dimensional multi-target online tracking method based on feature fusion and distance fusion provided by the invention.
Fig. 10 is a visualization diagram of a tracking result of the three-dimensional multi-target online tracking method based on feature fusion and distance fusion, provided by the invention, wherein the upper diagram is a visualization result of a 115 th frame of a KITTI data set video sequence 0, and the lower diagram is a visualization result of a 100 th frame of a video sequence 2.
Detailed Description
The present invention will be further described with reference to the accompanying drawings.
As shown in the figure, the three-dimensional multi-target online tracking method based on feature fusion and distance fusion comprises the following steps:
step 1: detecting vehicles of the current image sequence and the point cloud sequence by using a target detector to obtain detection data;
step 2: extracting two-dimensional characteristics, three-dimensional characteristics and position information of a target in a current frame;
and step 3: fusing the two-dimensional characteristic and the three-dimensional characteristic of the target in the current frame to obtain a fused characteristic of the target; predicting the position of the survival track in the current frame by using Kalman filtering;
and 4, step 4: calculating the Euclidean distance between the fusion characteristic distance and the position between the current frame and all targets of the historical track;
and 5: fusing characteristic distances and Euclidean distances between targets;
and 6: and performing data association according to the distance between the targets.
In this embodiment, step 1 specifically includes the following steps:
selecting a target detector DSA-PV-RCNN to perform target detection on a currently input image sequence and a point cloud sequence, wherein the training of the target detector is based on a public data set KITTI, and the detection category is a vehicle; acquiring M detection data bounding box data of a current frame:
Figure BDA0003640004010000051
where α represents the angle of the target relative to the sensor, u 1 And v 1 Representing the horizontal and vertical coordinates, u, of the upper left corner of the two-dimensional bounding box in the image 2 And v 2 The horizontal and vertical coordinates of the lower right corner of the two-dimensional surrounding frame in the image are represented, h, w and l represent the height, width and length of the three-dimensional surrounding frame, x, y and z represent the coordinates of the central point of the three-dimensional surrounding frame in the point cloud, and r represents the rotation angle of the three-dimensional surrounding frame.
In this embodiment, step 2 specifically includes the following steps:
step 2.1: according to detected data
Figure BDA0003640004010000052
Image processing methodThe picture at the corresponding position is cut off and is scaled to the fixed size 224 × 224; then inputting the picture into a two-dimensional feature extraction network to extract two-dimensional features of the target
Figure BDA0003640004010000061
Performing this operation on all M targets may result in a two-dimensional feature for each target, which may be expressed as
Figure BDA0003640004010000062
Step 2.2: according to (h) in the detection data i ,w i ,l i ,x i ,y i ,z i ,r i ) Extracting the point of the corresponding position in the point cloud, inputting the point into a three-dimensional feature extraction network, and extracting the three-dimensional feature of the target
Figure BDA0003640004010000063
Performing this operation on all M objects can obtain the three-dimensional features of each object, which can be expressed as
Figure BDA0003640004010000064
Step 2.3: according to (x) in the detection data i ,y i ,z i ) And obtaining the three-dimensional position information of the target.
In this embodiment, step 3 specifically includes the following steps:
step 3.1: using a squeeze actuated module N 1 Adjusting the weight of the target two-dimensional feature channel to obtain a new target two-dimensional feature
Figure BDA0003640004010000065
Figure BDA0003640004010000066
Step 3.2: using a squeeze actuated module N 2 Adjusting the weight of the target three-dimensional characteristic channel to obtain a new target three-dimensional characteristic
Figure BDA0003640004010000067
Figure BDA0003640004010000068
Step 3.3: splicing the new two-dimensional characteristic and the three-dimensional characteristic of the target to obtain the fusion characteristic of the target
Figure BDA0003640004010000069
Figure BDA00036400040100000610
Figure BDA00036400040100000611
Representing feature splicing;
step 3.4: establishing a state equation of the tracked target based on detection data in the historical track set, wherein the state equation comprises 6 variables:
Figure BDA00036400040100000612
(x, y, z) represents three-dimensional position information of the object,
Figure BDA00036400040100000613
a difference between the three-dimensional position information representing the target and the three-dimensional position information of the previous frame; then, Kalman filtering prediction is applied to obtain three-dimensional position information of the track in the current frame
Figure BDA00036400040100000614
In this embodiment, step 4 specifically includes the following steps:
step 4.1: estimating a network N by a characteristic distance 3 Calculating the characteristic distance between the ith target of the current frame and the jth track in the historical tracks
Figure BDA00036400040100000615
Figure BDA0003640004010000071
Figure BDA0003640004010000072
Represents the fusion characteristic of the ith target of the current frame,
Figure BDA0003640004010000073
showing the fusion characteristics of the jth historical track in all the historical tracks up to the t-1 th frame. Calculating the characteristic distance between the historical track and all the targets of the current frame to obtain a characteristic distance matrix
Figure BDA0003640004010000074
M represents the target number of the current frame, and N represents the target number of the adjacent previous frame;
step 4.2: according to the predicted position obtained in the step 3.4 and the detection position extracted in the step 2.3, the Euclidean distance can be calculated
Figure BDA0003640004010000075
Figure BDA0003640004010000076
(x i ,y i ,z i ) Three-dimensional position information of the ith object in the current frame is represented,
Figure BDA0003640004010000077
representing the predicted position of the jth historical track in the current frame; the Euclidean distance between the historical track and all targets of the current frame is calculated to obtain a Euclidean distance matrix
Figure BDA0003640004010000078
M represents the number of objects in the current frame, and N represents the number of objects in the adjacent previous frame;
In this embodiment, step 5 specifically includes the following steps:
step 5.1: estimating network N by feature distance weights 4 Calculating a characteristic distance weight A:
Figure BDA0003640004010000079
Figure BDA00036400040100000710
target fusion characteristics calculated in step 3.3;
step 5.2: and according to the weight, weighting and fusing the characteristic distance and the Euclidean distance between the targets:
Figure BDA00036400040100000711
where denotes multiplication of corresponding elements.
In this embodiment, step 6 specifically includes the following steps:
according to the distance matrix D between the current frame target and the historical track obtained in the step 5 final Selecting the column with the lowest distance from the first row of the matrix, if the distance is less than 2, regarding that the row and the column represent the corresponding targets, assigning the same Identity (ID), and setting the columns to be 100 to prevent the secondary matching; if it is higher than 2, the object represented by the row is considered to have no matching history object, and a new ID is assigned.
Examples
The invention relates to a three-dimensional multi-target online tracking method based on feature fusion and distance fusion.A DSA-PV-RCNN target detector is based on to obtain detection data on an image sequence and a point cloud sequence, and the type of a target of multi-target tracking is selected as a vehicle; the training dataset for the detector and tracker is the KITTI dataset, which includes 8008 pictures 24070 vehicles. The backbone network of the two-dimensional feature extraction network of the algorithm is VGGNet, the backbone network of the three-dimensional feature extraction network is PointNet, and the training losses of the distance estimation network and the feature distance weight estimation network are contrast loss and foldout loss.
Experiment: and testing the algorithm on the KITTI data set so as to verify the multi-target tracking effectiveness of the method provided by the method.
The above description is only of the preferred embodiments of the present invention, and it should be noted that: it will be apparent to those skilled in the art that various modifications and adaptations can be made without departing from the principles of the invention and these are intended to be within the scope of the invention.

Claims (7)

1. A three-dimensional multi-target online tracking method based on feature fusion and distance fusion is characterized in that:
step 1: detecting vehicles of the current image sequence and the point cloud sequence by using a target detector to obtain detection data;
step 2: extracting two-dimensional characteristics, three-dimensional characteristics and position information of a target in a current frame;
and step 3: fusing the two-dimensional characteristic and the three-dimensional characteristic of the target in the current frame to obtain a fused characteristic of the target; predicting the position of the survival track in the current frame by using Kalman filtering;
and 4, step 4: calculating the Euclidean distance between the fusion characteristic distance and the position between the current frame and all targets of the historical track;
and 5: fusing the characteristic distance and Euclidean distance between the targets;
and 6: and performing data association according to the distance between the targets.
2. The three-dimensional multi-target online tracking method based on feature fusion and distance fusion as claimed in claim 1, wherein the specific method of step 1 is as follows:
selecting a target detector DSA-PV-RCNN to perform target detection on a currently input image sequence and a point cloud sequence, wherein the training of the target detector is based on a public data set KITTI, and the detection category is a vehicle; acquiring M detection data bounding box data of a current frame:
Figure FDA0003640004000000011
where α represents the angle of the target relative to the sensor, u 1 And v 1 Representing the horizontal and vertical coordinates, u, of the upper left corner of the two-dimensional bounding box in the image 2 And v 2 The horizontal and vertical coordinates of the lower right corner of the two-dimensional surrounding frame in the image are represented, h, w and l represent the height, width and length of the three-dimensional surrounding frame, x, y and z represent the coordinates of the central point of the three-dimensional surrounding frame in the point cloud, and r represents the rotation angle of the three-dimensional surrounding frame.
3. The three-dimensional multi-target online tracking method based on feature fusion and distance fusion as claimed in claim 2, wherein the specific method of step 2 is:
step 2.1: according to detected data
Figure FDA0003640004000000012
Cutting down the picture at the corresponding position in the image, and zooming the picture to be 224 multiplied by 224 with a fixed size; then inputting the picture into a two-dimensional feature extraction network to extract two-dimensional features of the target
Figure FDA0003640004000000013
Performing this operation on all M objects yields a two-dimensional feature for each object, denoted as
Figure FDA0003640004000000014
Step 2.2: according to (h) in the detection data i ,w i ,l i ,x i ,y i ,z i ,r i ) Extracting the points at the corresponding positions in the point cloud, inputting the points into a three-dimensional feature extraction network, and extracting the three-dimensional features of the target
Figure FDA0003640004000000015
Performing this operation on all M objects obtains the three-dimensional features of each object, expressed as
Figure FDA0003640004000000016
Step 2.3: according to (x) in the detected data i ,y i ,z i ) And obtaining the three-dimensional position information of the target.
4. The three-dimensional multi-target online tracking method based on feature fusion and distance fusion as claimed in claim 3, wherein the specific method of step 3 is as follows:
step 3.1: using squeeze actuated modules N 1 Adjusting the weight of the target two-dimensional feature channel to obtain a new target two-dimensional feature
Figure FDA0003640004000000021
Figure FDA0003640004000000022
Step 3.2: using squeeze actuated modules N 2 Adjusting the weight of the target three-dimensional characteristic channel to obtain a new target three-dimensional characteristic
Figure FDA0003640004000000023
Figure FDA0003640004000000024
Step 3.3: splicing the new two-dimensional characteristics and the three-dimensional characteristics of the target to obtain the fusion characteristics of the target
Figure FDA0003640004000000025
Figure FDA0003640004000000026
Figure FDA0003640004000000027
Representing feature splicing;
step 3.4: establishing a state equation of the tracked target based on detection data in the historical track set, wherein the state equation comprises 6 variables:
Figure FDA0003640004000000028
(x, y, z) represents three-dimensional position information of the object,
Figure FDA0003640004000000029
a difference between the three-dimensional position information representing the target and the three-dimensional position information of the previous frame; then, Kalman filtering prediction is applied to obtain three-dimensional position information of the track in the current frame
Figure FDA00036400040000000210
5. The three-dimensional multi-target online tracking method based on feature fusion and distance fusion as claimed in claim 4, wherein the specific method of step 4 is as follows:
step 4.1: estimating a network N by a characteristic distance 3 Calculating the characteristic distance between the ith target of the current frame and the jth track in the historical tracks
Figure FDA00036400040000000211
Figure FDA00036400040000000212
Figure FDA00036400040000000213
Represents the fusion characteristic of the ith target of the current frame,
Figure FDA00036400040000000214
showing the fusion characteristics of the jth historical track in all the historical tracks up to the t-1 th frame; calculating the characteristic distance between the historical track and all the targets of the current frame to obtain a characteristic distance matrix
Figure FDA00036400040000000215
M represents the target number of the current frame, and N represents the target number of the adjacent previous frame;
step 4.2: according to the predicted position obtained in the step 3.4 and the detection position extracted in the step 2.3, the Euclidean distance is calculated
Figure FDA0003640004000000031
Figure FDA0003640004000000032
(x i ,y i ,z i ) Three-dimensional position information of the ith object in the current frame is represented,
Figure FDA0003640004000000033
representing the predicted position of the jth historical track in the current frame; calculating Euclidean distances between the historical track and all targets of the current frame to obtain an Euclidean distance matrix
Figure FDA0003640004000000034
M denotes the number of objects of the current frame and N denotes the number of objects of the adjacent previous frame.
6. The three-dimensional multi-target online tracking method based on feature fusion and distance fusion as claimed in claim 5, wherein the specific method of step 5 is:
step 5.1: estimating network N by feature distance weights 4 Calculating a characteristic distance weight A:
Figure FDA0003640004000000035
Figure FDA0003640004000000036
the target fusion characteristics are obtained by calculation in the step 3.3;
step 5.2: and according to the weight, weighting and fusing the characteristic distance and Euclidean distance between the targets:
Figure FDA0003640004000000037
where denotes multiplication of corresponding elements.
7. The three-dimensional multi-target online tracking method based on feature fusion and distance fusion as claimed in claim 6, wherein the specific method of step 6 is as follows:
obtaining a distance matrix D between the current frame target and the historical track according to the step 5 final Selecting the column with the lowest distance from the first row of the matrix, if the distance is less than 2, regarding that the row and the column represent the corresponding targets, assigning the same Identity (ID), and setting the columns to be 100 to prevent the secondary matching; if it is higher than 2, the target represented by the row is considered to have no history target matching with it, and a new ID is assigned to it.
CN202210512577.5A 2022-05-12 2022-05-12 Three-dimensional multi-target online tracking method based on feature fusion and distance fusion Pending CN114923491A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210512577.5A CN114923491A (en) 2022-05-12 2022-05-12 Three-dimensional multi-target online tracking method based on feature fusion and distance fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210512577.5A CN114923491A (en) 2022-05-12 2022-05-12 Three-dimensional multi-target online tracking method based on feature fusion and distance fusion

Publications (1)

Publication Number Publication Date
CN114923491A true CN114923491A (en) 2022-08-19

Family

ID=82808665

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210512577.5A Pending CN114923491A (en) 2022-05-12 2022-05-12 Three-dimensional multi-target online tracking method based on feature fusion and distance fusion

Country Status (1)

Country Link
CN (1) CN114923491A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116523970A (en) * 2023-07-05 2023-08-01 之江实验室 Dynamic three-dimensional target tracking method and device based on secondary implicit matching
CN117237401A (en) * 2023-11-08 2023-12-15 北京理工大学前沿技术研究院 Multi-target tracking method, system, medium and equipment for fusion of image and point cloud

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116523970A (en) * 2023-07-05 2023-08-01 之江实验室 Dynamic three-dimensional target tracking method and device based on secondary implicit matching
CN116523970B (en) * 2023-07-05 2023-10-20 之江实验室 Dynamic three-dimensional target tracking method and device based on secondary implicit matching
CN117237401A (en) * 2023-11-08 2023-12-15 北京理工大学前沿技术研究院 Multi-target tracking method, system, medium and equipment for fusion of image and point cloud
CN117237401B (en) * 2023-11-08 2024-02-13 北京理工大学前沿技术研究院 Multi-target tracking method, system, medium and equipment for fusion of image and point cloud

Similar Documents

Publication Publication Date Title
JP7147078B2 (en) Video frame information labeling method, apparatus, apparatus and computer program
CN109800689B (en) Target tracking method based on space-time feature fusion learning
CN110287826B (en) Video target detection method based on attention mechanism
US20220366576A1 (en) Method for target tracking, electronic device, and storage medium
CN112836639A (en) Pedestrian multi-target tracking video identification method based on improved YOLOv3 model
US20170316569A1 (en) Robust Anytime Tracking Combining 3D Shape, Color, and Motion with Annealed Dynamic Histograms
CN111127513A (en) Multi-target tracking method
CN114923491A (en) Three-dimensional multi-target online tracking method based on feature fusion and distance fusion
CN111797771B (en) Weak supervision video behavior detection method and system based on iterative learning
CN112489081B (en) Visual target tracking method and device
CN111626128A (en) Improved YOLOv 3-based pedestrian detection method in orchard environment
CN112884742B (en) Multi-target real-time detection, identification and tracking method based on multi-algorithm fusion
CN108564598B (en) Improved online Boosting target tracking method
CN110781262A (en) Semantic map construction method based on visual SLAM
CN113052873B (en) Single-target tracking method for on-line self-supervision learning scene adaptation
CN112069969A (en) Method and system for tracking highway monitoring video mirror-crossing vehicle
CN108734109B (en) Visual target tracking method and system for image sequence
CN112836640A (en) Single-camera multi-target pedestrian tracking method
CN113569882A (en) Knowledge distillation-based rapid pedestrian detection method
CN110827320B (en) Target tracking method and device based on time sequence prediction
CN114169425A (en) Training target tracking model and target tracking method and device
CN113129336A (en) End-to-end multi-vehicle tracking method, system and computer readable medium
Wang et al. Improving target detection by coupling it with tracking
CN117011346A (en) Blower image registration algorithm
CN101567088B (en) Method and device for detecting moving object

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination