CN111161315A - Multi-target tracking method and system based on graph neural network - Google Patents

Multi-target tracking method and system based on graph neural network Download PDF

Info

Publication number
CN111161315A
CN111161315A CN201911312114.9A CN201911312114A CN111161315A CN 111161315 A CN111161315 A CN 111161315A CN 201911312114 A CN201911312114 A CN 201911312114A CN 111161315 A CN111161315 A CN 111161315A
Authority
CN
China
Prior art keywords
neural network
appearance
motion
similarity matrix
edge
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911312114.9A
Other languages
Chinese (zh)
Other versions
CN111161315B (en
Inventor
蒋婷婷
高旭
李佳河
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Peking University
Original Assignee
Peking University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Peking University filed Critical Peking University
Priority to CN201911312114.9A priority Critical patent/CN111161315B/en
Publication of CN111161315A publication Critical patent/CN111161315A/en
Application granted granted Critical
Publication of CN111161315B publication Critical patent/CN111161315B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

The application discloses a multi-target tracking method and system based on a graph neural network, comprising the following steps: preprocessing a training set to obtain a node set, an edge set and a global variable; inputting the node set, the edge set and the global variable to a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix; training a graph neural network according to the loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network; processing the data set by using the trained neural network to obtain a similarity matrix; and calculating the similarity matrix by using a matching algorithm to obtain a matching result of the target in the data set. The node set, the edge set and the global variables can be updated by using the graph neural network and through an optimization algorithm and a loss function, and the global variables are introduced and updated outside the nodes and the edges, so that the graph neural network can capture the global information and improve the performance of multi-target tracking.

Description

Multi-target tracking method and system based on graph neural network
Technical Field
The application relates to the field of multi-target tracking, in particular to a multi-target tracking method and system based on a graph neural network.
Background
With the development of multimedia technology and the popularization of video acquisition equipment, the output of multimedia products such as videos is larger and larger, and the effect of the multimedia products in daily life is larger and larger. Tracking multiple targets in a video is a basic and important task in the fields of motion analysis, human-computer interaction, video monitoring (abnormal behavior recognition), automatic driving and the like. The video multi-target tracking technology has great research value in theoretical research and practical application.
The tracking task is divided into: single target tracking tasks and multi-target tracking tasks. The single-target tracking task is to track one initially specified target, and the multi-target tracking task is to track all targets in a field. The following difficulties to be overcome by the tracking task are: the tracking target is shielded, and the appearance characteristic is changed due to motion blurring. Although single target tracking methods can be applied directly to multi-target tracking tasks, simply applying the approach is not feasible. Compared with a single-target tracking task, the number of tracking targets in the multi-target tracking task is changed. The multi-target tracking method needs to determine the number of tracking targets in the current frame, which targets leave the visual field and which targets enter the visual field. Compared with a single-target tracking task, the targets in the multi-target tracking task are shielded more frequently, and simultaneously, the targets with similar appearances are more. If a single target tracking method is simply used for a multi-target tracking task, the tracking performance may be poor.
Video multi-target tracking methods are generally classified into off-line multi-target tracking, on-line multi-target tracking, and approximate on-line multi-target tracking. The off-line multi-target tracking method completes multi-target tracking by utilizing the information of all frames in the video. However, in practical applications, it is more necessary to track a plurality of targets in real time, rather than waiting for the entire video to be obtained and then tracking. At this time, an online multi-target tracking method is required. The on-line multi-target tracking method is to utilize the information of the current frame and the previous frame of the current frame to carry out multi-target tracking. Because the off-line multi-target tracking method utilizes more information, the tracking performance of the off-line multi-target tracking method is generally better than that of the on-line multi-target tracking method. In order to meet the application requirements of tracking performance and real-time performance, an approximate online multi-target tracking method is produced. The approximate on-line multi-target tracking method is a multi-target tracking method between an off-line multi-target tracking method and an on-line multi-target tracking method. Compared with an on-line multi-target tracking method, the approximate on-line multi-target tracking method also utilizes information of a period of time after the current frame for tracking. Compared with an online multi-target tracking method, the approximate online multi-target tracking method utilizes more information for tracking, and can obtain better tracking performance; compared with an off-line multi-target tracking method, the approximate on-line multi-target tracking method does not need to track after acquiring the whole video, and can carry out real-time tracking with a certain time delay.
The strategy of the multi-target tracking method comprises the following steps: detection-then-tracking strategies and model-less strategies. The detection-then-tracking strategy is as follows: firstly, detecting a target in each frame in a video, and giving the target in a form of a labeling frame; and then tracking is performed. The purpose of the multi-target tracking task adopting the detection-before-tracking strategy is as follows: and connecting the labeling frames belonging to the same target in different frames. The model-free strategy is: only the labeling frame of the target to be tracked in the first frame is given, and then the tracking method needs to perform detection first and then perform tracking. During the tracking process, a target may leave the field of view at any time, and a new target may enter the field of view. The modeless strategy, while more conforming to the tracked scenario, still requires that an initial annotation box be given for the newly entered target. Although less labeled boxes need to be provided for the wood model, an initial box is always needed for tracking the target. Therefore, the multi-target tracking method mostly adopts a strategy of firstly detecting and then tracking.
The existing multi-target tracking method mainly utilizes three types of characteristic information for tracking: appearance features, motion features, and interaction features. The appearance features mainly comprise traditional features and depth features. Wherein the traditional appearance characteristics include: color histogram (RGB histogram, HSV histogram, etc.), shape and texture. The depth features are extracted through a pre-trained deep neural network. The motion characteristics refer to modeling the motion speed of the target, and the motion models comprise a linear motion model and a nonlinear model. Wherein the linear model assumes that the object motion velocity is constant. Interactive features refer to modeling the relationship between multiple objects in a video, e.g., the direction and speed of motion of a group is the same. The appearance features, motion features and interaction features can provide useful information for the multi-target tracking method.
Multi-target tracking methods include two broad categories: the method comprises the steps of carrying out multi-target tracking on a traditional method and a multi-target tracking method based on a deep neural network. Because of the expansion of the basic data set and the improvement of the computer computing power, the deep neural network is put into application. The performance of deep neural network based methods has outperformed traditional methods, and even humans, on many visual tasks. For example, on the task of image classification, the classification accuracy of the deep neural network-based method is higher than that of human beings. Therefore, the multi-target tracking method based on the deep neural network also becomes the mainstream of the current multi-target tracking method. The multi-target tracking method mostly adopts a strategy of firstly detecting and then tracking, and a multi-target tracking task is converted into a bipartite graph matching problem. Thus, the multi-target tracking method needs to accurately calculate the similarity between two labeling frames. The similarity of the two labeling frames is calculated based on the characteristics of the labeling frames, and the more accurate characteristics can obtain more accurate similarity, so that a more accurate similarity matrix is obtained to obtain better tracking performance. In order to obtain better characteristics, the multi-target tracking method based on the deep neural network is integrated into the design of the multi-target tracking method based on a single-target tracker, an attention mechanism, LSTM and Re-ID characteristics and the like.
The multi-target tracking task can be completed by a graph model: minimum cost streams and conditional random fields. However, in these graph models, the structure of the graph is fixed, so that the nodes and edges are also fixed, which results in that the characteristic inaccurate nodes and edges cannot be corrected.
In view of the foregoing, it is desirable to provide a method and system for online or near online multi-target tracking capable of updating nodes and edges with high accuracy.
Disclosure of Invention
In order to solve the problems, the application provides a multi-target tracking method and system based on a graph neural network.
In one aspect, the application provides a multi-target tracking method based on a graph neural network, including:
preprocessing a training set to obtain a node set, an edge set and a global variable;
inputting the node set, the edge set and the global variable to a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix;
training the graph neural network according to a loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network;
processing the data set by using the trained neural network to obtain a similarity matrix;
and calculating the similarity matrix by using a matching algorithm to obtain a matching result of the targets in the data set.
Preferably, the preprocessing the training set to obtain a node set, an edge set and a global variable includes:
dividing the training set into a first training set and a second training set;
extracting depth features of all the labeled boxes in the frame of the first training set by using a preprocessing neural network to obtain an appearance node set;
calculating the overlapping degree of the labeling frames of all the adjacent frames in the first training set to obtain an edge set between the two labeling frames;
extracting the position, size and movement data of the labeling frames in all frames in the first training set to obtain the motion characteristics of each labeling frame;
normalizing all motion characteristics to obtain a motion node set;
one-dimensional vectors are initialized randomly as global variables.
Preferably, the inputting the node set, the edge set, and the global variable into a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix includes:
inputting an appearance node set, an edge set and a global variable to an appearance feature map neural network to obtain an appearance similarity matrix;
and inputting the motion node set, the edge set and the global variable to a motion characteristic graph neural network to obtain a motion similarity matrix.
Preferably, the inputting the appearance node set, the edge set and the global variable into an appearance feature map neural network to obtain an appearance similarity matrix includes:
inputting an edge set, two appearance nodes corresponding to each edge in the input edge set and a global variable to a first edge neural network to obtain a first updating edge;
inputting the first updating edge, two appearance nodes connected through the first updating edge and a global variable to a node neural network to obtain a first updating node;
fusing all the obtained first updating edges to obtain a second updating edge;
fusing all sending nodes in the appearance nodes and the obtained first updating node to obtain a second updating node;
fusing all the first updating edges to obtain a second updating edge;
inputting the global variable, a second updating node and a second updating edge to a first global neural network to obtain a first updating global variable;
and inputting the updated global variable, all the sending nodes in the appearance nodes, the first updated node and the first updated edge to a second edge neural network to obtain an appearance similarity matrix.
Preferably, the inputting the motion node set, the edge set, and the global variable to the motion feature map neural network to obtain a motion similarity matrix includes:
fusing all edges in the edge set to obtain a third updating edge;
fusing all motion nodes in the motion node set to obtain a third update node;
inputting the global variable, a third updating edge and a third updating node to a second global neural network to obtain a second updating global variable;
inputting an edge set, two motion nodes corresponding to each edge in the input edge set, and the second updated global variable to a third neural network to obtain a motion similarity matrix.
Preferably, the using optimization algorithm, training the graph neural network according to a loss function, the appearance similarity matrix and the motion similarity matrix, determining setting parameters of the graph neural network, and then training to obtain the trained graph neural network, includes:
training an appearance feature map neural network according to an appearance loss function and the appearance similarity matrix by using an optimization algorithm;
if the appearance loss is lower than the appearance loss threshold value, determining the setting parameters of the appearance feature map neural network by using a second training set;
training an appearance feature map neural network with set parameters according to an appearance loss function and a training set by using an optimization algorithm, and if the obtained appearance retraining loss is lower than an appearance loss threshold value, obtaining the trained appearance feature map neural network;
training a motion characteristic diagram neural network according to a motion loss function and the motion similarity matrix by using an optimization algorithm;
if the motion loss is lower than the motion loss threshold value, determining the setting parameters of the neural network of the motion characteristic diagram by using a second training set;
training a motion characteristic diagram neural network with set parameters according to a motion loss function and a training set by using an optimization algorithm, and if the obtained motion retraining loss is lower than a motion loss threshold value, obtaining the trained motion characteristic diagram neural network;
and finishing training of the appearance characteristic diagram neural network and the motion characteristic diagram neural network to obtain the trained diagram neural network.
Preferably, the determining of the appearance loss function comprises:
determining a first loss function according to the target similarity and the appearance similarity matrix;
determining node difference according to a receiving node in the appearance nodes and a first updating node corresponding to the receiving node;
determining a second loss function according to the node difference and the target similarity;
and determining an appearance loss function according to the first loss function and the second loss function.
Preferably, the processing the data set by using the trained neural network to obtain the similarity matrix includes:
processing the data set by using the trained neural network to obtain an appearance similarity matrix and a motion similarity matrix of the data set;
and carrying out weighted sum on the appearance similarity matrix and the motion similarity matrix of the data set to obtain a similarity matrix.
Preferably, after the calculating the similarity matrix by using the matching algorithm to obtain the matching result of the target in the dataset, the method further includes:
a single object tracker and/or a linear motion model is used to recover missing objects in the data set.
In a second aspect, the present application provides a multi-target tracking system based on a graph neural network, including:
the preprocessing module is used for preprocessing the training set to obtain a node set, an edge set and a global variable;
the training module is used for inputting the node set, the edge set and the global variable to a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix; training the graph neural network according to a loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network;
the processing module is used for processing the data set by using the trained graph neural network to obtain a similarity matrix; and calculating the similarity matrix by using a matching algorithm to obtain a matching result of the targets in the data set.
The application has the advantages that: the method has the advantages that the node set, the edge set and the global variable can be updated by using the graph neural network through an optimization algorithm and a loss function, the trained graph neural network can be obtained by training after the setting parameters of the graph neural network are determined, the accuracy of the graph neural network can be improved, the characteristic types are increased by introducing and updating the global variable outside the nodes and the edges, the graph neural network can capture global information, the accuracy of multi-target tracking similarity can be improved, and the multi-target tracking performance is improved.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating preferred embodiments and are not to be construed as limiting the application. Also, like reference numerals are used to denote like parts throughout the drawings. In the drawings:
FIG. 1 is a schematic diagram illustrating steps of a multi-target tracking method based on a graph neural network according to the present application;
FIG. 2 is a schematic diagram of an appearance characteristic diagram neural network of a multi-target tracking method based on the diagram neural network provided by the application;
FIG. 3 is a schematic diagram of a neural network of a motion characteristic diagram of a multi-target tracking method based on the neural network of the present application;
FIG. 4 is a schematic tracking flow diagram of a multi-target tracking method based on a graph neural network provided by the present application;
fig. 5 is a schematic diagram of a multi-target tracking system based on a graph neural network provided in the present application.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
In a first aspect, according to an embodiment of the present application, a multi-target tracking method based on a graph neural network is provided, as shown in fig. 1, including:
s101, preprocessing a training set to obtain a node set, an edge set and a global variable;
s102, inputting a node set, an edge set and a global variable to a Graph Neural Network (GNN) to obtain an appearance similarity matrix and a motion similarity matrix;
s103, training a graph neural network according to the loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining setting parameters of the graph neural network, and then training to obtain the trained graph neural network;
s104, processing the data set by using the trained neural network to obtain a similarity matrix;
and S105, calculating the similarity matrix by using a matching algorithm to obtain a matching result of the target in the data set.
Preprocessing a training set to obtain a node set, an edge set and a global variable, wherein the method comprises the following steps:
dividing the training set into a first training set and a second training set;
extracting depth features of all the labeled boxes in the frame of the first training set by using a preprocessing neural network to obtain an appearance node set;
calculating the overlapping degree of the labeling frames of all the adjacent frames in the first training set to obtain an edge set between the two labeling frames;
extracting the position, size and movement data of the labeling frames in all the frames in the first training set to obtain the motion characteristics of each labeling frame;
normalizing all motion characteristics to obtain a motion node set;
one-dimensional vectors are initialized randomly as global variables.
Inputting the node set, the edge set and the global variable into a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix, wherein the method comprises the following steps:
inputting an appearance node set, an edge set and a global variable to an appearance feature map neural network to obtain an appearance similarity matrix;
and inputting the motion node set, the edge set and the global variable to a neural network of a set motion characteristic diagram to obtain a motion similarity matrix.
Inputting an appearance node set, an edge set and a global variable into an appearance feature map neural network to obtain an appearance similarity matrix, wherein the method comprises the following steps:
inputting an edge set, two appearance nodes corresponding to each edge in the input edge set and a global variable to a first edge neural network to obtain a first updating edge;
inputting a first updating edge, two appearance nodes connected through the first updating edge and a global variable to a node neural network to obtain a first updating node;
fusing all the obtained first updating edges to obtain a second updating edge;
fusing all sending nodes in the appearance nodes and the obtained first updating node to obtain a second updating node;
fusing all the first updating edges to obtain a second updating edge;
inputting the global variable, a second updating node and a second updating edge to a first global neural network to obtain a first updating global variable;
and inputting the updated global variable, all the sending nodes in the appearance nodes, the first updated node and the first updated edge to a second edge neural network to obtain an appearance similarity matrix.
Inputting a motion node set, an edge set and a global variable to a motion characteristic graph neural network to obtain a motion similarity matrix, wherein the motion similarity matrix comprises the following steps:
fusing all edges in the edge set to obtain a third updating edge;
fusing all motion nodes in the motion node set to obtain a third update node;
inputting the global variable, a third updating edge and a third updating node to a second global neural network to obtain a second updating global variable;
inputting an edge set, two motion nodes corresponding to each edge in the input edge set, and a second updated global variable to a third neural network to obtain a motion similarity matrix.
Training a graph neural network according to the loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network, wherein the training comprises the following steps:
training an appearance characteristic diagram neural network according to an appearance loss function and an appearance similarity matrix by using an optimization algorithm;
if the appearance loss is lower than the appearance loss threshold value, determining the setting parameters of the appearance feature map neural network by using a second training set;
training an appearance feature map neural network with set parameters according to an appearance loss function and a training set by using an optimization algorithm, and if the obtained appearance retraining loss is lower than an appearance loss threshold value, obtaining the trained appearance feature map neural network;
training a motion characteristic diagram neural network according to a motion loss function and a motion similarity matrix by using an optimization algorithm;
if the motion loss is lower than the motion loss threshold value, determining the setting parameters of the neural network of the motion characteristic diagram by using a second training set;
training a motion characteristic diagram neural network with set parameters according to a motion loss function and a training set by using an optimization algorithm, and if the obtained motion retraining loss is lower than a motion loss threshold value, obtaining the trained motion characteristic diagram neural network;
and finishing training of the appearance characteristic diagram neural network and the motion characteristic diagram neural network to obtain the trained diagram neural network.
Determining an appearance loss function, comprising:
determining a first loss function according to the target similarity and the appearance similarity matrix;
determining node difference according to a receiving node in the appearance nodes and a first updating node corresponding to the receiving node;
determining a second loss function according to the node difference and the target similarity;
an appearance loss function is determined according to the first loss function and the second loss function.
Determination of a motion loss function, comprising:
and determining a motion loss function according to the target similarity and the motion similarity matrix.
Processing the data set by using the trained neural network to obtain a similarity matrix, wherein the similarity matrix comprises the following steps:
processing the data set by using the trained neural network to obtain an appearance similarity matrix and a motion similarity matrix of the data set;
and carrying out weighted sum on the appearance similarity matrix and the motion similarity matrix of the data set to obtain a similarity matrix.
After the similarity matrix is calculated by using a matching algorithm to obtain a matching result of the target in the data set, the method further comprises the following steps:
a single object tracker and/or a linear motion model is used to recover missing objects in the data set.
The missing target for the current frame (the frame currently being processed) is recovered back with the single target tracker. And if a first recovery threshold value exists, tracking the lost target of the current frame by using the single-target tracker, and when the single-target tracker gives a tracking result with a confidence coefficient higher than the first recovery threshold value, recovering the lost target of the current frame by using the tracking result given by the single-target tracker.
The object lost before the current frame and retrieved in the current frame is recovered by a linear motion model. Assume that there is a second recovery threshold and that the tracked object has a constant moving speed during the period of loss. If the number of lost frames is less than the second recovery threshold, the target in the lost frame is recovered by using the two connected marking frames, and if the number of lost frames is greater than the second recovery threshold, the target is blocked and does not need to be recovered.
The global variable is a one-dimensional vector and is used for representing global information captured by the neural network of the graph in the updating process.
The extracted depth features are used to obtain appearance features (appearance nodes) that do not include: position information and motion information.
The appearance node comprises the following steps according to the time of the frame: the system comprises a sending node and a receiving node, wherein the node at the front of time is the sending node, and the node at the back of time is the receiving node. An edge connects a node in each of two adjacent frames. At the same time as an edge is input, two nodes (a transmitting node and a receiving node) connected through the edge are input.
The first update node is a receiving node updated using a first update edge, two appearance nodes connected by the first update edge, and a global variable.
The difference function for calculating the node (node feature) difference includes: Mean-Square Error (MSE), Cosine Distance (Cosine Distance), and the like.
The labeling boxes in the training set are artificially labeled.
Take for example two adjacent frames a and b. Assuming that the frame a is a previous frame, in which there are 3 labeled frames, and the frame b is a next frame, in which there are 2 labeled frames, after the frame a and the frame b are input to the graph neural network, there are 6 edges in total, that is, the number of edges between each labeled frame in the frame a and each labeled frame in the frame b. Each edge represents the similarity between the two objects (labeled boxes) connected to it.
According to the characteristics of the multi-target tracking task, the appearance feature map neural network and the motion feature map neural network have respective sub-modules and a mode of updating the map neural network edge.
Next, as shown in fig. 2 and 3, the present embodiment will be further explained.
Firstly, an appearance feature map neural network based on appearance features and a motion feature map neural network based on motion features are constructed.
And (3) extracting the appearance features of the labeling frame by using a pre-trained neural network (pre-processing neural network) based on the appearance feature neural network of the appearance feature graph, and expressing the outer nodes by using the extracted appearance features to obtain an appearance node set. The edges of two labeling frames in adjacent frames use their overlapping degrees (IoU) as initialized edges to obtain an edge set. The overlapping degree is the ratio of the area of the common area of the two labeling boxes to the area of the coverage area of the two labeling boxes.
The motion feature extraction method based on the motion feature neural network of the motion feature map comprises the following steps: position, size, movement data, etc. Expressing the motion nodes by the normalized motion characteristics to obtain a motion node set; the edges of the two label boxes are also used as the initialized edges by using the overlapping degree of the edges to obtain an edge set.
The sub-modules of the appearance feature map neural network comprise: updating two neural networks of edge e in the edge set, the first edge neural network
Figure BDA0002324803720000111
And a second edge neural network
Figure BDA0002324803720000112
First global neural network for updating global variable u
Figure BDA0002324803720000113
Updating a receiving node vrNode neural network of
Figure BDA0002324803720000114
The node in the node set is v, and the sending node is vsThe receiving node is vr. The appearance feature based graph neural network uses symbols with superscript a.
The sub-modules of the motor characteristic map neural network comprise: second global neural network for updating global variables
Figure BDA0002324803720000115
Third edge neural network for updating edges
Figure BDA0002324803720000116
The motion feature based graph neural network uses symbols with superscript m.
The neural network in the appearance characteristic diagram neural network and the motion characteristic diagram neural network comprises: multi-tier perceptrons and/or RNNs, etc.
The updating mode of the neural network based on the map of the appearance features is that, as shown in figure 2,
Figure BDA0002324803720000121
using global variables uaTwo nodes connected by an edge
Figure BDA0002324803720000122
(the sending node) of the first node,
Figure BDA0002324803720000123
(receiving node) and edge eaTo update this edge to get the first updated edge ea′
Figure BDA0002324803720000124
Using global variables uaTwo nodes connected by an edge
Figure BDA0002324803720000125
Figure BDA0002324803720000126
And a first updated edge e obtained after updatinga′Updating the receiving node to obtain a first updated node
Figure BDA0002324803720000127
Figure BDA0002324803720000128
Representing, fusing all the sending nodes and the updated receiving nodes to obtain a second updated node
Figure BDA0002324803720000129
Figure BDA00023248037200001210
Representing and fusing all the updated edges to obtain a second updated edge
Figure BDA00023248037200001211
Figure BDA00023248037200001212
Using global variables uaAnd a second updating edge obtained after the fusion updating
Figure BDA00023248037200001213
And a second update node obtained after the fusion update
Figure BDA00023248037200001214
Updating the global variable to obtain a first update ua′
Figure BDA00023248037200001215
Using the first updated global variable u obtained after updatinga′Node connected by edge
Figure BDA00023248037200001216
Figure BDA00023248037200001217
And a first update edge ea′The edge is updated again to obtain the appearance edge moment ea″
The updating mode of the neural network based on the motion characteristic of the motion characteristic diagram is, as shown in figure 3,
Figure BDA00023248037200001218
using global variables umAnd a third update node obtained after fusion
Figure BDA00023248037200001219
And a third updated edge obtained after the fusion
Figure BDA00023248037200001220
Updating the global variable to obtain a second updated global variable um′
Figure BDA00023248037200001221
Updating the global variable u by using the second updated global variable obtained after updatingm′Two nodes connected by edges
Figure BDA00023248037200001222
Figure BDA00023248037200001223
And edge emTo update the edge to obtain a motion similarity matrix em′
Figure BDA00023248037200001224
Representing and fusing all the edges to obtain a third updated edge
Figure BDA00023248037200001225
Figure BDA00023248037200001226
Representing, fusing all the sending nodes and the receiving nodes to obtain a third updating node
Figure BDA00023248037200001227
The fusion mode comprises the following steps: addition and averaging, RNN fusion, etc.
And training the graph neural network by using the multi-target tracking training data set.
The training data set is divided into a training set (first training set) and a validation set (second training set).
An appearance loss function and a motion loss function are determined.
For the graph neural network, the cross entropy L of the similarity (target similarity) p of the real similarity and the similarity (appearance similarity matrix or motion similarity matrix) p' of the output of the appearance feature or motion feature graph neural network is adoptedmatch-plogp '- (1-p) log (1-p') as a first loss function. For a graph neural network with updated nodes of the appearance feature graph neural network, since it makes sense to update the node features only when two nodes belong to the same target, the two nodes should not be updated when they are not the same target. Therefore, the nerves are mapped to the appearance featuresSecond loss function for node update in the network: l isNodeUpdate=(1-p)Difference(vr,v′r) Wherein v isrDenotes a receiving node, v'rIs represented by the formularDifference represents a function for calculating the Difference in node characteristics with respect to an updated receiving node belonging to an object. The function that calculates the difference in node characteristics may be mean square error, cosine distance, and the like. Therefore, the appearance loss function of the appearance feature map neural network is: l isa=Lmatch+LNodeUpdate(ii) a The motion loss function of the neural network of the motion profile is: l ism=Lmatch. When the first loss function is used for the neural network of the appearance characteristic diagram, the similarity p 'is an appearance similarity matrix, and when the first loss function is used for the neural network of the motion characteristic diagram, the similarity p' is a motion similarity matrix.
The similarity p is a direct similarity matrix of each labeled frame in adjacent frames, and the numerical value is 1 or 0, i.e., the same is 1, and the different is 0. The similarity p' of the neural network output of the graph is a similarity matrix with similarity values between 0 and 1.
And based on the loss function, training the appearance feature map neural network and the motion feature map neural network respectively by using a first training set and an optimization algorithm. The optimization algorithm may preferably be a gradient-based optimization algorithm, such as a Stochastic Gradient Descent (SGD) method, Adam and adagard, and the like.
And when the appearance loss obtained according to the appearance loss function is smaller than an appearance loss threshold value, obtaining a trained appearance feature map neural network. Using the verification set, setting parameters of the neural network of the appearance feature map are determined.
And when the motion loss obtained according to the motion loss function is smaller than a motion loss threshold value, obtaining a trained motion characteristic diagram neural network. Using the verification set, setting parameters of the neural network of the appearance feature map are determined.
The setting parameters mainly comprise some hyper-parameters, such as \ alpha and some threshold settings. And selecting the setting of the appearance characteristic diagram neural network and the movement characteristic diagram neural network with the best tracking performance by running out the tracking performance (the tracking performance has special indexes: MOTA, IDF1, FP, FN, IDsw and the like) on the verification set.
Inputting training data comprising a training set and a verification set to the appearance feature map neural network with set parameters, retraining the appearance feature map neural network with the set parameters according to an appearance loss function by using an optimization algorithm, and obtaining the trained appearance feature map neural network when the obtained appearance retraining loss is lower than an appearance loss threshold.
Inputting training data comprising a training set and a verification set to a motion characteristic diagram neural network with set parameters, retraining the motion characteristic diagram neural network with the set parameters according to a motion loss function by using an optimization algorithm, and obtaining the trained motion characteristic diagram neural network when the obtained motion retraining loss is lower than an appearance loss threshold value.
And obtaining the trained neural network of the image after the neural network of the appearance characteristic diagram and the neural network of the motion characteristic diagram with the set parameters are trained.
During network training, global variables are initially obtained by random initialization and then updated according to training. And after the trained network is obtained, using the updated global variable as the global variable of the trained network to perform multi-target tracking.
And performing multi-target tracking on the test data set by using the trained graph neural network, as shown in FIG. 4.
And constructing nodes and edge stages of the neural network of the graph, and taking each labeling box in each frame as a node respectively and expressing the node by using an external feature or a motion feature. And extracting depth features by using a pre-trained network, marking the position, the size and the speed of a frame by using the motion features, and normalizing the motion features by using the size of the picture.
The connection from the previous frame of the label box to the current label box is represented by a directed edge, and IoU of the two connected label boxes is used for initializing the corresponding edge.
And respectively updating global variables of the appearance feature map neural network and the motion feature map neural network.
An updating edge stage of the graph neural network, wherein the appearance characteristic graph neural network and the motion characteristic graph neural network respectively update edges, and the appearance similarity matrix M of the similarity matrix is obtained according to the last updated edgeaAnd motion similarity matrix Mm
In the tracking stage, the two obtained similarity matrixes are weighted and combined to obtain a final similarity matrix M which is α Ma+(1-α)Mmα is a hyper-parameter (weight) set of optimal parameters chosen on the verification set.
And obtaining an optimal matching result (result with the maximum probability) by using a matching algorithm based on the obtained final similarity matrix, wherein the obtained matching result is the tracking result of the graph neural network. The matching algorithm comprises the following steps: hungarian algorithm, greedy algorithm, etc.
For the target which is not tracked in the current frame, firstly, a single target tracker is used for recovery. Will exceed tauoutThe target node is removed from the neural network of the graph; and adding the frame which is not connected with the target in the frame before the current frame as a new target into the graph neural network. Wherein, tauoutAnd the parameters are hyper-parameters, and the optimal parameters are selected from the verification set.
For some previous frames, the single-target tracker does not track the recovered targets, but the neural network tracks the recovered targets, and the linear motion model is used to recover the lost targets in the frames. In the tracking result, the frame before the current frame without matching the labeled frame in the frame before the current frame is recovered by using a single target tracker, such as DaSiamRPN. In particular, when the confidence given by the single-target tracker is higher than the first recovery threshold τconfAnd in time, the target is recovered by using a tracking result given by the single-target tracker. When the number of the frame of the lost label frame in the middle of the connection is less than the second recovery threshold taugapThen, the linear motion model is used for recovery; otherwise, these missing parts are considered occluded and not restored. Tau isconfAnd τgapHyper-parametric, selected optima on verification setAnd setting parameters.
In a second aspect, according to an embodiment of the present application, there is further provided a multi-target tracking system based on a graph neural network, as shown in fig. 5, including:
the preprocessing module 101 is configured to preprocess the training set to obtain a node set, an edge set, and a global variable;
the training module 102 is used for inputting the node set, the edge set and the global variable into a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix; training a graph neural network according to the loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network;
the processing module 103 is configured to process the data set by using the trained neural network to obtain a similarity matrix; and calculating the similarity matrix by using a matching algorithm to obtain a matching result of the target in the data set.
According to the method, the node set, the edge set and the global variable can be updated through an optimization algorithm and a loss function by using the graph neural network, the graph neural network is trained after the set parameters of the graph neural network are determined, the trained graph neural network is obtained, the accuracy of the graph neural network can be improved, the characteristic types are increased by introducing and updating the global variable outside the nodes and the edges, the global information can be captured by the graph neural network, the accuracy of the multi-target tracking similarity can be improved, and therefore the multi-target tracking performance is improved. And the similarity among the nodes in the neural network of the graph is updated through the neural network of the graph, so that more accurate similarity is obtained, and better tracking performance is obtained. Because nodes and edges can be updated in the graph neural network, in addition, the graph model does not model global information, and the introduction of global variables in the graph neural network enables the graph neural network to capture the global information. The method and the device are suitable for online and near-online multi-target tracking. The existing multi-target tracking method has no multi-target tracking method based on a graph neural network.
The above description is only for the preferred embodiment of the present application, but the scope of the present application is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present application should be covered within the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (10)

1. A multi-target tracking method based on a graph neural network is characterized by comprising the following steps:
preprocessing a training set to obtain a node set, an edge set and a global variable;
inputting the node set, the edge set and the global variable to a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix;
training the graph neural network according to a loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network;
processing the data set by using the trained neural network to obtain a similarity matrix;
and calculating the similarity matrix by using a matching algorithm to obtain a matching result of the targets in the data set.
2. The method of claim 1, wherein preprocessing the training set to obtain a set of nodes, a set of edges, and global variables comprises:
dividing the training set into a first training set and a second training set;
extracting depth features of all the labeled boxes in the frame of the first training set by using a preprocessing neural network to obtain an appearance node set;
calculating the overlapping degree of the labeling frames of all the adjacent frames in the first training set to obtain an edge set between the two labeling frames;
extracting the position, size and movement data of the labeling frames in all frames in the first training set to obtain the motion characteristics of each labeling frame;
normalizing all motion characteristics to obtain a motion node set;
one-dimensional vectors are initialized randomly as global variables.
3. The method of claim 1, wherein said inputting the set of nodes, the set of edges, and the global variables into a graph neural network, resulting in an appearance similarity matrix and a motion similarity matrix, comprises:
inputting an appearance node set, an edge set and a global variable to an appearance feature map neural network to obtain an appearance similarity matrix;
and inputting the motion node set, the edge set and the global variable to a motion characteristic graph neural network to obtain a motion similarity matrix.
4. The method of claim 3, wherein inputting the set of appearance nodes, the set of edges, and the global variable into an appearance feature map neural network, resulting in an appearance similarity matrix, comprises:
inputting an edge set, two appearance nodes corresponding to each edge in the input edge set and a global variable to a first edge neural network to obtain a first updating edge;
inputting the first updating edge, two appearance nodes connected through the first updating edge and a global variable to a node neural network to obtain a first updating node;
fusing all the obtained first updating edges to obtain a second updating edge;
fusing all sending nodes in the appearance nodes and the obtained first updating node to obtain a second updating node;
fusing all the first updating edges to obtain a second updating edge;
inputting the global variable, a second updating node and a second updating edge to a first global neural network to obtain a first updating global variable;
and inputting the updated global variable, all the sending nodes in the appearance nodes, the first updated node and the first updated edge to a second edge neural network to obtain an appearance similarity matrix.
5. The method of claim 3, wherein inputting the set of motion nodes, the set of edges, and the global variable to motion feature map neural network results in a motion similarity matrix comprising:
fusing all edges in the edge set to obtain a third updating edge;
fusing all motion nodes in the motion node set to obtain a third update node;
inputting the global variable, a third updating edge and a third updating node to a second global neural network to obtain a second updating global variable;
inputting an edge set, two motion nodes corresponding to each edge in the input edge set, and the second updated global variable to a third neural network to obtain a motion similarity matrix.
6. The method of claim 1, wherein the using an optimization algorithm to train the graph neural network according to a loss function, the appearance similarity matrix and the motion similarity matrix, determine setting parameters of the graph neural network, and train again to obtain the trained graph neural network comprises:
training an appearance feature map neural network according to an appearance loss function and the appearance similarity matrix by using an optimization algorithm;
if the appearance loss is lower than the appearance loss threshold value, determining the setting parameters of the appearance feature map neural network by using a second training set;
training an appearance feature map neural network with set parameters according to an appearance loss function and a training set by using an optimization algorithm, and if the obtained appearance retraining loss is lower than an appearance loss threshold value, obtaining the trained appearance feature map neural network;
training a motion characteristic diagram neural network according to a motion loss function and the motion similarity matrix by using an optimization algorithm;
if the motion loss is lower than the motion loss threshold value, determining the setting parameters of the neural network of the motion characteristic diagram by using a second training set;
training a motion characteristic diagram neural network with set parameters according to a motion loss function and a training set by using an optimization algorithm, and if the obtained motion retraining loss is lower than a motion loss threshold value, obtaining the trained motion characteristic diagram neural network;
and finishing training of the appearance characteristic diagram neural network and the motion characteristic diagram neural network to obtain the trained diagram neural network.
7. The method of claim 6, wherein the determining of the appearance loss function comprises:
determining a first loss function according to the target similarity and the appearance similarity matrix;
determining node difference according to a receiving node in the appearance nodes and a first updating node corresponding to the receiving node;
determining a second loss function according to the node difference and the target similarity;
and determining an appearance loss function according to the first loss function and the second loss function.
8. The method of claim 1, wherein processing the data set using the trained neural network to obtain a similarity matrix comprises:
processing the data set by using the trained neural network to obtain an appearance similarity matrix and a motion similarity matrix of the data set;
and carrying out weighted sum on the appearance similarity matrix and the motion similarity matrix of the data set to obtain a similarity matrix.
9. The method of claim 1, wherein after said calculating the similarity matrix using a matching algorithm to obtain a matching result for the objects in the dataset, further comprising:
a single object tracker and/or a linear motion model is used to recover missing objects in the data set.
10. A multi-target tracking system based on a graph neural network is characterized by comprising:
the preprocessing module is used for preprocessing the training set to obtain a node set, an edge set and a global variable;
the training module is used for inputting the node set, the edge set and the global variable to a graph neural network to obtain an appearance similarity matrix and a motion similarity matrix; training the graph neural network according to a loss function, the appearance similarity matrix and the motion similarity matrix by using an optimization algorithm, determining the setting parameters of the graph neural network, and then training to obtain the trained graph neural network;
the processing module is used for processing the data set by using the trained graph neural network to obtain a similarity matrix; and calculating the similarity matrix by using a matching algorithm to obtain a matching result of the targets in the data set.
CN201911312114.9A 2019-12-18 2019-12-18 Multi-target tracking method and system based on graph neural network Active CN111161315B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911312114.9A CN111161315B (en) 2019-12-18 2019-12-18 Multi-target tracking method and system based on graph neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911312114.9A CN111161315B (en) 2019-12-18 2019-12-18 Multi-target tracking method and system based on graph neural network

Publications (2)

Publication Number Publication Date
CN111161315A true CN111161315A (en) 2020-05-15
CN111161315B CN111161315B (en) 2023-01-03

Family

ID=70557249

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911312114.9A Active CN111161315B (en) 2019-12-18 2019-12-18 Multi-target tracking method and system based on graph neural network

Country Status (1)

Country Link
CN (1) CN111161315B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111626194A (en) * 2020-05-26 2020-09-04 佛山市南海区广工大数控装备协同创新研究院 Pedestrian multi-target tracking method using depth correlation measurement
CN111665861A (en) * 2020-05-19 2020-09-15 中国农业大学 Trajectory tracking control method, apparatus, device and storage medium
CN111862156A (en) * 2020-07-17 2020-10-30 中南民族大学 Multi-target tracking method and system based on graph matching
CN112215869A (en) * 2020-10-12 2021-01-12 华中科技大学 Group target tracking method and system based on graph similarity constraint
CN112414401A (en) * 2020-11-06 2021-02-26 北京理工大学 Unmanned aerial vehicle cooperative positioning system and method based on graph neural network
CN112465006A (en) * 2020-11-24 2021-03-09 中国人民解放军海军航空大学 Graph neural network target tracking method and device
CN112561963A (en) * 2020-12-18 2021-03-26 北京百度网讯科技有限公司 Target tracking method and device, road side equipment and storage medium
CN113362368A (en) * 2021-07-26 2021-09-07 北京邮电大学 Crowd trajectory prediction method based on multi-level space-time diagram neural network
CN115668877A (en) * 2020-06-03 2023-01-31 华为技术有限公司 Method and system for generating a network configuration using a graph neural network

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105976400A (en) * 2016-05-10 2016-09-28 北京旷视科技有限公司 Object tracking method and device based on neural network model
CN107861123A (en) * 2017-10-24 2018-03-30 武汉大学 A kind of through-wall radar is under complex environment to the method for multiple mobile object real-time tracking
CN108682022A (en) * 2018-04-25 2018-10-19 清华大学 Based on the visual tracking method and system to anti-migration network
CN109508000A (en) * 2018-12-16 2019-03-22 西南电子技术研究所(中国电子科技集团公司第十研究所) Isomery multi-sensor multi-target tracking method
US20190217476A1 (en) * 2018-01-12 2019-07-18 Futurewei Technologies, Inc. Robot navigation and object tracking
CN110276783A (en) * 2019-04-23 2019-09-24 上海高重信息科技有限公司 A kind of multi-object tracking method, device and computer system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105976400A (en) * 2016-05-10 2016-09-28 北京旷视科技有限公司 Object tracking method and device based on neural network model
CN107861123A (en) * 2017-10-24 2018-03-30 武汉大学 A kind of through-wall radar is under complex environment to the method for multiple mobile object real-time tracking
US20190217476A1 (en) * 2018-01-12 2019-07-18 Futurewei Technologies, Inc. Robot navigation and object tracking
CN108682022A (en) * 2018-04-25 2018-10-19 清华大学 Based on the visual tracking method and system to anti-migration network
CN109508000A (en) * 2018-12-16 2019-03-22 西南电子技术研究所(中国电子科技集团公司第十研究所) Isomery multi-sensor multi-target tracking method
CN110276783A (en) * 2019-04-23 2019-09-24 上海高重信息科技有限公司 A kind of multi-object tracking method, device and computer system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CONG MA等: "Deep Association: End-to-end Graph-Based Learning for Multiple Object Tracking with Conv-Graph Neural Network", 《ICMR ’19》 *
XIAOLONG JIANG等: "Graph Neural Based End-to-end Data Association Framework for Online Multiple-Object Tracking", 《ARXIV:1907.05315V1》 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111665861A (en) * 2020-05-19 2020-09-15 中国农业大学 Trajectory tracking control method, apparatus, device and storage medium
CN111626194A (en) * 2020-05-26 2020-09-04 佛山市南海区广工大数控装备协同创新研究院 Pedestrian multi-target tracking method using depth correlation measurement
CN111626194B (en) * 2020-05-26 2024-02-02 佛山市南海区广工大数控装备协同创新研究院 Pedestrian multi-target tracking method using depth correlation measurement
CN115668877A (en) * 2020-06-03 2023-01-31 华为技术有限公司 Method and system for generating a network configuration using a graph neural network
CN111862156A (en) * 2020-07-17 2020-10-30 中南民族大学 Multi-target tracking method and system based on graph matching
CN111862156B (en) * 2020-07-17 2021-02-26 中南民族大学 Multi-target tracking method and system based on graph matching
CN112215869A (en) * 2020-10-12 2021-01-12 华中科技大学 Group target tracking method and system based on graph similarity constraint
CN112414401A (en) * 2020-11-06 2021-02-26 北京理工大学 Unmanned aerial vehicle cooperative positioning system and method based on graph neural network
CN112414401B (en) * 2020-11-06 2023-02-28 北京理工大学 Unmanned aerial vehicle cooperative positioning system and method based on graph neural network
CN112465006A (en) * 2020-11-24 2021-03-09 中国人民解放军海军航空大学 Graph neural network target tracking method and device
CN112465006B (en) * 2020-11-24 2022-08-05 中国人民解放军海军航空大学 Target tracking method and device for graph neural network
CN112561963A (en) * 2020-12-18 2021-03-26 北京百度网讯科技有限公司 Target tracking method and device, road side equipment and storage medium
CN113362368A (en) * 2021-07-26 2021-09-07 北京邮电大学 Crowd trajectory prediction method based on multi-level space-time diagram neural network

Also Published As

Publication number Publication date
CN111161315B (en) 2023-01-03

Similar Documents

Publication Publication Date Title
CN111161315B (en) Multi-target tracking method and system based on graph neural network
CN110414432B (en) Training method of object recognition model, object recognition method and corresponding device
US8345984B2 (en) 3D convolutional neural networks for automatic human action recognition
CN107145862B (en) Multi-feature matching multi-target tracking method based on Hough forest
CN110728698B (en) Multi-target tracking system based on composite cyclic neural network system
CN111583263A (en) Point cloud segmentation method based on joint dynamic graph convolution
CN112801019B (en) Method and system for eliminating re-identification deviation of unsupervised vehicle based on synthetic data
CN112507859B (en) Visual tracking method for mobile robot
CN112507778B (en) Loop detection method of improved bag-of-words model based on line characteristics
CN114092517A (en) Multi-target tracking method based on traditional and deep learning algorithm
KR20230171966A (en) Image processing method and device and computer-readable storage medium
Yang et al. Learning relation by graph neural network for SAR image few-shot learning
CN104778699A (en) Adaptive object feature tracking method
Hammam et al. DeepPet: A pet animal tracking system in internet of things using deep neural networks
Getahun et al. A deep learning approach for lane detection
CN112507893A (en) Distributed unsupervised pedestrian re-identification method based on edge calculation
Yin Object Detection Based on Deep Learning: A Brief Review
CN103500456A (en) Object tracking method and equipment based on dynamic Bayes model network
Sriram et al. Analytical review and study on object detection techniques in the image
Chu et al. Illumination-guided transformer-based network for multispectral pedestrian detection
CN113129336A (en) End-to-end multi-vehicle tracking method, system and computer readable medium
Feng et al. DAMUN: A domain adaptive human activity recognition network based on multimodal feature fusion
Lu et al. Hybrid deep learning based moving object detection via motion prediction
Pang et al. Target tracking based on siamese convolution neural networks
Cheng et al. Class attendance checking system based on deep learning and global optimization

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant