CN106971401B - Multi-target tracking device and method - Google Patents

Multi-target tracking device and method Download PDF

Info

Publication number
CN106971401B
CN106971401B CN201710203912.2A CN201710203912A CN106971401B CN 106971401 B CN106971401 B CN 106971401B CN 201710203912 A CN201710203912 A CN 201710203912A CN 106971401 B CN106971401 B CN 106971401B
Authority
CN
China
Prior art keywords
tracking
target
module
detection
queue
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710203912.2A
Other languages
Chinese (zh)
Other versions
CN106971401A (en
Inventor
邹李兵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Original Assignee
Lenovo Beijing Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd filed Critical Lenovo Beijing Ltd
Priority to CN201710203912.2A priority Critical patent/CN106971401B/en
Publication of CN106971401A publication Critical patent/CN106971401A/en
Application granted granted Critical
Publication of CN106971401B publication Critical patent/CN106971401B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Abstract

The application provides a multi-target tracking device and a multi-target tracking method. The multi-target tracking apparatus includes: a tracking module configured to receive video frame data and acquire indication data indicating a plurality of objects to be tracked, and track the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects so as to generate a tracking result; a detection module configured to receive the video frame data and detect the plurality of objects in each frame of the video frame data to generate a detection result; the learning module is configured to receive the video frame data, update the detection module according to the tracking result of the tracking module, and update the tracking module according to the detection result of the detection module; and an integration module configured to receive a tracking result of the tracking module and a detection result of the detection module to generate indication data indicating a plurality of targets being tracked.

Description

Multi-target tracking device and method
Technical Field
The application relates to a multi-target tracking device and method.
Background
TLD (Tracking-Learning-Detection) is a new single target long time (long Tracking) Tracking algorithm. The TLD is remarkably different from the traditional tracking method in that the traditional tracking algorithm and the traditional detection algorithm are combined to solve the problems of deformation, partial shielding and the like of the tracked target in the tracking process, and the tracking effect is more stable, robust and reliable. However, the original TLD is a single target tracking method, and is not applicable in a multi-target tracking application scenario.
For this reason, it is desirable to provide a multi-target tracking apparatus and method capable of tracking a plurality of targets.
Disclosure of Invention
According to an embodiment of the present application, there is provided a multi-target tracking apparatus including:
a tracking module configured to receive video frame data and acquire indication data indicating a plurality of objects to be tracked, and track the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects so as to generate a tracking result;
a detection module configured to receive the video frame data and detect the plurality of objects in each frame of the video frame data to generate a detection result;
the learning module is configured to receive the video frame data, update the detection module according to the tracking result of the tracking module, and update the tracking module according to the detection result of the detection module; and
an integration module configured to receive a tracking result of the tracking module and a detection result of the detection module to generate indication data indicating a plurality of targets being tracked.
Optionally, the tracking module comprises a multi-target tracking manager configured to:
performing addition of a tracking target according to a predetermined operation;
performing an update of a plurality of tracking targets; and
and managing the bidirectional mapping relation between the first queues of the tracking target frames of the plurality of tracking targets and the second queues of the target area characteristic points of the plurality of tracking targets.
Optionally, the multi-target tracking manager is further configured to:
tracking all the feature points in the second queue of the feature points of the target area by using an optical flow method so as to generate a tracking result;
determining whether the tracking target frame of each of the plurality of tracking targets succeeds in tracking according to the tracking result and the reverse mapping relation between the characteristic points and the tracking target frame;
if the result is successful, recalculating the tracking target frame and updating the feature points, otherwise deleting the tracking target frame of the tracking target and the corresponding feature points from the first queue and the second queue respectively.
Optionally, the learning module comprises:
a sample queue generator configured to generate a positive sample queue and a negative sample queue, wherein the positive sample queue includes a tracking target frame of each tracking target tracked by the tracking module or a detection target frame of each tracking target detected by the detection module in each frame of the video frame data, the negative sample queue is a common negative sample stored, and the negative sample is an area not intersecting the positive sample within a predetermined range near the positive sample, wherein, for each of the plurality of tracking targets, the positive sample queue includes the tracking target frame and the detection target frame of the tracking target, and the negative sample queue includes the common negative sample and positive samples belonging to the positive sample queues of the other tracking targets; and
the classifier is configured to perform similarity comparison on the positive sample queue and the negative sample queue of each tracking target respectively; acquiring the number and the attribution of the positive sample queue and the negative sample queue according to a set threshold value, and executing normalization processing; and calculating probabilities of positive and negative samples belonging to each tracked object.
Optionally, the detection module is further configured to:
for each frame of the video frame data, sliding a detection box to detect a negative sample;
determining the area in which the negative sample is detected as a subsequent non-detection area; and
for the remaining regions, positive sample detection is performed to determine a tracking target.
According to another embodiment of the present application, there is provided a multi-target tracking method including:
receiving, by a tracking module, video frame data and acquiring indication data indicating a plurality of objects to be tracked, and tracking the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects to generate a tracking result;
receiving, by a detection module, the video frame data and detecting the plurality of objects in each frame of the video frame data to generate a detection result;
receiving the video frame data through a learning module, updating the detection module according to the tracking result of the tracking module, and updating the tracking module according to the detection result of the detection module; and
receiving, by an integration module, a tracking result of the tracking module and a detection result of the detection module to generate indication data indicating a plurality of targets being tracked.
Optionally, the tracking module comprises a multi-target tracking manager configured to:
performing addition of a tracking target according to a predetermined operation;
updating a plurality of tracking targets according to a preset operation; and
and managing the bidirectional mapping relation between the first queues of the tracking target frames of the plurality of tracking targets and the second queues of the target area characteristic points of the plurality of tracking targets.
Optionally, the multi-target tracking manager is further configured to:
tracking all the feature points in the second queue of the feature points of the target area by using an optical flow method so as to generate a tracking result;
determining whether the tracking target frame of each of the plurality of tracking targets succeeds in tracking according to the tracking result and the reverse mapping relation between the characteristic points and the tracking target frame;
if the result is successful, recalculating the tracking target frame and updating the feature points, otherwise deleting the tracking target frame of the tracking target and the corresponding feature points from the first queue and the second queue respectively.
Optionally, the learning module comprises:
a sample queue generator configured to generate a positive sample queue and a negative sample queue, wherein the positive sample queue includes a tracking target frame of each tracking target tracked by the tracking module or a detection target frame of each tracking target detected by the detection module in each frame of the video frame data, the negative sample queue is a common negative sample stored, and the negative sample is an area not intersecting the positive sample within a predetermined range near the positive sample, wherein, for each of the plurality of tracking targets, the positive sample queue includes the tracking target frame and the detection target frame of the tracking target, and the negative sample queue includes the common negative sample and positive samples belonging to the positive sample queues of the other tracking targets; and
the classifier is configured to perform similarity comparison on the positive sample queue and the negative sample queue of each tracking target respectively; acquiring the number and the attribution of the positive sample queue and the negative sample queue according to a set threshold value, and executing normalization processing; and calculating probabilities of positive and negative samples belonging to each tracked object.
Optionally, the detection module is further configured to:
for each frame of the video frame data, sliding a detection box to detect a negative sample;
determining the area in which the negative sample is detected as a subsequent non-detection area; and
for the remaining regions, positive sample detection is performed to determine a tracking target.
Therefore, by using the multi-target tracking device and method according to the embodiment of the application, a plurality of targets can be tracked.
Drawings
FIG. 1 is a block diagram illustrating a prior art TLD algorithm;
FIG. 2 is a block diagram illustrating a functional configuration of a multi-target tracking device according to an embodiment of the present application;
FIG. 3 is a diagram illustrating a mapping relationship of a trace target box queue and a feature point queue;
FIG. 4 is a schematic diagram illustrating a positive sample queue and a negative sample queue; and
FIG. 5 is a flow diagram illustrating a multi-target tracking method according to an embodiment of the application.
Detailed Description
Before describing the multi-target tracking apparatus and the multi-target tracking method according to the embodiment of the present application, the TLD algorithm is first briefly introduced.
As is well known to those skilled in the art, the TLD algorithm is a long-term, online learning, single-target tracking method with minimal a priori information. The TLD algorithm consists essentially of three parts: tracker, detector and learner. The tracker part is composed of a short-period self-adaptive tracker and is used for predicting the movement of the target between continuous frames under the condition that the movement between the frames is limited and the target is visible. The detector part is an efficient cascade classifier, simple and effective image features are created and used, and the target can be detected in real time, and meanwhile, the tracker is corrected under the necessary condition. The learner evaluates the performance of the tracker and detector and completes the updating of the detector by generating valid training samples, eliminating detector errors.
The framework of the TLD algorithm is shown in fig. 1. In the initial frame of tracking the target, the initialization of the TLD algorithm is completed by giving the position and the size of the target. In the subsequent tracking process, each frame image is processed in parallel by the tracker and the detector together.
Specifically, the tracker 101 estimates the position of the target in the current frame according to the position information of the target in the previous frame, the detector 103 globally scans the window of the current frame to detect one or more possible target positions, and the detection result and the tracking result are input to the synthesizer 104. The integrator 104 gives information on whether the current frame has a target, a target position, and whether a tracking track to the current frame is valid. The results of these integration processes, the joint detection results, and the tracking results are input to the learner 102, which completes the updating of the tracker and detector.
However, one drawback of existing TLD algorithms is that only a single target can be tracked. Therefore, the method and the device improve the existing TLD algorithm, and realize multi-target tracking by respectively improving the tracker, the detector and the learner of the existing TLD algorithm.
Hereinafter, a multi-target tracking apparatus according to a first embodiment of the present application will be described in detail with reference to the accompanying drawings. As shown in fig. 2, the multi-target tracking apparatus 200 according to the first embodiment of the present application includes:
a tracking module 201 configured to receive video frame data and acquire indication data indicating a plurality of objects to be tracked, and track the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects so as to generate a tracking result;
a detection module 202 configured to receive the video frame data and detect the plurality of objects in each frame of the video frame data to generate a detection result;
a learning module 203 configured to receive the video frame data, update the detection module according to the tracking result of the tracking module, and update the tracking module according to the detection result of the detection module; and
an integration module 204 configured to receive the tracking result of the tracking module and the detection result of the detection module to generate indication data indicating the tracked plurality of targets.
Unlike the existing TLD, the tracking module 201 according to the embodiment of the present application acquires indication data indicating a plurality of objects to be tracked in addition to receiving video frame data.
The indication data may be a tracking box indicating the position and the size of the target, or may be a mark for identifying whether the tracked target is visible.
In one embodiment, for each of the plurality of tracking targets, a rectangle containing the tracking target may be manually determined in the first frame image, and initial coordinates and width and height information of the rectangle may be obtained as the indication data.
In another embodiment, the information of the object to be tracked may be stored in advance, the object to be tracked may be identified in the first frame image by means of image identification, a rectangle containing the tracking target may be generated, and the initial coordinates and the width and height information of the rectangle may be obtained as the indication data.
The tracking module 201 may also include a multi-target tracking manager 2011.
The multi-target tracking manager 2011 is configured to: performing addition of a tracking target according to a predetermined operation; performing an update of a plurality of tracking targets; and managing a bidirectional mapping relation between a first queue of the tracking target frames of the plurality of tracking targets and a second queue of the target area feature points of the plurality of tracking targets.
Specifically, for example, during tracking, by manually determining a rectangle containing a tracking target in the current frame image and obtaining initial coordinates and width and height information of the rectangle as indicating data, the multi-target tracking manager 2011 may add the manually determined tracking target as a tracking target to be tracked next.
Alternatively, the multi-target tracking manager 2011 may add a tracking target to be tracked next by transmitting information of the tracking target to be tracked to the tracker so that a rectangle containing the tracking target is generated in the current frame image during tracking and initial coordinates and width and height information of the rectangle are obtained as indicating data.
In addition, the multi-target tracking manager 2011 may perform updates of multiple tracking targets. The updating includes recalculating the tracking frames of the tracking targets and the corresponding feature points.
The multi-target tracking manager 2011 may further manage a bidirectional mapping relationship between a first queue of the tracking target frames of the plurality of tracking targets and a second queue of the target area feature points of the plurality of tracking targets.
The bidirectional mapping relationship will be described in detail below with reference to fig. 3.
As shown in fig. 3, for a plurality of tracked targets to be tracked (N is assumed, and N is an integer greater than 1), a tracking frame is generated for each tracked target, respectively. The trace frames of the plurality of trace targets form a first queue, i.e., obj1, obj2, … …, obj n in fig. 3.
The feature points are a number of pixel points selected in a tracking frame of the video data frame. As shown in fig. 3, each of the tracking frames obj1, obj2, … …, objN includes a plurality of pixel points as feature points. And each tracking frame and the corresponding characteristic point have a bidirectional mapping relation. By such a bidirectional mapping relationship, the corresponding tracking frame and feature point can be determined.
In addition, the multi-target tracking manager 2011 is further configured to:
tracking all the feature points in the second queue of the feature points of the target area by using an optical flow method so as to generate a tracking result;
determining whether the tracking target frame of each of the plurality of tracking targets succeeds in tracking according to the tracking result and the reverse mapping relation between the characteristic points and the tracking target frame;
if the result is successful, recalculating the tracking target frame and updating the feature points, otherwise deleting the tracking target frame of the tracking target and the corresponding feature points from the first queue and the second queue respectively.
In addition, the learning module 203 includes a sample queue generator 2031 and a classifier 2032.
The sample queue generator 2031 is configured to generate a positive sample queue and a negative sample queue. The positive sample queue includes a tracking target frame of each tracking target tracked by the tracking module 201 or a detection target frame of each tracking target detected by the detection module 202 in each frame of the video frame data.
Specifically, in each frame, when the confidence of the tracking target frame of the tracking target tracked by the tracking module 201 is high, the tracking frame tracked by the tracking module 201 may be selected as a positive sample. On the other hand, when the confidence of the tracking target frame of the tracking target tracked by the tracking module 201 is low and the confidence of the object frame detected by the detection module 202 is high, the detection frame detected by the detection module 202 may be selected as the positive sample. On the other hand, when the confidence of the tracking target frame of the tracking target tracked by the tracking module 201 is low and the confidence of the object frame detected by the detection module 202 is low, no positive sample is selected in the frame.
Thus, as shown in fig. 4, for obj1, obj2, … …, obj n, by the current frame, the positive samples of tracked object 1 include P11, P12, P13, P14, P15. The positive samples of the tracked object 2 include P21, P22, P23. … … are provided. Positive samples for trace object N include Pn1, Pn 2.
On the other hand, as shown in FIG. 4, the negative examples queue is a common negative example of storage. The negative sample is a region not intersecting the positive sample within a predetermined range around the positive sample. For example, for the positive sample P11 of the tracked object 1, a region of, for example, 10 pixels in size around the positive sample P11 may be taken as a negative sample. Note that the predetermined range may be freely set as needed.
Further, it is to be noted that, in the present embodiment, for each of a plurality of trace targets, the positive sample queue includes the trace target frame or the detection target frame of the trace target obtained in the above-described manner. On the other hand, for each of the plurality of trace targets, the negative sample queue includes, in addition to the common negative sample, positive samples belonging to positive sample queues of other trace targets. That is, for a particular tracked object, the positive samples of the other tracked objects are also negative samples for it.
In addition, the classifier 2032 may employ a multi-class K-nearest neighbor (KNN, K-nearest neighbor) classifier, for example. The KNN classifier is configured to perform similarity comparison on the positive sample queue and the negative sample queue of each tracking target respectively. Then, according to a set threshold, the number and the attribution of the positive sample queue and the negative sample queue are obtained, and normalization processing is performed to calculate the probability of the positive samples and the negative samples belonging to the tracking object objN, wherein the negative samples comprise the negative samples belonging to the negative sample queue and all the positive samples not belonging to the objN in the positive sample queue.
The detection module 203 is further configured to, for each frame of the video frame data, slide a detection box to detect a negative sample. Unlike the TLD algorithm of the prior art, because the sample queue generator 2031 has already generated the negative sample queue, the detection module 203 can reduce the area to be detected by filtering the area where the negative sample exists by detecting the negative sample first.
For example, if the detection box determines that the similarity of the negative examples is > 70%, the region may be determined to be free from the tracking object, i.e., the region in which the negative examples are detected is determined to be a subsequent non-detection region.
Finally, for the remaining area of the video frame, positive sample detection is performed to determine a tracking target. Thus, compared to a manner in which a plurality of detectors are simply provided for each object, the number of detectors in the embodiment of the present application is not increased, that is, there is only one detector. In this way, the computing and storage resources of the tracking system will be greatly saved.
On the other hand, by filtering the negative sample regions first, the detection region of the image frame can be greatly reduced, and thus the performance of the detector is improved.
As with the prior art TLD algorithm, the integration module 204 receives the tracking results of the tracking module and the detection results of the detection module to generate indication data indicative of the plurality of targets being tracked.
Specifically, the integration module 204 integrates the target frames obtained by the detector and the tracker and provides the integrated target frames as the final output of the multi-target tracking device. If neither the tracker nor the detector obtains the target frame, the integration module 204 determines that the tracked target is not present in the current frame. Otherwise, the integration module 204 takes the target box with the greatest conservative similarity as the final target box position.
Therefore, by using the multi-target tracking device according to the embodiment of the application, a plurality of targets can be tracked.
< second embodiment >
Next, a multi-target tracking method according to a second embodiment of the present application will be described with reference to fig. 5.
As shown in fig. 5, the multi-target tracking method 500 according to the present embodiment includes:
step S501: receiving, by a tracking module, video frame data and acquiring indication data indicating a plurality of objects to be tracked, and tracking the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects to generate a tracking result;
step S502: receiving, by a detection module, the video frame data and detecting the plurality of objects in each frame of the video frame data to generate a detection result;
step S503: receiving the video frame data through a learning module, updating the detection module according to the tracking result of the tracking module, and updating the tracking module according to the detection result of the detection module; and
step S504: receiving, by an integration module, a tracking result of the tracking module and a detection result of the detection module to generate indication data indicating a plurality of targets being tracked.
The multi-target tracking method 500 according to the present embodiment can be implemented by the multi-target tracking apparatus 200 according to the first embodiment. Accordingly, detailed descriptions of the respective modules for performing the respective steps of the multi-target tracking method are omitted herein.
Optionally, the tracking module comprises a multi-target tracking manager configured to:
performing addition of a tracking target according to a predetermined operation;
updating a plurality of tracking targets according to a preset operation; and
and managing the bidirectional mapping relation between the first queues of the tracking target frames of the plurality of tracking targets and the second queues of the target area characteristic points of the plurality of tracking targets.
Optionally, the multi-target tracking manager is further configured to:
tracking all the feature points in the second queue of the feature points of the target area by using an optical flow method so as to generate a tracking result;
determining whether the tracking target frame of each of the plurality of tracking targets succeeds in tracking according to the tracking result and the reverse mapping relation between the characteristic points and the tracking target frame;
if the result is successful, recalculating the tracking target frame and updating the feature points, otherwise deleting the tracking target frame of the tracking target and the corresponding feature points from the first queue and the second queue respectively.
Optionally, the learning module comprises:
a sample queue generator configured to generate a positive sample queue and a negative sample queue, wherein the positive sample queue includes a tracking target frame of each tracking target tracked by the tracking module or a detection target frame of each tracking target detected by the detection module in each frame of the video frame data, the negative sample queue is a common negative sample stored, and the negative sample is an area not intersecting the positive sample within a predetermined range near the positive sample, wherein, for each of the plurality of tracking targets, the positive sample queue includes the tracking target frame and the detection target frame of the tracking target, and the negative sample queue includes the common negative sample and positive samples belonging to the positive sample queues of the other tracking targets; and
the classifier is configured to perform similarity comparison on the positive sample queue and the negative sample queue of each tracking target respectively; acquiring the number and the attribution of the positive sample queue and the negative sample queue according to a set threshold value, and executing normalization processing; and calculating probabilities of positive and negative samples belonging to each tracked object.
Optionally, the detection module is further configured to:
for each frame of the video frame data, sliding a detection box to detect a negative sample;
determining the area in which the negative sample is detected as a subsequent non-detection area; and
for the remaining regions, positive sample detection is performed to determine a tracking target.
Therefore, by using the multi-target tracking method according to the embodiment of the application, a plurality of targets can be tracked.
It is to be noted that the above embodiments are merely examples, and the present invention is not limited to such examples, but may be variously modified.
It should be noted that, in the present specification, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
Finally, it should be noted that the series of processes described above includes not only processes performed in time series in the order described herein, but also processes performed in parallel or individually, rather than in time series.
Through the above description of the embodiments, those skilled in the art will clearly understand that the present invention may be implemented by software plus a necessary hardware platform, and may also be implemented by hardware entirely. With this understanding, all or part of the technical solutions of the present invention that contribute to the background can be embodied in the form of a software product, which can be stored in a storage medium, such as a ROM (read only memory)/RAM (random access memory), a magnetic disk, an optical disk, etc., and includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the method according to each embodiment or some parts of the embodiments of the present invention.
The present invention has been described in detail, and the principle and embodiments of the present invention are explained herein by using specific examples, which are only used to help understand the method and the core idea of the present invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.

Claims (10)

1. A multi-target tracking apparatus, comprising:
a tracking module configured to receive video frame data and acquire indication data indicating a plurality of objects to be tracked, and track the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects so as to generate a tracking result;
a detection module configured to receive the video frame data and detect the plurality of objects in each frame of the video frame data to generate a detection result;
the learning module is configured to receive the video frame data, update the detection module according to the tracking result of the tracking module, and update the tracking module according to the detection result of the detection module; and
an integration module configured to receive a tracking result of the tracking module and a detection result of the detection module to generate indication data indicating a plurality of targets being tracked,
the learning module includes:
a sample queue generator configured to generate a positive sample queue and a negative sample queue, respectively, for each of a plurality of tracked targets, wherein the negative sample queue comprises stored common negative samples and positive samples of positive sample queues belonging to other tracked targets.
2. The apparatus of claim 1, wherein the tracking module comprises a multi-target tracking manager configured to:
performing addition of a tracking target according to a predetermined operation;
performing an update of a plurality of tracking targets; and
and managing the bidirectional mapping relation between the first queues of the tracking target frames of the plurality of tracking targets and the second queues of the target area characteristic points of the plurality of tracking targets.
3. The apparatus of claim 2, wherein the multi-target tracking manager is further configured to:
tracking all the feature points in the second queue of the feature points of the target area by using an optical flow method so as to generate a tracking result;
determining whether the tracking target frame of each of the plurality of tracking targets succeeds in tracking according to the tracking result and the reverse mapping relation between the characteristic points and the tracking target frame;
if the result is successful, recalculating the tracking target frame and updating the feature points, otherwise deleting the tracking target frame of the tracking target and the corresponding feature points from the first queue and the second queue respectively.
4. The apparatus of claim 3, wherein the positive sample queue includes a tracking target frame of each tracking target tracked by the tracking module or a detection target frame of each tracking target detected by the detection module in each frame of the video frame data, and the negative sample is a region within a predetermined range around the positive sample that does not intersect the positive sample, wherein the positive sample queue includes a tracking target frame and a detection target frame of the tracking target for each of the plurality of tracking targets,
the learning module further comprises:
the classifier is configured to perform similarity comparison on the positive sample queue and the negative sample queue of each tracking target respectively; acquiring the number and the attribution of the positive sample queue and the negative sample queue according to a set threshold value, and executing normalization processing; and calculating probabilities of positive and negative samples belonging to each tracked object.
5. The apparatus of claim 4, wherein the detection module is further configured to:
for each frame of the video frame data, sliding a detection box to detect a negative sample;
determining the area in which the negative sample is detected as a subsequent non-detection area; and
for the remaining regions, positive sample detection is performed to determine a tracking target.
6. A multi-target tracking method comprises the following steps:
receiving, by a tracking module, video frame data and acquiring indication data indicating a plurality of objects to be tracked, and tracking the plurality of objects in each frame of the video frame data according to the indication data of the plurality of objects to generate a tracking result;
receiving, by a detection module, the video frame data and detecting the plurality of objects in each frame of the video frame data to generate a detection result;
receiving the video frame data through a learning module, updating the detection module according to the tracking result of the tracking module, and updating the tracking module according to the detection result of the detection module; and
receiving, by an integration module, a tracking result of the tracking module and a detection result of the detection module to generate indication data indicating a plurality of targets being tracked,
the learning module includes:
a sample queue generator configured to generate a positive sample queue and a negative sample queue, respectively, for each of a plurality of tracked targets, wherein the negative sample queue comprises stored common negative samples and positive samples of positive sample queues belonging to other tracked targets.
7. The method of claim 6, wherein the tracking module comprises a multi-target tracking manager configured to:
performing addition of a tracking target according to a predetermined operation;
updating a plurality of tracking targets according to a preset operation; and
and managing the bidirectional mapping relation between the first queues of the tracking target frames of the plurality of tracking targets and the second queues of the target area characteristic points of the plurality of tracking targets.
8. The method of claim 7, wherein the multi-target tracking manager is further configured to:
tracking all the feature points in the second queue of the feature points of the target area by using an optical flow method so as to generate a tracking result;
determining whether the tracking target frame of each of the plurality of tracking targets succeeds in tracking according to the tracking result and the reverse mapping relation between the characteristic points and the tracking target frame;
if the result is successful, recalculating the tracking target frame and updating the feature points, otherwise deleting the tracking target frame of the tracking target and the corresponding feature points from the first queue and the second queue respectively.
9. The method of claim 8, wherein the positive sample queue includes a tracking target frame of each tracking target tracked by the tracking module or a detection target frame of each tracking target detected by the detection module in each frame of the video frame data, and the negative sample is a region within a predetermined range around the positive sample that does not intersect the positive sample, wherein the positive sample queue includes the tracking target frame and the detection target frame of the tracking target for each of the plurality of tracking targets,
the learning module further comprises:
the classifier is configured to perform similarity comparison on the positive sample queue and the negative sample queue of each tracking target respectively; acquiring the number and the attribution of the positive sample queue and the negative sample queue according to a set threshold value, and executing normalization processing; and calculating probabilities of positive and negative samples belonging to each tracked object.
10. The method of claim 9, wherein the detection module is further configured to:
for each frame of the video frame data, sliding a detection box to detect a negative sample;
determining the area in which the negative sample is detected as a subsequent non-detection area; and
for the remaining regions, positive sample detection is performed to determine a tracking target.
CN201710203912.2A 2017-03-30 2017-03-30 Multi-target tracking device and method Active CN106971401B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710203912.2A CN106971401B (en) 2017-03-30 2017-03-30 Multi-target tracking device and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710203912.2A CN106971401B (en) 2017-03-30 2017-03-30 Multi-target tracking device and method

Publications (2)

Publication Number Publication Date
CN106971401A CN106971401A (en) 2017-07-21
CN106971401B true CN106971401B (en) 2020-09-25

Family

ID=59337132

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710203912.2A Active CN106971401B (en) 2017-03-30 2017-03-30 Multi-target tracking device and method

Country Status (1)

Country Link
CN (1) CN106971401B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108470332B (en) * 2018-01-24 2023-07-07 博云视觉(北京)科技有限公司 Multi-target tracking method and device
CN108734107B (en) * 2018-04-24 2021-11-05 武汉幻视智能科技有限公司 Multi-target tracking method and system based on human face
CN109325961B (en) * 2018-08-27 2021-07-09 北京悦图数据科技发展有限公司 Unmanned aerial vehicle video multi-target tracking method and device
CN109829397B (en) * 2019-01-16 2021-04-02 创新奇智(北京)科技有限公司 Video annotation method and system based on image clustering and electronic equipment
CN110400329B (en) * 2019-06-17 2022-04-05 北京百度网讯科技有限公司 People flow counting method and system
CN111401228B (en) * 2020-03-13 2023-12-19 中科创达软件股份有限公司 Video target labeling method and device and electronic equipment
CN111445501B (en) * 2020-03-25 2023-03-24 苏州科达科技股份有限公司 Multi-target tracking method, device and storage medium
CN113486820B (en) * 2021-07-09 2023-06-06 厦门理工学院 Bidirectional target tracking method and system based on efficient template updating and selecting mechanism
CN113570637B (en) * 2021-08-10 2023-09-19 中山大学 Multi-target tracking method, device, equipment and storage medium
CN117011736A (en) * 2022-04-28 2023-11-07 北京字跳网络技术有限公司 Multi-target tracking method, device, equipment and readable storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101794382A (en) * 2010-03-12 2010-08-04 华中科技大学 Method for counting passenger flow of buses in real time
CN104331901A (en) * 2014-11-26 2015-02-04 北京邮电大学 TLD-based multi-view target tracking device and method
CN105787484A (en) * 2014-12-24 2016-07-20 深圳市Tcl高新技术开发有限公司 Object tracking or identifying method and object tracking or identifying device
CN106250878A (en) * 2016-08-19 2016-12-21 中山大学 A kind of combination visible ray and the multi-modal method for tracking target of infrared image

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106204649B (en) * 2016-07-05 2019-06-28 西安电子科技大学 A kind of method for tracking target based on TLD algorithm

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101794382A (en) * 2010-03-12 2010-08-04 华中科技大学 Method for counting passenger flow of buses in real time
CN104331901A (en) * 2014-11-26 2015-02-04 北京邮电大学 TLD-based multi-view target tracking device and method
CN105787484A (en) * 2014-12-24 2016-07-20 深圳市Tcl高新技术开发有限公司 Object tracking or identifying method and object tracking or identifying device
CN106250878A (en) * 2016-08-19 2016-12-21 中山大学 A kind of combination visible ray and the multi-modal method for tracking target of infrared image

Also Published As

Publication number Publication date
CN106971401A (en) 2017-07-21

Similar Documents

Publication Publication Date Title
CN106971401B (en) Multi-target tracking device and method
US10885365B2 (en) Method and apparatus for detecting object keypoint, and electronic device
CN107886048B (en) Target tracking method and system, storage medium and electronic terminal
CN108388879B (en) Target detection method, device and storage medium
CN110766724B (en) Target tracking network training and tracking method and device, electronic equipment and medium
US8355576B2 (en) Method and system for crowd segmentation
JP7093427B2 (en) Object tracking methods and equipment, electronic equipment and storage media
US20120250983A1 (en) Object detecting apparatus and method
CN111104925B (en) Image processing method, image processing apparatus, storage medium, and electronic device
CN111512317A (en) Multi-target real-time tracking method and device and electronic equipment
US11783384B2 (en) Computer vision systems and methods for automatically detecting, classifying, and pricing objects captured in images or videos
CN110363165B (en) Multi-target tracking method and device based on TSK fuzzy system and storage medium
CN115375917B (en) Target edge feature extraction method, device, terminal and storage medium
CN110942473A (en) Moving target tracking detection method based on characteristic point gridding matching
CN110490058B (en) Training method, device and system of pedestrian detection model and computer readable medium
CN111932545A (en) Image processing method, target counting method and related device thereof
CN113052019A (en) Target tracking method and device, intelligent equipment and computer storage medium
JP2019021297A (en) Image processing device and method, and electronic apparatus
CN113033305B (en) Living body detection method, living body detection device, terminal equipment and storage medium
KR20150137698A (en) Method and apparatus for movement trajectory tracking of moving object on animal farm
JP2014203133A (en) Image processing device and image processing method
CN111860261A (en) Passenger flow value statistical method, device, equipment and medium
CN111640076A (en) Image completion method and device and electronic equipment
CN111967403A (en) Video moving area determining method and device and electronic equipment
KR102485359B1 (en) Method for enhancing in situ adaptive artfitial intelligent model

Legal Events

Date Code Title Description
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant