CN105261040A - Multi-target tracking method and apparatus - Google Patents

Multi-target tracking method and apparatus Download PDF

Info

Publication number
CN105261040A
CN105261040A CN201510676986.9A CN201510676986A CN105261040A CN 105261040 A CN105261040 A CN 105261040A CN 201510676986 A CN201510676986 A CN 201510676986A CN 105261040 A CN105261040 A CN 105261040A
Authority
CN
China
Prior art keywords
sub
image area
frame
image
image frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510676986.9A
Other languages
Chinese (zh)
Other versions
CN105261040B (en
Inventor
傅慧源
马华东
张�诚
方瑞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Posts and Telecommunications
Original Assignee
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Posts and Telecommunications filed Critical Beijing University of Posts and Telecommunications
Priority to CN201510676986.9A priority Critical patent/CN105261040B/en
Publication of CN105261040A publication Critical patent/CN105261040A/en
Application granted granted Critical
Publication of CN105261040B publication Critical patent/CN105261040B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24147Distances to closest patterns, e.g. nearest neighbour classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the invention provides a multi-target tracking method and apparatus. The method comprises: determining the first type of prediction areas in a current target picture frame of each target object in a reference picture frame, and the gradient amplitude graph corresponding to the target picture frame; utilizing a preserved window to successively slide along pixels in the gradient amplitude graph to obtain a plurality of sub image areas generated in the sliding process of the preserved window; screening a first type of sub image areas including objects from the sub image areas; and respectively coupling the first type of sub image areas in the target picture frame with the first type of prediction areas in a current target picture frame of each target object in a reference picture frame, and determining the area of the target object in the target image frame in a coupling mode, instead of directly predicting only based on the area information of the target object of a previous frame of image, thereby improving multi-target tracking accuracy.

Description

A kind of multi-object tracking method and device
Technical field
The present invention relates to computer vision field, particularly a kind of multi-object tracking method and device.
Background technology
Computer vision is mainly studied the identification of destination object, the location of destination object and the motion analysis to destination object, multiple target tracking is one of basic problem of computer vision field, its application is very extensive, the main task of multiple target tracking is in one section of continuous print video file, orient the position at interested target place in every piece image, thus the tracking realized interested target, multiple target tracking is computer vision field very part and parcel, only have the accurate tracking realized target, succeeding target behavioural analysis work could be launched.
Wherein, existing Multitarget Tracking be according to destination object at the area information target of prediction object of previous frame image in the region of current image frame, using doped region as the region of destination object in current image frame.
But because the change in location of destination object has ambiguity, the area information of the destination object of this only foundation previous frame image carrys out direct prediction mode, causes multiple target tracking accuracy rate lower.
Summary of the invention
The object of the embodiment of the present invention is to provide a kind of multi-object tracking method and device, and improve multiple target tracking accuracy rate, technical scheme is as follows:
Determine that in reference image frame, each destination object is in the first kind estimation range of current target image frame, wherein, described reference image frame and described target image frame are the different images frame in video file, and reference image frame is consecutive previous frame image with target image frame; Wherein, described destination object is at least one in multiple subject included in described reference image frame;
Determine the gradient magnitude figure corresponding to described target image frame;
Utilize predetermined window to slide by pixel in described gradient magnitude figure, obtain multiple sub-image areas that described predetermined window generates in sliding process;
From multiple sub-image area, screening obtains the first kind sub-image area comprising subject;
Perform presumptive area analysis operation to each first kind sub-image area, wherein, described predetermined analysis operation at least comprises:
Images match is carried out with first kind estimation range corresponding to each destination object respectively in current first kind sub-image area, if there is the target first kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target first kind estimation range.
Optionally, described regional analysis operation also comprises:
If there is no matching rate is greater than the target first kind estimation range of predetermined threshold value, then when there is at least one two field picture before judging described reference image frame:
From before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame corresponding to this first kind sub-image area, and determine that in current auxiliary view picture frame, each destination object is in the Equations of The Second Kind estimation range of current target image frame, images match is carried out with each Equations of The Second Kind estimation range respectively in this first kind sub-image area, if there is the target Equations of The Second Kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target second estimation range, otherwise, continue to perform from before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame step corresponding to this first kind sub-image area, until each two field picture before reference image frame is all by as the auxiliary view picture frame corresponding to this first kind sub-image area.
Optionally, the described gradient magnitude figure determined corresponding to described target image frame, comprising:
Calculate the gradient magnitude of each pixel position in described target image frame;
Based on the gradient magnitude of calculated each pixel position, obtain the gradient magnitude figure corresponding to described target image frame;
Wherein, the formula that the gradient magnitude calculating each pixel position in described target image frame utilizes is:
f=min(|g x|+|g y|,255)
g x=I(x+1,y)-I(x,y)
g y=I(x,y+1)-I(x,y)
Wherein, f is the gradient magnitude of pixel position (x, y), g xfor the horizontal gradient of image slices vegetarian refreshments position (x, y), g yfor the VG (vertical gradient) of this pixel position (x, y) of image, I (x, y) is the pixel of pixel position (x, y).
Optionally, described from multiple sub-image area, screening obtains the first kind sub-image area comprising subject, comprising:
Calculate the score of each sub-image area in described multiple subimage;
Based on described score, descending sort is carried out to multiple sub-image area, extract at least one sub-image area that sorting position belongs to precalculated position scope;
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the first kind sub-image area comprising subject;
Wherein, the formula that the score of each sub-image area calculated in described multiple subimage utilizes is:
S = ω · f T = Σ k = 1 64 x k y k
Wherein, S is sub-image area score, ω=(x 1, x 2..., x 64) for presetting proper vector, f=(y 1, y 2..., y 64) vector form of the two-value gradient magnitude of sub-image area that generates in sliding process for window, k is dimension.
Optionally, described use Supplement device and nearest neighbor classifier, screen at least one extracted sub-image area, obtains the first kind sub-image area comprising subject, comprising:
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the selection result;
Using the selection result as the first kind sub-image area comprising subject;
Or,
Described use Supplement device and nearest neighbor classifier, screen at least one extracted sub-image area, obtains the first kind sub-image area comprising subject, comprising:
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the selection result;
Judge whether each sub-image area in the selection result is predetermined background image region;
The sub-image area of background image region will not be belonged to as the first kind sub-image area comprising subject in the selection result.
Embodiments provide a kind of multiple target tracking device, comprising:
First determination module, for determining that in reference image frame, each destination object is in the first kind estimation range of current target image frame, wherein, described reference image frame and described target image frame are the different images frame in video file, and reference image frame is consecutive previous frame image with target image frame; Wherein, described destination object is at least one in multiple subject included in described reference image frame;
Second determination module, for determining the gradient magnitude figure corresponding to described target image frame;
Sub-image area generation module, for utilizing predetermined window to slide by pixel in described gradient magnitude figure, obtains multiple sub-image areas that described predetermined window generates in sliding process;
Screening module, for from multiple sub-image area, screens and obtains the first kind sub-image area comprising subject;
Analysis module, for performing presumptive area analysis operation to each first kind sub-image area, wherein, described predetermined analysis operation at least comprises:
Images match is carried out with first kind estimation range corresponding to each destination object respectively in current first kind sub-image area, if there is the target first kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target first kind estimation range.
Optionally, the presumptive area analysis operation performed by described analysis module also comprises:
If there is no matching rate is greater than the target first kind estimation range of predetermined threshold value, then when there is at least one two field picture before judging described reference image frame:
From before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame corresponding to this first kind sub-image area, and determine that in current auxiliary view picture frame, each destination object is in the Equations of The Second Kind estimation range of current target image frame, images match is carried out with each Equations of The Second Kind estimation range respectively in this first kind sub-image area, if there is the target Equations of The Second Kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target second estimation range, otherwise, continue to perform from before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame step corresponding to this first kind sub-image area, until each two field picture before reference image frame is all by as the auxiliary view picture frame corresponding to this first kind sub-image area.
Optionally, described second determination module comprises:
First calculating sub module, for calculating the gradient magnitude of each pixel position in described target image frame;
Obtain submodule, for the gradient magnitude based on calculated each pixel position, obtain the gradient magnitude figure corresponding to described target image frame;
Wherein, the formula that the gradient magnitude calculating each pixel position in described target image frame utilizes is:
f=min(|g x|+|g y|,255)
g x=I(x+1,y)-I(x,y)
g y=I(x,y+1)-I(x,y)
Wherein, f is the gradient magnitude of pixel position (x, y), g xfor the horizontal gradient of image slices vegetarian refreshments position (x, y), g yfor the VG (vertical gradient) of this pixel position (x, y) of image, I (x, y) is the pixel of pixel position (x, y).
Optionally, described screening module, comprising:
Second calculating sub module, for calculating the score of each sub-image area in described multiple subimage;
Extracting submodule, for carrying out descending sort based on described score to multiple sub-image area, extracting at least one sub-image area that sorting position belongs to precalculated position scope;
Classification submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the first kind sub-image area comprising subject;
Wherein, the formula that the score of each sub-image area calculated in described multiple subimage utilizes is:
S = ω · f T = Σ k = 1 64 x k y k
Wherein, S is sub-image area score, ω=(x 1, x 2..., x 64) for presetting proper vector, f=(y 1, y 2..., y 64) vector form of the two-value gradient magnitude of sub-image area that generates in sliding process for window, k is dimension.
Optionally, described sort module, comprising:
First screening submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the selection result;
First obtains submodule, for using the selection result as the first kind sub-image area comprising subject;
Or,
Described sort module, comprising:
Second screening submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the selection result;
Judge submodule, for judging whether each sub-image area in the selection result is predetermined background image region;
Second obtains submodule, for not belonging to the sub-image area of background image region in the selection result as the first kind sub-image area comprising subject.
Visible, in the embodiment of the present invention, propose a kind of multi-object tracking method and device, by the first kind sub-image area in target image frame is carried out images match with each destination object in reference image frame in the first kind estimation range of current target image frame respectively, by the region of mode determination destination object in described target image frame of coupling, and not only directly predict according to the area information of the destination object of previous frame image, therefore, it is possible to improve multiple target tracking accuracy rate.Certainly, arbitrary product of the present invention is implemented or method must not necessarily need to reach above-described all advantages simultaneously.
Accompanying drawing explanation
In order to be illustrated more clearly in the embodiment of the present invention or technical scheme of the prior art, be briefly described to the accompanying drawing used required in embodiment or description of the prior art below, apparently, accompanying drawing in the following describes is only some embodiments of the present invention, for those of ordinary skill in the art, under the prerequisite not paying creative work, other accompanying drawing can also be obtained according to these accompanying drawings.
The schematic flow sheet of a kind of multi-object tracking method that Fig. 1 provides for the embodiment of the present invention;
Fig. 2 is the intermediate result proper vector ω=(x of training linear SVM classifier 1, x 2..., x 64) graph-based figure;
Fig. 3 is that a kind of multi-object tracking method in this paper and original TLD method are at process one frame temporal comparison diagram used, a () is for the AVSS-PV-Medium video requency frame data collection in AVSS2007 file at process one frame temporal comparison diagram used, (b) is at process one frame temporal comparison diagram used for the AVSS-PV-Hard video requency frame data collection in AVSS2007 file;
Fig. 4 is a kind of multi-object tracking method in this paper and the comparison diagram of original TLD method in accuracy rate, a () is for the comparison diagram of AVSS-PV-Medium video requency frame data collection in accuracy rate in AVSS2007 file, (b) is for the comparison diagram of AVSS-PV-Hard video requency frame data collection in accuracy rate in AVSS2007 file;
Fig. 5 is the tracking results schematic diagram of a kind of multi-object tracking method in this paper;
The structural representation of a kind of multiple target tracking device that Fig. 6 provides for the embodiment of the present invention.
Embodiment
Below in conjunction with the accompanying drawing in the embodiment of the present invention, be clearly and completely described the technical scheme in the embodiment of the present invention, obviously, described embodiment is only the present invention's part embodiment, instead of whole embodiments.Based on the embodiment in the present invention, those of ordinary skill in the art, not making the every other embodiment obtained under creative work prerequisite, belong to the scope of protection of the invention.
Below by specific embodiment, the present invention is described in detail.
The schematic flow sheet of a kind of multi-object tracking method that Fig. 1 provides for the embodiment of the present invention, comprises the steps:
S101: determine that in reference image frame, each destination object is in the first kind estimation range of current target image frame.
Wherein, described reference image frame and described target image frame are the different images frame in video file, and reference image frame is consecutive previous frame image with target image frame; Wherein, described destination object is at least one in multiple subject included in described reference image frame.
Read the two field picture in video file, using this two field picture as reference picture frame, force to be adjusted to fixed measure with reference to image frame dimension according to actual service condition, to accelerate computing, such as: fixed measure is 320 × 240 pixels, use and select the optical flow method tracker of tracking point to predict first estimation range of each destination object at current target image frame with FB (Forward-BackwardError just swings to mistake) error approach, concrete steps are as follows:
Evenly get a little according to grid node in reference image frame targeted object region, such as: sizing grid is 10 × 10 pixels;
Then, for wherein each pixel position (x1, y1), follow the trail of its position to current target image frame with optical flow method tracker, then the position (x2, y2) of reference image frame is gone back in backward tracing, calculates FB error:
F B - E R R O R = ( x 2 - x 1 ) 2 - ( y 2 - y 1 ) 2
Wherein, FB-ERROR is the FB error between pixel position (x1, y1) and pixel position (x2, y2).
Filter out this FB error minimum front 50% pixel, go out the first kind estimation range of destination object at current target image frame according to the changes in coordinates of these pixels and the change calculations of distance, namely destination object is in the predicted position of current target image frame.
In addition, the residual error of each pixel position displacement is calculated:
d=|d i-d m|
Wherein, d is the residual error of pixel position displacement, d ifor the displacement of a pixel position in grid, d mfor the median of pixel position displacement amounts all in grid.
If the median of the residual error of all pixel position displacements is more than 10 pixel units in grid, so optical flow method tracker follows the trail of the objective object unsuccessfully, now, cannot obtain the first kind estimation range of destination object at current target image frame.
S102: determine the gradient magnitude figure corresponding to described target image frame.
Concrete, the described gradient magnitude figure determined corresponding to described target image frame, comprising:
Calculate the gradient magnitude of each pixel position in described target image frame;
Based on the gradient magnitude of calculated each pixel position, obtain the gradient magnitude figure corresponding to described target image frame;
Wherein, the formula that the gradient magnitude calculating each pixel position in described target image frame utilizes is:
f=min(|g x|+|g y|,255)
g x=I(x+1,y)-I(x,y)
g y=I(x,y+1)-I(x,y)
Wherein, f is the gradient magnitude of pixel position (x, y), g xfor the horizontal gradient of image slices vegetarian refreshments position (x, y), g yfor the VG (vertical gradient) of this pixel position (x, y) of image, I (x, y) is the pixel of pixel position (x, y).
Further, according to actual service condition, target image frame size can be scaled fixed measure, so that the gradient magnitude carrying out each pixel position in target image frame calculates.
S103: utilize predetermined window to slide by pixel in described gradient magnitude figure, obtain multiple sub-image areas that described predetermined window generates in sliding process.
According to actual service condition, described predetermined window can be the window of 8 × 8 pixels, and can be the window of 10 × 10 pixels, certain embodiment be not limited in this.
According to actual service condition, the mode that described predetermined window is slided by pixel in described gradient magnitude figure can be from top to bottom, and order from left to right, can also be from the bottom up, the order of turning left from the right side, certainly, embodiment be not limited in this.
S104: from multiple sub-image area, screening obtains the first kind sub-image area comprising subject.
Concrete, calculate the score of each sub-image area in described multiple subimage;
Based on described score, descending sort is carried out to multiple sub-image area, extract at least one sub-image area that sorting position belongs to precalculated position scope;
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the first kind sub-image area comprising subject;
Wherein, the formula that the score of each sub-image area calculated in described multiple subimage utilizes is:
S = ω · f T = Σ k = 1 64 x k y k
Wherein, S is sub-image area score, ω=(x 1, x 2..., x 64) for presetting proper vector, f=(y 1, y 2..., y 64) vector form of the two-value gradient magnitude of sub-image area that generates in sliding process for window, k is dimension.
Concrete, described use Supplement device and nearest neighbor classifier, screen at least one extracted sub-image area, obtains the first kind sub-image area comprising subject, comprising:
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the selection result;
Using the selection result as the first kind sub-image area comprising subject;
Or,
Described use Supplement device and nearest neighbor classifier, screen at least one extracted sub-image area, obtains the first kind sub-image area comprising subject, comprising:
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the selection result;
Judge whether each sub-image area in the selection result is predetermined background image region;
The sub-image area of background image region will not be belonged to as the first kind sub-image area comprising subject in the selection result.
Wherein, default proper vector ω=(x is obtained 1, x 2..., x 64) step as follows:
Obtain Linear SVM (SupportVectorMachine, support vector machine) sorter;
Obtain the region of all objects object in training plan picture frame;
Determine the gradient magnitude figure corresponding to training plan picture frame;
By the area reduction of subject to preset range, such as: 8 × 8 pixels, accelerating algorithm is used to calculate the two-value gradient magnitude in each subject region;
Use the two-value gradient magnitude training linear SVM classifier of the two-value gradient magnitude in subject region and the background image region of grab sample respectively, output characteristic vector ω=(x 1, x 2..., x 64), by the proper vector ω=(x of this 64 dimension 1, x 2..., x 64) as presetting proper vector, the intermediate result proper vector ω=(x of training linear SVM classifier 1, x 2..., x 64) graph-based as shown in Figure 2.
For precalculated position scope, determine according to actual service condition, such as: from 1-2000 position range.
Further, according to actual service condition, the target image frame size at its place can be scaled former target image frame size at least one sub-image area of extracting, use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened.
Described fern sorter comprises n fundamental classifier, and each fundamental classifier is made m to sub-image area and compared grey scale pixel value, and producing length is the binary code of m, and this binary code one has 2 mkind.Wherein, the first kind sub-image area comprising subject produces the binary code x that length is m.When sub-image area quantity is sufficient, the posterior probability that sub-image area is the first kind sub-image area comprising subject must be arrived:
p i = # p # p + # n
Wherein, p ifor arriving i-th fundamental classifier the posterior probability that sub-image area is the first kind sub-image area comprising subject, i is i-th fundamental classifier, #p meets the first kind sub-image area quantity comprising subject that binary code is x, the sub-image area quantity of #n to be that not meet binary code be the sub-image area of x be background image.
The corresponding posterior probability that each sub-image area obtains through each fundamental classifier is averaged, average is greater than the sub-image area of 50% as the first kind sub-image area comprising subject.
Then, being screened described sub-image area further by nearest neighbor classifier, is pre-set specifications form according to actual service condition adjustment sub-image area, such as: 15 × 15 pixels, then sub-image area is the vector of 255 dimensions, and the similarity between two sub-image areas is:
S(j,k)=0.5(NCC(j,k)+1)
Wherein, S (j, k) is normalized correlation coefficient for the similarity between sub-image area j and sub-image area k, NCC, utilizes following formula to obtain NCC (j, k):
N C C ( j , k ) = Σ n = 1 255 j ( n ) k ( n ) Σ n = 1 255 j ( n ) 2 Σ n = 1 255 k ( n ) 2
Wherein, NCC (j, k) is the normalized correlation coefficient between sub-image area j and sub-image area k, the n-th vector that j (n) is sub-image area j, the n-th vector that k (n) is sub-image area k.
Suppose that sub-image area be the sub-image area of the first kind sub-image area comprising subject is positive sub-image region, sub-image area is the sub-image area of background image is negative sub-image area, calculates positive arest neighbors similarity and negative arest neighbors similarity respectively:
S + ( j , M ) = max j + ∈ M S ( j , j + )
S - ( j , M ) = max j - ∈ M S ( j , j - )
Wherein, S +the positive arest neighbors similarity that (j, M) is sub-image area j, S -the negative arest neighbors similarity that (j, M) is sub-image area j, M is the set M=(j of positive and negative sub-image area composition +, k +, z +, j -, k -, z -), j +for positive sub-image region, j -for negative sub-image area.
Calculate the relative similarity of sub-image area j:
S j r = S + ( j , M ) S + ( j , M ) + S - ( j , M )
Wherein, for the relative similarity of sub-image area j, S +the positive arest neighbors similarity that (j, M) is sub-image area j, S -the negative arest neighbors similarity that (j, M) is sub-image area j.
This relative similarity span is [0,1], and value shows more greatly sub-image area more close to the region of destination object in target image frame.
If the relative similarity of this sub-image area j meet:
S j r > θ N N
Wherein, θ nNfor predetermined coefficient, value is 0.6;
Then determine that this sub-image area is the first kind sub-image area comprising subject.
In another embodiment of the invention, sub-image area due to described extraction can be the sub-image area of inclusion body object, also can for comprising the sub-image area of background image, therefore further judgement is done to screening each sub-image area obtained, judge to screen whether each sub-image area obtained is predetermined background image region, will not belong to the sub-image area of background image region in the selection result as the first kind sub-image area comprising subject.
S105: presumptive area analysis operation is performed to each first kind sub-image area.
Wherein, described predetermined analysis operation at least comprises:
Images match is carried out with first kind estimation range corresponding to each destination object respectively in current first kind sub-image area, if there is the target first kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target first kind estimation range.
Concrete, described regional analysis operation also comprises:
If there is no matching rate is greater than the target first kind estimation range of predetermined threshold value, then when there is at least one two field picture before judging described reference image frame:
From before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame corresponding to this first kind sub-image area, and determine that in current auxiliary view picture frame, each destination object is in the Equations of The Second Kind estimation range of current target image frame, images match is carried out with each Equations of The Second Kind estimation range respectively in this first kind sub-image area, if there is the target Equations of The Second Kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target second estimation range, otherwise, continue to perform from before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame step corresponding to this first kind sub-image area, until each two field picture before reference image frame is all by as the auxiliary view picture frame corresponding to this first kind sub-image area.
Such as: using the 5th frame in video file as reference picture frame, then the 6th frame is current target image frame, 4th frame is auxiliary view picture frame, destination object A is had in reference image frame, if there is no the matching rate that in the 6th frame, in the first kind sub-image area of destination object A and the 5th frame, the first kind estimation range of destination object A matches is greater than the first kind estimation range of the destination object A of predetermined threshold value, then select the 3rd frame as current auxiliary view picture frame, and determine the Equations of The Second Kind estimation range of destination object A in the 4th frame in the 3rd frame, the first kind estimation range of destination object A in the 6th frame is mated with the Equations of The Second Kind estimation range of destination object A in the 4th frame, if there is the Equations of The Second Kind estimation range that matching rate is greater than the destination object A of predetermined threshold value, then by the first kind sub-image area of this destination object A, as the region of described destination object A in the 5th frame, otherwise, continue to perform from the 4th frame to the 1st frame, select near the 5th frame a two field picture as destination object A first kind sub-image area corresponding to current auxiliary view picture frame step, such as continue selection the 2nd frame as current auxiliary view picture frame, until the auxiliary view picture frame of each two field picture all corresponding to the first kind sub-image area as this destination object A before the 4th frame.
Visible, in the embodiment of the present invention, propose a kind of multi-object tracking method, by the first kind sub-image area in target image frame is carried out images match with each destination object in reference image frame in the first kind estimation range of current target image frame respectively, by the region of mode determination destination object in described target image frame of coupling, and not only directly predict according to the area information of the destination object of previous frame image, therefore, it is possible to improve multiple target tracking accuracy rate.
Fig. 3 is that a kind of multi-object tracking method in this paper and original TLD method are at process one frame temporal comparison diagram used, a () is for the AVSS-PV-Medium video requency frame data collection in AVSS2007 file at process one frame temporal comparison diagram used, (b) is at process one frame temporal comparison diagram used for the AVSS-PV-Hard video requency frame data collection in AVSS2007 file.
Wherein, original TLD (Tracking-Learning-Detection, follow the trail of study to detect) method is a kind of method for multiple target tracking, indicated by the solid line in the drawings, proposed method is represented by dashed line, AVSS2007 (AdvancedVideoandSignalbasedSurveillance2007, advanced video and input 2007) file is video requency frame data collection file for multiple target tracking, AVSS-PV-Medium video requency frame data collection wherein and AVSS-PV-Hard video requency frame data collection are two kinds of difficulty type video requency frame data collection in video requency frame data collection file, i.e. medium difficulty video requency frame data collection and difficult difficulty video requency frame data collection.
Visible, along with the increase of tracking target number of objects, the processing time of proposed multi-object tracking method is also in increase, but total processing time remain hundred Milliseconds other, when tracking target number of objects is increased to 7 destination objects, the time of proposed multi-object tracking method process one frame is not also more than 160ms, when destination object number is more than 2, the processing time of original TLD method has been thousand milliseconds of ranks, so proposed multi-object tracking method is compared, TLD method has very large lifting on the processing time.
Fig. 4 is a kind of multi-object tracking method in this paper and the comparison diagram of original TLD method in accuracy rate, a () is for the comparison diagram of AVSS-PV-Medium video requency frame data collection in accuracy rate in AVSS2007 file, (b) is for the comparison diagram of AVSS-PV-Hard video requency frame data collection in accuracy rate in AVSS2007 file.
Wherein, original TLD method is indicated by the solid line in the drawings, and proposed method is represented by dashed line.
If the Euclidean distance between the destination object center position coordinates of prediction and destination object real center position coordinates exceedes threshold value, be defined as target of prediction object's position correct, accuracy rate is the ratio of the correct destination object number of prediction and destination object sum.
Visible, when destination object number is little, such as 1 or 2 destination objects time, multi-object tracking method in this paper can keep high accuracy rate, but when destination object number starts to increase gradually, the accuracy rate of multi-object tracking method in this paper has decline in various degree under different scene, under medium Medium difficulty, even if destination object number is increased to 7, multi-object tracking method in this paper still can keep the accuracy rate of more than 86%; Under difficult Hard difficulty, along with the increase of destination object number, it is very fast that the accuracy rate of tracking declines, and when destination object number is increased to 7, the accuracy rate of tracking drops to about 80%.
The tracking results schematic diagram of Fig. 5 a kind of multi-object tracking method in this paper.
Wherein, the current frame number of numeral in every width figure lower right corner.
The structural representation of a kind of multiple target tracking device that Fig. 6 provides for the embodiment of the present invention, corresponding with the process flow diagram shown in Fig. 1, comprise the first determination module 601, second determination module 602, sub-image area generation module 603, screening module 604 and analysis module 605.
Wherein, described first determination module 601, for determining that in reference image frame, each destination object is in the first kind estimation range of current target image frame, wherein, described reference image frame and described target image frame are the different images frame in video file, and reference image frame is consecutive previous frame image with target image frame; Wherein, described destination object is at least one in multiple subject included in described reference image frame;
Described second determination module 602, for determining the gradient magnitude figure corresponding to described target image frame;
Described sub-image area generation module 603, for utilizing predetermined window to slide by pixel in described gradient magnitude figure, obtains multiple sub-image areas that described predetermined window generates in sliding process;
Described screening module 604, for from multiple sub-image area, screens and obtains the first kind sub-image area comprising subject;
Described analysis module 605, for performing presumptive area analysis operation to each first kind sub-image area, wherein, described predetermined analysis operation at least comprises:
Images match is carried out with first kind estimation range corresponding to each destination object respectively in current first kind sub-image area, if there is the target first kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target first kind estimation range.
Presumptive area analysis operation performed by described analysis module also comprises:
If there is no matching rate is greater than the target first kind estimation range of predetermined threshold value, then when there is at least one two field picture before judging described reference image frame:
From before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame corresponding to this first kind sub-image area, and determine that in current auxiliary view picture frame, each destination object is in the Equations of The Second Kind estimation range of current target image frame, images match is carried out with each Equations of The Second Kind estimation range respectively in this first kind sub-image area, if there is the target Equations of The Second Kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target second estimation range, otherwise, continue to perform from before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame step corresponding to this first kind sub-image area, until each two field picture before reference image frame is all by as the auxiliary view picture frame corresponding to this first kind sub-image area.
Described second determination module 602 comprises:
First calculating sub module, for calculating the gradient magnitude of each pixel position in described target image frame;
Obtain submodule, for the gradient magnitude based on calculated each pixel position, obtain the gradient magnitude figure corresponding to described target image frame;
Wherein, the formula that the gradient magnitude calculating each pixel position in described target image frame utilizes is:
f=min(|g x|+|g y|,255)
g x=I(x+1,y)-I(x,y)
g y=I(x,y+1)-I(x,y)
Wherein, f is the gradient magnitude of pixel position (x, y), g xfor the horizontal gradient of image slices vegetarian refreshments position (x, y), g yfor the VG (vertical gradient) of this pixel position (x, y) of image, I (x, y) is the pixel of pixel position (x, y).
Described screening module 604, comprising:
Second calculating sub module, for calculating the score of each sub-image area in described multiple subimage;
Extracting submodule, for carrying out descending sort based on described score to multiple sub-image area, extracting at least one sub-image area that sorting position belongs to precalculated position scope;
Classification submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the first kind sub-image area comprising subject;
Wherein, the formula that the score of each sub-image area calculated in described multiple subimage utilizes is:
S = ω · f T = Σ k = 1 64 x k y k
Wherein, S is sub-image area score, ω=(x 1, x 2..., x 64) for presetting proper vector, f=(y 1, y 2..., y 64) vector form of the two-value gradient magnitude of sub-image area that generates in sliding process for window, k is dimension.
Described sort module, comprising:
First screening submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the selection result;
First obtains submodule, for using the selection result as the first kind sub-image area comprising subject;
Or,
Described sort module, comprising:
Second screening submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the selection result;
Judge submodule, for judging whether each sub-image area in the selection result is predetermined background image region;
Second obtains submodule, for not belonging to the sub-image area of background image region in the selection result as the first kind sub-image area comprising subject.
Visible, in the embodiment of the present invention, propose a kind of multi-object tracking method and device, by the first kind sub-image area in target image frame is carried out images match with each destination object in reference image frame in the first kind estimation range of current target image frame respectively, by the region of mode determination destination object in described target image frame of coupling, and not only directly predict according to the area information of the destination object of previous frame image, therefore, it is possible to improve multiple target tracking accuracy rate.
It should be noted that, in this article, the relational terms of such as first and lower first-class and so on is only used for an entity or operation to separate with another entity or operational zone, and not necessarily requires or imply the relation that there is any this reality between these entities or operation or sequentially.And, term " comprises ", " comprising " or its any other variant are intended to contain comprising of nonexcludability, thus make to comprise the process of a series of key element, method, article or equipment and not only comprise those key elements, but also comprise other key elements clearly do not listed, or also comprise by the intrinsic key element of this process, method, article or equipment.When not more restrictions, the key element limited by statement " comprising ... ", and be not precluded within process, method, article or the equipment comprising described key element and also there is other identical element.
Each embodiment in this instructions all adopts relevant mode to describe, between each embodiment identical similar part mutually see, what each embodiment stressed is the difference with other embodiments.Especially, for system embodiment, because it is substantially similar to embodiment of the method, so description is fairly simple, relevant part illustrates see the part of embodiment of the method.
The foregoing is only preferred embodiment of the present invention, be not intended to limit protection scope of the present invention.All any amendments done within the spirit and principles in the present invention, equivalent replacement, improvement etc., be all included in protection scope of the present invention.

Claims (10)

1. a multi-object tracking method, is characterized in that, comprising:
Determine that in reference image frame, each destination object is in the first kind estimation range of current target image frame, wherein, described reference image frame and described target image frame are the different images frame in video file, and reference image frame is consecutive previous frame image with target image frame; Wherein, described destination object is at least one in multiple subject included in described reference image frame;
Determine the gradient magnitude figure corresponding to described target image frame;
Utilize predetermined window to slide by pixel in described gradient magnitude figure, obtain multiple sub-image areas that described predetermined window generates in sliding process;
From multiple sub-image area, screening obtains the first kind sub-image area comprising subject;
Perform presumptive area analysis operation to each first kind sub-image area, wherein, described predetermined analysis operation at least comprises:
Images match is carried out with first kind estimation range corresponding to each destination object respectively in current first kind sub-image area, if there is the target first kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target first kind estimation range.
2. method according to claim 1, is characterized in that, described regional analysis operation also comprises:
If there is no matching rate is greater than the target first kind estimation range of predetermined threshold value, then when there is at least one two field picture before judging described reference image frame:
From before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame corresponding to this first kind sub-image area, and determine that in current auxiliary view picture frame, each destination object is in the Equations of The Second Kind estimation range of current target image frame, images match is carried out with each Equations of The Second Kind estimation range respectively in this first kind sub-image area, if there is the target Equations of The Second Kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target second estimation range, otherwise, continue to perform from before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame step corresponding to this first kind sub-image area, until each two field picture before reference image frame is all by as the auxiliary view picture frame corresponding to this first kind sub-image area.
3. method according to claim 1, is characterized in that, the described gradient magnitude figure determined corresponding to described target image frame, comprising:
Calculate the gradient magnitude of each pixel position in described target image frame;
Based on the gradient magnitude of calculated each pixel position, obtain the gradient magnitude figure corresponding to described target image frame;
Wherein, the formula that the gradient magnitude calculating each pixel position in described target image frame utilizes is:
f=min(|g x|+|g y|,255)
g x=I(x+1,y)-I(x,y)
g y=I(x,y+1)-I(x,y)
Wherein, f is the gradient magnitude of pixel position (x, y), g xfor the horizontal gradient of image slices vegetarian refreshments position (x, y), g yfor the VG (vertical gradient) of this pixel position (x, y) of image, I (x, y) is the pixel of pixel position (x, y).
4. method according to claim 1, is characterized in that, described from multiple sub-image area, and screening obtains the first kind sub-image area comprising subject, comprising:
Calculate the score of each sub-image area in described multiple subimage;
Based on described score, descending sort is carried out to multiple sub-image area, extract at least one sub-image area that sorting position belongs to precalculated position scope;
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the first kind sub-image area comprising subject;
Wherein, the formula that the score of each sub-image area calculated in described multiple subimage utilizes is:
S = ω · f T = Σ k = 1 64 x k y k
Wherein, S is sub-image area score, ω=(x 1, x 2..., x 64) for presetting proper vector, f=(y 1, y 2..., y 64) vector form of the two-value gradient magnitude of sub-image area that generates in sliding process for window, k is dimension.
5. method according to claim 4, is characterized in that, described use Supplement device and nearest neighbor classifier, screens at least one extracted sub-image area, obtains the first kind sub-image area comprising subject, comprising:
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the selection result;
Using the selection result as the first kind sub-image area comprising subject;
Or,
Described use Supplement device and nearest neighbor classifier, screen at least one extracted sub-image area, obtains the first kind sub-image area comprising subject, comprising:
Use Supplement device and nearest neighbor classifier, at least one extracted sub-image area is screened, obtains the selection result;
Judge whether each sub-image area in the selection result is predetermined background image region;
The sub-image area of background image region will not be belonged to as the first kind sub-image area comprising subject in the selection result.
6. a multiple target tracking device, is characterized in that, comprising:
First determination module, for determining that in reference image frame, each destination object is in the first kind estimation range of current target image frame, wherein, described reference image frame and described target image frame are the different images frame in video file, and reference image frame is consecutive previous frame image with target image frame; Wherein, described destination object is at least one in multiple subject included in described reference image frame;
Second determination module, for determining the gradient magnitude figure corresponding to described target image frame;
Sub-image area generation module, for utilizing predetermined window to slide by pixel in described gradient magnitude figure, obtains multiple sub-image areas that described predetermined window generates in sliding process;
Screening module, for from multiple sub-image area, screens and obtains the first kind sub-image area comprising subject;
Analysis module, for performing presumptive area analysis operation to each first kind sub-image area, wherein, described predetermined analysis operation at least comprises:
Images match is carried out with first kind estimation range corresponding to each destination object respectively in current first kind sub-image area, if there is the target first kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target first kind estimation range.
7. device according to claim 6, is characterized in that, the presumptive area analysis operation performed by described analysis module also comprises:
If there is no matching rate is greater than the target first kind estimation range of predetermined threshold value, then when there is at least one two field picture before judging described reference image frame:
From before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame corresponding to this first kind sub-image area, and determine that in current auxiliary view picture frame, each destination object is in the Equations of The Second Kind estimation range of current target image frame, images match is carried out with each Equations of The Second Kind estimation range respectively in this first kind sub-image area, if there is the target Equations of The Second Kind estimation range that matching rate is greater than predetermined threshold value, then by this first kind sub-image area, as the region of destination object in described target image frame corresponding to described target second estimation range, otherwise, continue to perform from before described reference image frame not by least one two field picture of auxiliary view picture frame corresponding to this first kind sub-image area, select near a two field picture of described reference image frame as the current auxiliary view picture frame step corresponding to this first kind sub-image area, until each two field picture before reference image frame is all by as the auxiliary view picture frame corresponding to this first kind sub-image area.
8. device according to claim 6, is characterized in that, described second determination module comprises:
First calculating sub module, for calculating the gradient magnitude of each pixel position in described target image frame;
Obtain submodule, for the gradient magnitude based on calculated each pixel position, obtain the gradient magnitude figure corresponding to described target image frame;
Wherein, the formula that the gradient magnitude calculating each pixel position in described target image frame utilizes is:
f=min(|g x|+|g y|,255)
g x=I(x+1,y)-I(x,y)
g y=I(x,y+1)-I(x,y)
Wherein, f is the gradient magnitude of pixel position (x, y), g xfor the horizontal gradient of image slices vegetarian refreshments position (x, y), g yfor the VG (vertical gradient) of this pixel position (x, y) of image, I (x, y) is the pixel of pixel position (x, y).
9. device according to claim 6, is characterized in that, described screening module, comprising:
Second calculating sub module, for calculating the score of each sub-image area in described multiple subimage;
Extracting submodule, for carrying out descending sort based on described score to multiple sub-image area, extracting at least one sub-image area that sorting position belongs to precalculated position scope;
Classification submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the first kind sub-image area comprising subject;
Wherein, the formula that the score of each sub-image area calculated in described multiple subimage utilizes is:
S = ω · f T = Σ k = 1 64 x k y k
Wherein, S is sub-image area score, ω=(x 1, x 2..., x 64) for presetting proper vector, f=(y 1, y 2..., y 64) vector form of the two-value gradient magnitude of sub-image area that generates in sliding process for window, k is dimension.
10. device according to claim 9, is characterized in that, described sort module, comprising:
First screening submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the selection result;
First obtains submodule, for using the selection result as the first kind sub-image area comprising subject;
Or,
Described sort module, comprising:
Second screening submodule, for using Supplement device and nearest neighbor classifier, screening at least one extracted sub-image area, obtaining the selection result;
Judge submodule, for judging whether each sub-image area in the selection result is predetermined background image region;
Second obtains submodule, for not belonging to the sub-image area of background image region in the selection result as the first kind sub-image area comprising subject.
CN201510676986.9A 2015-10-19 2015-10-19 A kind of multi-object tracking method and device Active CN105261040B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510676986.9A CN105261040B (en) 2015-10-19 2015-10-19 A kind of multi-object tracking method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510676986.9A CN105261040B (en) 2015-10-19 2015-10-19 A kind of multi-object tracking method and device

Publications (2)

Publication Number Publication Date
CN105261040A true CN105261040A (en) 2016-01-20
CN105261040B CN105261040B (en) 2018-01-05

Family

ID=55100711

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510676986.9A Active CN105261040B (en) 2015-10-19 2015-10-19 A kind of multi-object tracking method and device

Country Status (1)

Country Link
CN (1) CN105261040B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106875417A (en) * 2017-01-10 2017-06-20 南京邮电大学 A kind of multi-object tracking method associated across time domain based on high-order figure
CN107452028A (en) * 2017-07-28 2017-12-08 浙江华睿科技有限公司 A kind of method and device for determining target image positional information
CN108303094A (en) * 2018-01-31 2018-07-20 深圳市拓灵者科技有限公司 The Position Fixing Navigation System and its positioning navigation method of array are merged based on multiple vision sensor
CN110222585A (en) * 2019-05-15 2019-09-10 华中科技大学 A kind of motion target tracking method based on cascade detectors

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109934849B (en) * 2019-03-08 2022-05-31 西北工业大学 Online multi-target tracking method based on trajectory metric learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080252725A1 (en) * 2005-09-26 2008-10-16 Koninklijke Philips Electronics, N.V. Method and Device for Tracking a Movement of an Object or of a Person
CN101493889A (en) * 2008-01-23 2009-07-29 华为技术有限公司 Method and apparatus for tracking video object
CN102982340A (en) * 2012-10-31 2013-03-20 中国科学院长春光学精密机械与物理研究所 Target tracking method based on semi-supervised learning and random fern classifier

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080252725A1 (en) * 2005-09-26 2008-10-16 Koninklijke Philips Electronics, N.V. Method and Device for Tracking a Movement of an Object or of a Person
CN101493889A (en) * 2008-01-23 2009-07-29 华为技术有限公司 Method and apparatus for tracking video object
CN102982340A (en) * 2012-10-31 2013-03-20 中国科学院长春光学精密机械与物理研究所 Target tracking method based on semi-supervised learning and random fern classifier

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZDENEK KALAL ET AL: "Tracking-Learning-Detection", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
周鑫 等: "改进后的TLD视频目标跟踪方法", 《中国图象图形学报》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106875417A (en) * 2017-01-10 2017-06-20 南京邮电大学 A kind of multi-object tracking method associated across time domain based on high-order figure
CN106875417B (en) * 2017-01-10 2019-10-08 南京邮电大学 One kind is based on high-order figure across the associated multi-object tracking method of time domain
CN107452028A (en) * 2017-07-28 2017-12-08 浙江华睿科技有限公司 A kind of method and device for determining target image positional information
CN108303094A (en) * 2018-01-31 2018-07-20 深圳市拓灵者科技有限公司 The Position Fixing Navigation System and its positioning navigation method of array are merged based on multiple vision sensor
CN110222585A (en) * 2019-05-15 2019-09-10 华中科技大学 A kind of motion target tracking method based on cascade detectors
CN110222585B (en) * 2019-05-15 2021-07-27 华中科技大学 Moving target tracking method based on cascade detector

Also Published As

Publication number Publication date
CN105261040B (en) 2018-01-05

Similar Documents

Publication Publication Date Title
US9965719B2 (en) Subcategory-aware convolutional neural networks for object detection
CN110287826B (en) Video target detection method based on attention mechanism
CN105261040A (en) Multi-target tracking method and apparatus
CN102542289B (en) Pedestrian volume statistical method based on plurality of Gaussian counting models
CN103886325B (en) Cyclic matrix video tracking method with partition
CN108346159A (en) A kind of visual target tracking method based on tracking-study-detection
CN101835037B (en) Method and system for carrying out reliability classification on motion vector in video
CN103208008A (en) Fast adaptation method for traffic video monitoring target detection based on machine vision
CN104978567B (en) Vehicle checking method based on scene classification
CN104517095B (en) A kind of number of people dividing method based on depth image
CN101807257A (en) Method for identifying information of image tag
CN102521565A (en) Garment identification method and system for low-resolution video
CN102496001A (en) Method of video monitor object automatic detection and system thereof
CN104077596A (en) Landmark-free tracking registering method
CN109242883B (en) Optical remote sensing video target tracking method based on depth SR-KCF filtering
CN104834942A (en) Remote sensing image change detection method and system based on mask classification
CN106778570B (en) A kind of pedestrian detection and tracking in real time
CN105279772A (en) Trackability distinguishing method of infrared sequence image
CN103473786A (en) Gray level image segmentation method based on multi-objective fuzzy clustering
CN112001878A (en) Deep learning ore scale measuring method based on binarization neural network and application system
CN103413149B (en) Method for detecting and identifying static target in complicated background
CN106991686A (en) A kind of level set contour tracing method based on super-pixel optical flow field
CN103886585A (en) Video tracking method based on rank learning
CN115272652A (en) Dense object image detection method based on multiple regression and adaptive focus loss
CN105513080A (en) Infrared image target salience evaluating method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant