CN106485732B - A kind of method for tracking target of video sequence - Google Patents
A kind of method for tracking target of video sequence Download PDFInfo
- Publication number
- CN106485732B CN106485732B CN201610813832.4A CN201610813832A CN106485732B CN 106485732 B CN106485732 B CN 106485732B CN 201610813832 A CN201610813832 A CN 201610813832A CN 106485732 B CN106485732 B CN 106485732B
- Authority
- CN
- China
- Prior art keywords
- target
- space
- time
- context model
- formula
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 35
- 230000008569 process Effects 0.000 claims abstract description 22
- 230000007246 mechanism Effects 0.000 claims abstract description 8
- 230000000903 blocking effect Effects 0.000 claims abstract description 4
- 230000009466 transformation Effects 0.000 claims description 9
- 230000006870 function Effects 0.000 claims description 6
- 238000005286 illumination Methods 0.000 claims description 4
- 238000005562 fading Methods 0.000 claims description 3
- 230000004048 modification Effects 0.000 claims description 3
- 238000012986 modification Methods 0.000 claims description 3
- 238000009825 accumulation Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 230000002401 inhibitory effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
Landscapes
- Image Analysis (AREA)
Abstract
The invention discloses a kind of method for tracking target of video sequence, and target image is normalized first;Then according to the target position of initial frame, the spatio-temporal context information of target is extracted, space-time context model is constructed, the relationship of space-time context model Yu target position confidence map is established, to carry out target following;Then the learning rate for determining whether to update space-time context model using the mechanism of blocking updates space-time context model using the learning rate of update;Scale renewal process is finally established according to the time and space significance in object tracking process between before and after frames.The present invention is set out using space time information, is tracked in conjunction with treatment mechanism and time and space significance is blocked to target, be can effectively improve target following robustness and real-time.
Description
Technical field
The invention belongs to computer visions, digital image processing field, in particular to a kind of target of video sequence with
Track method.
Background technique
Target following is one of hot research topic of field of machine vision, be widely used in human-computer interaction, video with
The fields such as track, navigation, while being also the basis of the follow-up works such as target identification, Activity recognition in video, therefore target following has
Wide application prospect and practical value, receive the highest attention of researcher all over the world.
Currently, since the external factor such as complex background, illumination and target such as rotate, block at the influence of internal factors, make
It obtains tracking process to be typically located under a uncontrolled environment, therefore target following is still a challenging problem.
Summary of the invention
In order to solve above-mentioned background technique propose the technical issues of, the present invention is intended to provide a kind of target of video sequence with
Track method, is set out using space time information, is tracked, can be effectively mentioned to target in conjunction with treatment mechanism and time and space significance is blocked
High target following robustness and real-time.
In order to achieve the above technical purposes, the technical solution of the present invention is as follows:
A kind of method for tracking target of video sequence, comprising the following steps:
(1) target image is normalized, reduces illumination interference;
(2) according to the target position of initial frame, the spatio-temporal context information of target is extracted, constructs space-time context model,
The relationship of space-time context model Yu target position confidence map is established, to carry out target following;
(3) determine whether to update the learning rate of space-time context model using the mechanism of blocking, using update learning rate more
New space-time context model;
(4) scale renewal process is established according to the time and space significance in object tracking process between before and after frames.
Further, detailed process is as follows for step (2):
(a) relationship of spatial context model Yu target position confidence map is established:
In formula (1), and P (x | m (z), o) it is spatial context model, indicate that the space of target and surrounding context feature is closed
System, x are certain point position coordinates in target, and z indicates that the contextual location coordinate of target, o are tracking target, XcFor context spy
Sign defines Xc=m (z)=(I (z), z) | ∈ Ωc(x*), I (z) is the gray value in image at z, Ωc(x*) it is around target
Context area, P (m (z) | o) indicates target local context prior probability;
(b) P (x | m (z), o)=h is enabledsc(x-z) (2)
P (m (z) | o)=I (z) ωσ(z-x*) (3)
In formula (2), (3), hsc(x-z) be relative distance and direction about target and local contextual location z function,
x*For target's center position, ωσ(z-x*) it is Weighted Gauss function, is defined as:
In formula (4), a represents normaliztion constant, and σ represents scale parameter;
Formula (2)-(4) are substituted into formula (1), are obtained
In formula (5), subscript t indicates t frame;
(c) space-time context model is obtained according to spatial context model:
In formula (6),It is spatial context modelFourier transformation,It is space-time context modelFourier transformation, ρtRepresent learning rate;
(d) space-time context model is established according to formula (5)With the relationship of target position confidence map:
In formula (7), F indicates Fourier transformation;
(e) target's centerBy seeking target position confidence map mt+1(x) extreme value obtains:
Further, detailed process is as follows for step (3):
Define the peak sidelobe ratio of t frame:
In formula (9), μsl-tAnd σsl-tIt is confidence map m respectivelyt(x) mean value and standard deviation around peak value in 12 × 12 neighborhoods;
It enables
Δ m=mt-mt-1 (11)
WhenAnd ppsr-t≥pth, Δ m < Mtol, indicate that target is in gradually to walk out and block, model should be carried out more at this time
Newly;WhenAnd ppsr-t≥pth, Δ m > Mtol, indicate that target following is in good condition, cope with model at this time and be updated;WhenOr ppsr-t< pthWhen, indicate that target is in serious shielding or full occlusion state, at this time without model modification;Wherein, pth
For the given threshold of peak sidelobe ratio, MtolFor the given threshold of the variable quantity of target confidence map;
Update the learning rate ρ of space-time context modelt:
Space-time context model is updated according to the learning rate of update:
Further, pthValue be 2.5 × 10-3。
Further, detailed process is as follows for step (4):
In formula (14), η represents fading factor, and n indicates to calculate target scale, s in every n frametIndicate t frame target scale, σt
Indicate t frame scale parameter.
Further, η=0.51, n=5.
By adopting the above technical scheme bring the utility model has the advantages that
(1) present invention using Bayesian frame to the time-space relationship of the target to be tracked and its local context region into
Row modeling, obtains the statistic correlation of target He its peripheral region low-level features, robustness with higher;
(2) present invention copes with complex scene, has judgement to circumstance of occlusion using treatment mechanism is blocked, can
Judged according to object variations process, therefore application blocks treatment mechanism to a certain extent and reduces the accumulation of error, and
There can be certain inhibiting effect to the drift of target, improve the robustness for target following under complex situations;
(3) present invention carries out real-time update to learning rate in object tracking process, can effectively reduce the accumulation of error,
Object module updates and target scale renewal process all refers to learning rate, and it is selectable that the update of learning rate can be such that model carries out
Update it is more accurate, scale it is correct update also have significant impact to the extraction of target signature, the extraction of error characteristic will affect
The tracking of target is caused to drift about or be lost;
(4) present invention is to utilize the connection of target before and after frames to the application of space-time significance measure, is updated to target scale
With great influence, the real-time of object tracking process mesoscale update is improved, the influence of interference information is reduced.
Detailed description of the invention
Fig. 1 is flow chart of the invention.
Specific embodiment
Below with reference to attached drawing, technical solution of the present invention is described in detail.
As shown in Figure 1, a kind of method for tracking target of video sequence, comprising the following steps:
Step 1: target image being normalized, illumination interference is reduced, improves operational efficiency.
Step 2: according to the target position of initial frame, extracting the spatio-temporal context information of target, construct space-time context mould
Type establishes the relationship of space-time context model Yu target position confidence map, to carry out target following.Detailed process is as follows:
A, the relationship of spatial context model Yu target position confidence map is established:
In formula (1), and P (x | m (z), o) it is spatial context model, indicate that the space of target and surrounding context feature is closed
System, x are certain point position coordinates in target, and z indicates that the contextual location coordinate of target, o are tracking target, XcFor context spy
Sign defines Xc=m (z)=(I (z), z) | ∈ Ωc(x*), I (z) is the gray value in image at z, Ωc(x*) it is around target
Context area, P (m (z) | o) indicates target local context prior probability.
B, P (x | m (z), o)=h is enabledsc(x-z) (2)
P (m (z) | o)=I (z) ωσ(z-x*) (3)
In formula (2), (3), hsc(x-z) be relative distance and direction about target and local contextual location z function,
x*For target's center position, ωσ(z-x*) it is Weighted Gauss function, is defined as:
In formula (4), a represents normaliztion constant, and σ represents scale parameter.
Formula (2)-(4) are substituted into formula (1), are obtained
In formula (5), subscript t indicates t frame.
C, space-time context model is obtained according to spatial context model:
In formula (6),It is spatial context modelFourier transformation,It is space-time context modelFourier transformation, ρtRepresent learning rate.
D, space-time context model is established according to formula (5)With the relationship of target position confidence map:
In formula (7), F indicates Fourier transformation.
E, target's centerBy seeking target position confidence map mt+1(x) extreme value obtains:
Step 3: determining whether to update the learning rate of space-time context model using the mechanism of blocking, utilize the learning rate of update
Update space-time context model.Detailed process is as follows:
Define the peak sidelobe ratio of t frame:
In formula (9), μsl-tAnd σsl-tIt is confidence map m respectivelyt(x) mean value and standard deviation around peak value in 12 × 12 neighborhoods;
It enables
Δ m=mt-mt-1 (11)
As shown in table 1, whenAnd ppsr-t≥pth, Δ m < Mtol, indicate that target is in gradually to walk out and block, cope at this time
Model is updated;WhenAnd ppsr-t≥pth, Δ m > Mtol, indicate that target following is in good condition, model should be carried out at this time
It updates;WhenOr ppsr-t< pthWhen, indicate that target is in serious shielding or full occlusion state, at this time without model modification;
Wherein, pthFor the given threshold of peak sidelobe ratio, MtolFor the given threshold of the variable quantity of target confidence map.
Update the learning rate ρ of space-time context modelt:
Space-time context model is updated according to the learning rate of update:
Table 1
Step 4: scale renewal process, purpose are established according to the time and space significance in object tracking process between before and after frames
It is to improve the real-time that object tracking process mesoscale updates, reduces the influence of interference information.Detailed process is as follows:
In formula (14), η represents fading factor, ηiIt is reduced with the increase of i, indicates influence of the historical frames to present frame at any time
Between change, in the present embodiment, η=0.51;N indicates calculating target scale in every n frame, in the present embodiment, n=5, because
In 5 frames, target scale variation will not be larger, can reduce calculation amount and improve the real-time of the patent;, stIndicate t frame target
Scale, σtIndicate t frame scale parameter.
The above examples only illustrate the technical idea of the present invention, and this does not limit the scope of protection of the present invention, all
According to the technical idea provided by the invention, any changes made on the basis of the technical scheme each falls within the scope of the present invention
Within.
Claims (4)
1. a kind of method for tracking target of video sequence, which comprises the following steps:
(1) target image is normalized, reduces illumination interference;
(2) according to the target position of initial frame, the spatio-temporal context information of target is extracted, constructs space-time context model, is established
The relationship of space-time context model and target position confidence map, to carry out target following;Detailed process is as follows for the step:
(a) relationship of spatial context model Yu target position confidence map is established:
In formula (1), and P (x | m (z), o) it is spatial context model, indicate the spatial relationship of target and surrounding context feature, x
It is certain point position coordinates in target, z indicates that the contextual location coordinate of target, o are tracking target, XcIt is fixed for contextual feature
Adopted Xc=m (z)=(I (z), z) | ∈ Ωc(x*), I (z) is the gray value in image at z, Ωc(x*) it is upper around target
Context area, and P (m (z) | o) indicate target local context prior probability;
(b) P (x | m (z), o)=h is enabledsc(x-z) (2)
P (m (z) | o)=I (z) ωσ(z-x*) (3)
In formula (2), (3), hscIt (x-z) is about target and the relative distance of local contextual location z and the function in direction, x*For
Target's center position, ωσ(z-x*) it is Weighted Gauss function, is defined as:
In formula (4), a represents normaliztion constant, and σ represents scale parameter;
Formula (2)-(4) are substituted into formula (1), are obtained
In formula (5), subscript t indicates t frame;
(c) space-time context model is obtained according to spatial context model:
In formula (6),It is spatial context modelFourier transformation,It is space-time context modelFourier transformation, ρtRepresent learning rate;
(d) space-time context model is established according to formula (5)With the relationship of target position confidence map:
In formula (7), F indicates Fourier transformation;
(e) target's centerBy seeking target position confidence map mt+1(x) extreme value obtains:
(3) learning rate for determining whether to update space-time context model using the mechanism of blocking, when being updated using the learning rate of update
Empty context model;Detailed process is as follows for the step:
Define the peak sidelobe ratio of t frame:
In formula (9), μsl-tAnd σsl-tIt is confidence map m respectivelyt(x) mean value and standard deviation around peak value in 12 × 12 neighborhoods;
It enables
Δ m=mt-mt-1(11)
WhenAnd ppsr-t≥pth, Δ m < Mtol, indicate that target is in gradually to walk out and block, cope with model at this time and be updated;
WhenAnd ppsr-t≥pth, Δ m > Mtol, indicate that target following is in good condition, cope with model at this time and be updated;WhenOr
ppsr-t< pthWhen, indicate that target is in serious shielding or full occlusion state, at this time without model modification;Wherein, pthFor peak value
The given threshold of secondary lobe ratio, MtolFor the given threshold of the variable quantity of target confidence map;
Update the learning rate ρ of space-time context modelt:
Space-time context model is updated according to the learning rate of update:
(4) scale renewal process is established according to the time and space significance in object tracking process between before and after frames.
2. a kind of method for tracking target of video sequence according to claim 1, it is characterised in that: pthValue be 2.5 ×
10-3。
3. a kind of method for tracking target of video sequence according to claim 1, which is characterized in that the specific mistake of step (4)
Journey is as follows:
In formula (14), η represents fading factor, and n indicates to calculate target scale, s in every n frametIndicate t frame target scale, σtIt indicates
T frame scale parameter.
4. a kind of method for tracking target of video sequence according to claim 3, it is characterised in that: η=0.51, n=5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610813832.4A CN106485732B (en) | 2016-09-09 | 2016-09-09 | A kind of method for tracking target of video sequence |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610813832.4A CN106485732B (en) | 2016-09-09 | 2016-09-09 | A kind of method for tracking target of video sequence |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106485732A CN106485732A (en) | 2017-03-08 |
CN106485732B true CN106485732B (en) | 2019-04-16 |
Family
ID=58273728
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610813832.4A Active CN106485732B (en) | 2016-09-09 | 2016-09-09 | A kind of method for tracking target of video sequence |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106485732B (en) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107093189A (en) * | 2017-04-18 | 2017-08-25 | 山东大学 | Method for tracking target and system based on adaptive color feature and space-time context |
CN107240120B (en) * | 2017-04-18 | 2019-12-17 | 上海体育学院 | Method and device for tracking moving target in video |
CN107146238B (en) * | 2017-04-24 | 2019-10-11 | 西安电子科技大学 | Based on the preferred motion target tracking method of characteristic block |
CN107045726B (en) * | 2017-04-27 | 2019-12-24 | 南京航空航天大学 | Target tracking method based on space-time context model |
CN107146240B (en) * | 2017-05-05 | 2019-08-13 | 西北工业大学 | The video target tracking method of taking photo by plane detected based on correlation filtering and conspicuousness |
CN107346548A (en) * | 2017-07-06 | 2017-11-14 | 电子科技大学 | A kind of tracking for electric transmission line isolator |
CN107424175B (en) * | 2017-07-20 | 2020-09-08 | 西安电子科技大学 | Target tracking method combined with space-time context information |
CN107452015B (en) * | 2017-07-28 | 2020-09-25 | 南京工业职业技术学院 | Target tracking system with re-detection mechanism |
CN108022254B (en) * | 2017-11-09 | 2022-02-15 | 华南理工大学 | Feature point assistance-based space-time context target tracking method |
CN110084233B (en) * | 2018-01-25 | 2021-02-09 | 华中科技大学 | Method and system for quickly capturing target in production line video sequence |
CN108320298B (en) * | 2018-04-28 | 2022-01-28 | 亮风台(北京)信息科技有限公司 | Visual target tracking method and equipment |
CN109285179B (en) * | 2018-07-26 | 2021-05-14 | 昆明理工大学 | Moving target tracking method based on multi-feature fusion |
CN109325966B (en) * | 2018-09-05 | 2022-06-03 | 华侨大学 | Method for carrying out visual tracking through space-time context |
CN110111370B (en) * | 2019-05-15 | 2023-05-30 | 重庆大学 | Visual object tracking method based on TLD and depth multi-scale space-time features |
CN112733770A (en) * | 2021-01-18 | 2021-04-30 | 全程(上海)智能科技有限公司 | Regional intrusion monitoring method and device |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105117720A (en) * | 2015-09-29 | 2015-12-02 | 江南大学 | Object scale self-adaption tracking method based on spatial-temporal model |
CN105654518A (en) * | 2016-03-23 | 2016-06-08 | 上海博康智能信息技术有限公司 | Trace template self-adaption method based on variance estimation |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011517228A (en) * | 2008-04-11 | 2011-05-26 | トムソン ライセンシング | System and method for improving visibility of objects in digital images |
-
2016
- 2016-09-09 CN CN201610813832.4A patent/CN106485732B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105117720A (en) * | 2015-09-29 | 2015-12-02 | 江南大学 | Object scale self-adaption tracking method based on spatial-temporal model |
CN105654518A (en) * | 2016-03-23 | 2016-06-08 | 上海博康智能信息技术有限公司 | Trace template self-adaption method based on variance estimation |
Non-Patent Citations (3)
Title |
---|
Fast visual tracking via dense spatio-temporal context learning;Zhang K H等;《13th European Conference on Computer Vision》;20141231;全文 |
基于置信图特性的改进时空上下文目标跟踪;张雷等;《计算机工程》;20160831;第42卷(第8期);第277-281页 |
融合时空上下文的超像素跟踪算法研究;杜安安;《中国优秀硕士学位论文全文数据库》;20150815;第I138-1393页 |
Also Published As
Publication number | Publication date |
---|---|
CN106485732A (en) | 2017-03-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106485732B (en) | A kind of method for tracking target of video sequence | |
CN101840507B (en) | Target tracking method based on character feature invariant and graph theory clustering | |
CN106875431B (en) | Image tracking method with movement prediction and augmented reality implementation method | |
CN107301657B (en) | A kind of video target tracking method considering target movable information | |
CN111860352B (en) | Multi-lens vehicle track full tracking system and method | |
CN106952294B (en) | A kind of video tracing method based on RGB-D data | |
CN103971386A (en) | Method for foreground detection in dynamic background scenario | |
CN105374049B (en) | Multi-corner point tracking method and device based on sparse optical flow method | |
Cho et al. | Real-time people counting system for customer movement analysis | |
CN105261034A (en) | Method and device for calculating traffic flow on highway | |
CN103942542A (en) | Human eye tracking method and device | |
CN103391430B (en) | DSP (digital signal processor) based relevant tracking method and special device | |
US20180340788A1 (en) | A navigation apparatus and associated methods | |
Wang et al. | Improving RGB-D SLAM accuracy in dynamic environments based on semantic and geometric constraints | |
CN106920253A (en) | It is a kind of based on the multi-object tracking method for blocking layering | |
CN103281476A (en) | Television image moving target-based automatic tracking method | |
CN108320298A (en) | A kind of visual target tracking method and equipment | |
CN102855637A (en) | Covariance tracking method for fusing bilateral filtering in video images | |
CN105187801B (en) | System and method for generating abstract video | |
CN104361607A (en) | Real-time multidirectional pedestrian counting and tracking method | |
CN114037087B (en) | Model training method and device, depth prediction method and device, equipment and medium | |
Haggui et al. | Centroid human tracking via oriented detection in overhead fisheye sequences | |
CN103578121A (en) | Motion detection method based on shared Gaussian model in disturbed motion environment | |
CN116883460A (en) | Visual perception positioning method and device, electronic equipment and storage medium | |
CN114518106B (en) | Method, system, medium and equipment for detecting update of vertical elements of high-precision map |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |