WO2015163830A1 - Target localization and size estimation via multiple model learning in visual tracking - Google Patents

Target localization and size estimation via multiple model learning in visual tracking Download PDF

Info

Publication number
WO2015163830A1
WO2015163830A1 PCT/TR2014/000117 TR2014000117W WO2015163830A1 WO 2015163830 A1 WO2015163830 A1 WO 2015163830A1 TR 2014000117 W TR2014000117 W TR 2014000117W WO 2015163830 A1 WO2015163830 A1 WO 2015163830A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
visual tracking
real
filter group
model
Prior art date
Application number
PCT/TR2014/000117
Other languages
French (fr)
Inventor
Erhan GUNDOGDU
Emre TUNALI
Gökhan TANISIK
Sinan OZ
Original Assignee
Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi filed Critical Aselsan Elektronik Sanayi Ve Ticaret Anonim Sirketi
Priority to PCT/TR2014/000117 priority Critical patent/WO2015163830A1/en
Publication of WO2015163830A1 publication Critical patent/WO2015163830A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/251Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/215Motion-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20076Probabilistic image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Abstract

Visual target tracking has many challenges such as robustness to occlusion, noise, drifts, stabilization etc. Although various algorithms are proposed as a remedy for these problems, the solutions should be narrowed to algorithms with low computational cost when real time systems are in consideration. In this manner, the family of tracking methods based on correlation filters is a prominent option since many of the algorithms in this family are efficient and simple to implement. In order to achieve efficient and robust tracking system, the present invention relates a correlation based tracker with a target localization and size estimation method with a feedback mechanism. In this sense, target model is dynamically learnt and extracted in the tracker window encapsulating the actual target and this model is used both for target localization and size estimation in addition to track window correction which introduces robustness to improper initializations. Moreover, a multiple model visual tracking methodology is also presented in order to adapt to changes in target model with different rates that are either caused by changes in the target or its surroundings. The overall system can be used as a real-time visual tracking system with adaptive learning mechanism and provides minimum sized target bounding box as output. Furthermore, the method presented in this invention is capable of target model extraction which can be considered as a preprocessing step of a shape based object classification algorithm.

Description

TARGET LOCALIZATION AND SIZE ESTIMATION VIA MULTIPLE MODEL LEARNING IN VISUAL TRACKING
Field of the Invention
The present invention relates to a target localization and size estimation method for visual tracking purposes using an intelligent system including a dynamic and adaptive localization algorithm, which is robust to improper target initializations, as well as a multiple model structure with a model selection algorithm for tracking.
Background of the Invention
Visual target tracking is a well known topic in computer vision and machine learning disciplines. As well as many machine vision problems, visual target tracking has also trade-offs such as computational complexity and robustness to various problems including occlusion, noise, drifts etc. Although diverse set of algorithms exists as solutions of mentioned problems, they may not be appropriate for real-time systems since they tend to be computationally costly. To achieve target tracking in real-time systems with less computational burden, using trackers that are based on correlation filters can be considered as an option.
In correlation filter based trackers; there is plurality of templates used to find the location of an object by searching a predefined area in a video frame. According to a cost function, one can find the location of a target using the predefined space and the prior information provided at the beginning of the tracking.
In most scenarios, trackers based on correlation filters assume fixed object size and limits target search to a predefined window. Moreover, most of them do not use the appropriate computer vision tools to extract the semantic information behind the data taken from the sensors. Furthermore, many algorithms model target objects and surroundings which may differ in time. These model based algorithms may have limits on adaptation to changes in the scene which results in conflicts with the current model parameters, hence the performance of the algorithms are confined in a limited range.
This disclosure provides solutions for previously mentioned problems by using a biologically inspired framework which is capable of target model extraction in awareness of changes in the scene. Actually, the proposed disclosure interprets the target model and decides on its learning rates for both localization and size estimation which yields better track maintenance. In addition, a multiple model visual tracking method is also proposed to extend limitations of adaptation on changes in the target and its background.
United States patent document US8520956B2 discloses an efficient method for visual tracking based on correlation filters. In this approach, the prior information is plurality of images used for learning a correlation filter which is optimum for a defined cost function. They have basically three different options to find the optimum correlation filter as ASEF (Average of Synthetic Exact Filters), MOSSE (Minimizing Output Sum of Squared Error) Filter and a cost function minimizing filter. In all of the methods of this patent document, the filter is assumed to have a fixed size. Although the basic idea is novel and works well in many scenarios, the assumption of the fixed size object does not hold all the time. As the object starts to magnify in the region of interest, the tracker may not compensate for the enlarged object inside the tracker window or vice versa. Moreover, the method presented in US8520956B2 selects target to be nearly the whole image patch in the window. Hence, the target together with its background is matched in consecutive frames. Since, the shape (boundaries) of the target is not extracted in this method; neither background suppression is utilized in consideration of target size, nor any mechanism is included to centralize the tracked target. However; in the cases of erroneous track initialization or mismatching between consecutive frames, the target may shift from the center of track window which is not desired and may be result in premature track loss if it is not corrected. Centralization of the target might be achieved by obtaining the silhouette of the object. For this purpose, the saliency map calculation is exploited in the target window and the target model is learnt by using saliency maps in time. Then, the most salient object in the updated saliency map is defined as the target. Selection of most salient region also means more distinctive target which also increases probability of longer track maintenance.
United States patent document US2008304740 discloses methods for detecting a salient object in an input image are described. A set of local, regional, and global representations are exploited such as including multi-scale contrast, center- surround 20 histogram, and color spatial distribution to find the salient region in the image. Conditional random field learning methodology is used to combine relationships between the features mentioned above. The learned conditional random field helps to locate the salient object in the image. Although image segmentation is proposed to find the salient region in the scene, it may not be appropriate for real time applications.
United States patent document US20130101210 discloses a method for determining a region to crop from a digital image using a saliency map. In their method, the image or the region of interest is processed to obtain the saliency map. The pixels in the image are assigned with saliency values within a number range. Their method further contains analyzing the saliency map to find a potential cropping rectangle. In order to find a rectangle to crop, candidate rectangles are extracted. Every candidate rectangle has a score corresponding to the sum of saliency values inside the rectangle. The rectangle with the highest saliency score is chosen to be cropped. However, instead of finding a potential cropping rectangle, the method presented here exploits the connected component labeling of the binarized saliency maps as well as the saliency values of each connected component to find a dominant connected component to be used in the target boundary calculation. Another important contribution of the proposed disclosure is temporal refinement of saliency map with an adaptive learning rate. More clearly, decision does not depend on a single frame unlike the method in US20130101210; on the contrary saliency map is generated from weighted average of previous frames in which weights depend on adaptively changing object model.
Change of environmental conditions such as clutter in background, noise, deformations of the object is another important problem which should be compensated. In order to solve these problems, adaptive algorithms are proposed such as the method in US8520956B2. There are also methods which uses models to track objects such as the method in US8477998B1. In their method, an input image is taken with a target selection at the first frame. Plurality of images chosen based on the selected target region is processed to construct a generalized target model with different poses. In each frame, the model is used to find a recognition score in many candidate regions. Moreover, the model is adaptively trained using the content of the most likely object region. The adaptive model of the method in US8477998B1 helps to reduce the drifts.
Although adaptive tracking methods can solve drifting problems up to a scale, infinitesimal drifting of targets or sensors (camera) may not be compensated using standard adaptive algorithms. A real world example of such a problem is very small drifting of the camera such as one pixel drift in 100 frames. As a matter of fact, the slightly shifted target is considered as the original target and learnt in adaptive tracking methods. This can cause considerable drifts as the actual target model starts to deform. In order to handle both abrupt and indiscernible appearance changes, multiple models for different adaptation rates can be used interchangeably. Therefore, in this disclosure a multiple model visual tracking system is also presented with a model selection mechanism including two filter groups with different learning parameters. References
[1] Nobuyoki Otsu, "A threshold selection method from gray-level histograms", IEEE Trans. Sys. Man., Cyber. 9, 1979 (1), 62-66.
doi: 10.1109/TMSC.1979.4310076
[2] Y. Wei, F. Wen, W. Zhu and J.Sun, "Geodesic Saliency Using Background Priors", IEEE, ICCV, 2012.
Summary of the Invention
This invention proposes a method of generation of minimum sized bounding box for the target, which means target location and size estimation, in a visual tracking system. Moreover, a multiple model learning methodology is also presented to improve adaptation limits of the system. Since the proposed methodologies require learning of the target model, they can be adapted to any model based tracking algorithm.
In the presented visual tracking framework, track initialization (101) is given by a user or system by inputting an image patch that includes target to be tracked. Since the target bounding box generation requires the condition that the track window should contain the target completely, the drifts, caused by the tracker, should be prevented in order not to dissatisfy this condition. Therefore, a tracker robust to drifts should be preferred to provide the whole target in the track window. In order to prevent drifts, the track is maintained by multiple model visual target tracking procedure is designed to adopt to different rates of temporal variations of the target (102) in which the model updates are achieved by different learning rates to compensate for abrupt and indiscernible changes in the target appearance and appropriate model is interchangeably selected.
The target bounding box generation and feedback decision (103) procedure is utilized by using the updated saliency map. In each frame a saliency map is calculated in the region of interest, which is determined by the tracker, is referred as current saliency map. Since the aim is to use temporal information of the saliency map, another saliency map is defined as updated saliency map and calculated using the current saliency map with an adaptive learning rate. One should note that using a constant learning rate may cause mislearning of target silhouette in cases of full/partial occlusions or with noisy data. Hence, an adaptive learning rate is utilized. The details of the calculation of the learning rate will be explained in the next section. After updating the saliency map with the calculated learning rate, the target selection is performed. At the output of the target selection process, the silhouette of the target, which is used for size estimation, as well as the location of the target is determined. Using the current saliency map and the updated saliency map a feedback is applied to the system. In the consecutive frames the procedure explained above is applied and target bounding box is obtained from updated saliency map for each frame.
Brief Description of Contents
A system and method realized to fulfill the objective of the present invention is illustrated in the accompanying figures, in which:
Figure 1 shows the flowchart of the overall system.
Figure 2 shows the multiple model visual tracking algorithm steps.
Figure 3 shows the target bounding box generation and feedback decision steps.
Figure 4 shows the target bounding box with the track window.
Figure 5 shows the current and updated saliency maps with their binarized images at the beginning of the tracking.
Figure 6 shows the current and updated saliency maps with their binarized images after 380 frames.
Figure 7 shows the current and updated saliency maps with their binarized images after 510 frames. Detailed Description of the Invention
Hereinafter, the present invention will be described in more detail with reference to attached drawings. The following embodiments are presented only for the purpose of describing the present invention, and are not intended to limit the scope of the present invention.
The real-time visual tracking method for target localization and size estimation, based on correlation filter with multiple model, the method comprises the following steps:
- Given track initialization (101),
- Multiple model visual target tracking for the purpose of responding to abrupt and indiscernible changes of the target appearance (102),
- Target bounding box generation and feedback decision using the updated saliency map (103).
Multiple Model Target Tracking
Visual tracking is the method for tracking a selected region of interest throughout the video frames. The output of the visual tracker is a region of interest in which the saliency map is to be calculated. In each frame, the output of the target localization system is used to feed the visual tracker with the target size and location if necessary conditions are satisfied. The details of the necessary conditions are explained in the part namely learning rate calculation and temporal refinement of the saliency map.
In the present invention, a multiple model visual tracker is designed to respond successfully to abrupt and indiscernible changes of the target appearance. To achieve this goal, two filter groups are constructed. The first filter group, is
Figure imgf000008_0001
intended to solve the drifts, which are undetectably small for the second filter group, The second filter group is designed to adapt to high variations in target appearance. The flow of the algorithm is shown in Figure 2. The presented tracking algorithm takes a region of interest in the new frame as input. Then, the necessary and appropriate preprocessing procedure for incoming frame, such as low-pass filtering, contrast stretching, spatial windowing etc., (201) is performed according to the tracking method.
Figure imgf000009_0001
In the above equations, the F represents the filter groups with subscript denoting the filter group ID. t and t+1 denote previous and the next filters, respectively.
Figure imgf000009_0002
represents the currently calculated value of the filter using the current frame. The first filter group takes the preprocessed image part as input and the tracker algorithm is run according to the first filter group by correlation matching using the first group of the multiple model visual tracking (202). The output of the filter generates a quality measure for the resulting target location. If the quality of the response of the first filter group to the target in the current frame is higher than a predefined threshold (203), the target location is updated according to the first filter group (204) and the first filter group is updated with a low learning rate and the second filter group is updated with a high learning rate (207) as in Eqn.l and Eqn.3. Yet, if the quality measure of the first filter is less than the predefined threshold (203), the second filter group generates a target location output by correlation matching using the second filter group using multiple model visual tracking (205) and a quality measure. If the quality of the response of the second filter group in the current frame is higher than a predefined threshold (206), the target location is updated with respect to the second filter group (208) and both of the filter groups are updated with the learning rate of a high learning rate (210, 211:) as in Eqn.2 and Eqn.3. If the quality of the second filter response is not high enough compared to the predefined threshold (206), the system detects occlusion (209). When the new frame is taken with the target information, the same procedure is applied as it is shown in Figure 2. The thresholds used to decide on the quality of the response of the filter groups are design parameters and can be adjusted according to the specific system requirements.
To summarize the philosophy behind this methodology, the filter group which is designed for adapting to smaller changes has the priority to be used as the actual response since abrupt changes are not expected normally. When the rapid changes starts to occur, the filter group designed to adapt to abrupt changes becomes active if the filter group for smaller changes cannot satisfy the quality requirements. In order to maintain the sustainability of the two filter groups, the filter group for smaller changes starts to be updated with the update parameters of the other filter group for abrupt changes. Hence, interaction between the multiple filter groups is used to tolerate the errors of each other in different conditions by using their corresponding learning rates when one of the filter groups starts to give low quality tracking results.
The procedure of the multiple visual tracking can be summarized as follows:
- A necessary and appropriate preprocessing procedure for incoming frame (201),
- Correlation matching using the first filter group of the multiple model visual tracking (202),
- Querying the quality of the response of the first filter group to the target in the current frame (203),
- Updating the target location according to the first filter group (204) if the querying result of the response of the first filter group is greater than a selected threshold (203),
- Updating the first filter group with a low learning rate and updating the second filter group with a high learning rate (207), - Correlation matching using the second filter group of the multiple model visual tracking (205) , if the result of the querying the quality of the first filter group is less than the selected threshold,
Querying the quality of the response of the second filter group to the target in the current frame (206),
- Updating the target location according to the second filter group if the result of the querying the quality of the first filter group is greater than a predefined threshold (208),
- Updating the first filter group with a high learning rate (210).
- Updating the second filter group with a high learning rate (211),
Detecting occlusion (209), if the result of querying the second filter group is less than the predefined threshold (206).
Target Bounding Box Generation
Target bounding box generation actually means target location and size estimation and divided into three substages: saliency map generation, learning rate calculation and temporal refinement of the saliency map, target selection.
Target bounding box generation method provides the ability of detecting the scale changes of the target through the video frames, adapting to the scale changes and updating the visual target model more appropriately than the updating the tracking model of a visual tracking method without the ability of detecting the scale changes. Moreover, this phase also provides an adaptive learning rate selection algorithm, which is designed to prevent mislearning of the target model in the cases of clutter or occlusion. By avoiding the mislearning of the target through the frames, a target model which is appropriate to be used in order to boost redetection of target after the target is lost due to occlusion, clutter or noise, is constructed. This extracted target model is also appropriate to be used in a shape based classifier. Saliency Map Generation
The saliency map of the region of interest, which is selected by the algorithm above, can be extracted by a saliency score calculation method. The recently proposed fast saliency extraction method by Y. Wei et.al., "Geodesic saliency using background priors", 2012, ECCV, in which the saliency problem is tackled from different perspective by focusing on background more than the object, can be an used as a saliency calculation tool. This method is capable of extracting a saliency map within few milliseconds even in embedded systems; however, it has two basic assumptions for the input image that should be guaranteed, namely boundary and connectivity. The boundary assumption is reflection of a basic tendency that photographer/cameraman do not crop salient objects among the frames. Therefore, the image boundary is usually background. The connectivity assumption comes from the fact that background regions are generally tend to be large and homogenous, i.e. sky, grass. In other words, most image patches can be easily connected to each other piecewise. In our case, these two assumptions are fulfilled during tracking by simply selecting initial target window including target, roughly centralized, and its immediate surroundings. Satisfying these two conditions, the salient regions are assumed to be the patches that are extracted by downscaling or by any super pixel extraction algorithm with high geodesic distance from the boundaries of the image that is assumed to correspond to piecewise-connected background regions. The geodesic saliency of a patch p is the accumulated edge weights along the shortest path from p to virtual background node b in an undirected weighted graph as in Eqn.4.
Figure imgf000012_0001
Figure imgf000012_0002
Note that, since patches close to the center of the image requires a longer path in order to reach the background, accumulations of weights tend to be larger in the center patches. Therefore, this method also favors the center image regions as more salient which is reasonable since salient regions tend to occur around the center of image.
Learning Rate Calculation and Temporal Refinement of the Saliency Map
Since target tracking is a continual process, it includes temporal information which can be used for target localization and size estimation. In the proposed methodology, current saliency map calculation is performed for each frame (301) and then considering previously generated saliency maps the updated saliency map is continuously learnt and given as input to the system for the next frame (302). The important thing is each saliency map may not represent the target with the same quality. Therefore, samples of higher quality should be weighted more in the updated saliency map. Actually, this quality measures the temporal consistency of the target. In this sense, adaptive learning rate calculation becomes very important especially for two reasons: First, due to noise or any imperfection of the sensor data saliency map may deviate from frame to frame. However, learning saliency map from deviated version will extract the common structure as target model; hence it can compensate the noise. Second, when the target is fully or partially occluded, the abrupt change in the saliency map is detected and learning rate is adjusted in a way to prevent target model that exists before occlusion.
For learning rate calculation, two parameters, saliency ratio and correlation score, are used which are calculated by using both updated and current saliency maps. Firstly, current saliency map is calculated and binarized for each image. Then the first parameter for learning rate, saliency ratio
Figure imgf000013_0002
is calculated as in Eqn.5 for current saliency map (303), where dominant components represents the saliency values greater than binarization threshold and
Figure imgf000013_0001
is the current saliency map. To be clearer, this metric is designed to measure distinctiveness of the target. In the cases where only the target has high saliency values, the saliency ratio will be 1 which means the target in scene is very distinctive. Hence, for extraction of target model this frame is very reliable and should be learnt with high learning rate.
Figure imgf000014_0001
The second metric is correlation score, DNCC (t), which is a very strong cue for detection of abrupt changes from the updated target model. To achieve this goal, normalized cross correlation between the target models, selected from updated saliency map using the target selection procedure (307) that will be described in detail in target selection part, and current saliency map is taken as in Eqn.6 (304).
Figure imgf000014_0002
Using the saliency ratio and the correlation score, the learning rate is calculated at each frame as in the Eqn.7 (305) and represented with symbol λ (t) at time t. Note that ranges of both metrics extends from 0 to 1 and if both are 1 the current target is overwritten to the updated target which is not desired since it clears out all temporal information. In order to prevent this, a maximum learning rate is restricted to alpha, a. Moreover, in order to prevent mislearning of target, a penalization constant, β, is used whenever target model and the best possible match has resemblance below the feedback threshold, which simply
Figure imgf000014_0005
means system updates target model whenever the measurement is considered to be secure.
Figure imgf000014_0003
After calculation of learning rate, the saliency map is updated,
Figure imgf000014_0004
according to the calculated learning rate at each frame as in the Eqn.8 (306). The natural response of such a learning framework is to learn more if the current salient component is worth considering. Moreover, the components which are consistent with the learnt saliency map are also considered to be learnt more.
Figure imgf000015_0001
Since correlation metric shows resemblance between target and current saliency map, it is also used to answer to the question 'when the feedback should be given to the visual tracking system as the actual location of the object'. The formulation in Eqn.9 is used for querying tracking feedback (308).
Figure imgf000015_0002
In Eqn.9, the IsFeedBack variable is a binary variable controlling the decision of giving feedback or not (Give Feedback if 1 else 0). If the correlation score is high enough, then the current saliency map is consistent with the previous behavior of the region of interest. This results in giving feedback to the visual tracking system since it is the signature of a secure measurement. Moreover, exploiting feedback mechanism results in compensation for false target initialization via its learning mechanism where the targets, which are not well-localized in initialization, are centralized via the feedback mechanism. On the other hand, the system should not accept location information coming from the current saliency map, since it indicates existence of obstacles or occlusion at the current frame.
One should note that all these saliency calculations are utilized in the window which is the output of the visual tracking algorithm, referred as the region of interest in this context, and illustrated as outer bounding box in Figure 4. After applying the presented target bounding box generation method, the inner bounding, which is capable of identifying the target location and size, is generated. This results in locating the object in the region of interest by generating a minimum size bounding box including the target instead of the region of interest given by the real time visual tracking method and the target bounding box and region of interest illustrations are given in Figure 4.
A simple illustration is given in Figure 5, Figure 6 Error! Reference source not found.and Figure 7, the illustrations for the 68th frame, 437th frame and 567th frames, respectively. In these figures, the top-left window is dedicated for the original gray-scale image in which track bounding box, larger rectangle determining region of interest, is visualized together with target bounding box, small rectangle revealing target location and size. The top-middle figure shows the updated saliency map and the top-right figure illustrates the binarization result of the updated saliency map. The bottom-middle figure shows the current saliency map and the bottom-right figure shows the binarization of the current saliency map. The bottom-left figure shows the normalized cross correlation result of the updated and current saliency maps. When the target is partly or fully occluded as in the case shown in Figure 6, the updated and current saliency maps, top-middle and bottom-middle, respectively, in Figure 6 would be dissimilar. This is the case, when the temporal consistency is spoilt. The dissimilarity would yield low cross correlation between target model and current saliency map which prevents giving the location feedback. Moreover, the learning rate is decreased with the penalty term β as in Eqn.7 to prevent target model. After 510 frames (Figure 7), the occlusions coming from the trees and the moving person disappear and the system firstly starts to increase the learning rate due to effect of the saliency ratio metric. Then, this yields increase in correlation score and when the upper condition is satisfied in Eqn.7 system both omits the penalization term in Eqn.7, β, and starts to give the location of the target as feedback to the tracker. By this way, the algorithm presented here is not affected by occlusion and clutter more than the classical feedback mechanism for the visual tracker. Target Selection
Target selection procedure (307) is achieved in two steps: binarization and maximization of the regularization energy. Then, the target bounding box is outputted as the bounding box of the selected connected component.
Although minimum computational cost is desired in each step, using static threshold or suboptimal methods for binarization may be problematic. Thus, Otsu's method is used with slight refinement. The method of N. Otsu, "A threshold selection method from gray-level histogram", 1979, can be either defined as an exhaustive search for the threshold that either minimizes the within- class variance or maximizes between-class variance. The between-class variance is often calculated as given in Eqn.10:
Figure imgf000017_0002
where is referred as class probabilities and are class means. After
Figure imgf000017_0006
Figure imgf000017_0005
some manipulations Eqn.10 can be written as in Eqn.ll.
Figure imgf000017_0003
where μ is the mean value of the histogram. Since the purpose is to calculate the optimal threshold value T that maximizes , the problem can be solved by either
Figure imgf000017_0001
inserting Eqn.10 or Eqn.ll into the Eqn.12.
Figure imgf000017_0004
Note that using Eqn.10 and Eqn.ll directly results in Eqn.13 and Eqn.14 respectively;
Figure imgf000018_0001
where the number of pixels with gray level i
Figure imgf000018_0003
is given with As it
Figure imgf000018_0004
can be seen using Eqn.11 becomes slightly advantageous since constant μ term is dropped out. This slight modification results in one less multiplication in Eqn.14 than Eqn.13 which results in L less multiplication in exhaustive search used in Otsu's methodology.
After thresholding the saliency map, the connected component maximizing the regulanzation energy given by Eqn.15, i.e. the most salient region with minimum distance to the center, is selected as the target.
Figure imgf000018_0002
where is the vectorized form obtained by raster scanning the 2D label matrix with values 1 and 0 as foreground and background respectively, S is the saliency map vectorized similarly and are the centers of each connected
Figure imgf000018_0005
component and the initial window respectively.
The target bounding box generation and feedback decision procedure can be summarized as:
- Calculating current saliency map (301), - Calculating saliency ratio for the current saliency map (303),
Calculating the correlation score using normalized cross correlation between the target model selected (304) from the updated saliency map using target selection procedure (307) and current saliency map,
Calculating the learning rate using the saliency ratio and the correlation score (305),
- Updating the saliency map according to the calculated learning rate (306), to obtain the updated saliency map as input to the system for the next frame (302),
- Selecting target according to the updated saliency map (307),
- Querying tracking feedback (308).

Claims

1. The real-time visual tracking method for target localization and size estimation, based on correlation filter with multiple model, the method comprises the following steps:
- Given track initialization (101),
- Multiple model visual target tracking for the purpose of responding to abrupt and indiscernible changes of the target appearance (102),
- Target bounding box generation and feedback decision using the updated saliency map (103).
2. The real-time visual tracking method according to claim 1, wherein the multiple model visual tracking step further comprises the sub-steps of:
- A necessary and appropriate preprocessing procedure for incoming frame (201),
- Correlation matching using the first filter group of the multiple model visual tracking (202),
- Querying the quality of the response of the first filter group to the target in the current frame (203),
- Updating the target location according to the first filter group (204) if the querying result of the response of the first filter group is greater than a selected threshold (203),
- Updating the first filter group with a low learning rate and updating the second filter group with a high learning rate (207),
3. The real-time visual tracking method according to claim 1, wherein the multiple model visual tracking step further comprises the sub-steps of:
- Correlation matching using the second filter group of the multiple model visual tracking (205) , if the result of the querying the quality of the first filter group is less than the selected threshold, - Querying the quality of the response of the second filter group to the target in the current frame (206),
- Updating the target location according to the second filter group if the result of the querying the quality of the first filter group is greater than a predefined threshold (208),
- Updating the first filter group with a high learning rate (210).
- Updating the second filter group with a high learning rate (211),
4. The real-time visual tracking method according to claim 1, wherein the multiple model visual tracking step further comprising:
detecting occlusion (209), if the result of querying the second filter group is less than the predefined threshold (206).
5. The multiple model visual tracking according to claim 2 and claim 3, the low learning rate is used for learning the target more slowly than learning the instantaneous target with the high learning rate.
6. The real-time visual tracking method according to claim 1, wherein the target bounding box generation and feedback decision step further comprises the sub-steps of:
- Calculating current saliency map (301),
- Calculating saliency ratio for the current saliency map (303),
- Calculating the correlation score using normalized cross correlation between the target model selected (304) from the updated saliency map using target selection procedure (307) and current saliency map,
- Calculating the learning rate using the saliency ratio and the correlation score (305),
- Updating the saliency map according to the calculated learning rate (306), to obtain the updated saliency map as input , to the system for the next frame (302), - Selecting target according to the updated saliency map (307),
- Querying tracking feedback (308).
7. The real-time visual tracking method according to claim 6, wherein the target bounding box generation and feedback decision step further comprising:
locating the object in the region of interest by generating a minimum size bounding box including the target instead of the region of interest given by the real time visual tracking method.
8. The real-time visual tracking method according to claim 6, wherein the target bounding box generation and feedback decision step further comprising:
compensating for false target initialization via its learning mechanism where the targets, which are not well-localized in initialization, are centralized via the feedback mechanism.
9. The real-time visual tracking method according to claim 6, wherein the target bounding box generation and feedback decision step further comprising:
detecting the scale changes of the target through the video frames, adapting to the scale changes and updating the visual target model more appropriately than the updating the tracking model of a visual tracking method without the ability of detecting the scale changes.
10. The real-time visual tracking method according to claim 6, wherein the target bounding box generation and feedback decision step further comprising:
using an adaptive learning rate selection algorithm, which prevents mislearning of the target model in the cases of clutter or occlusion.
11. The real-time visual tracking method according to claim 10, even though redetection of the target is the merit of any tracking system, the method further comprising:
constructing a target model which is appropriate to be used in order to boost redetection of target after the target is lost due to occlusion, clutter or noise.
12. The method of binarization in claim 6 is the modified version of the equation , which
Figure imgf000023_0001
results in L, number of histogram bins, less multiplication.
13. The real-time visual tracking method according to claim 1, wherein the multiple model visual target tracking further comprising:
adapting to abrupt changes of the region of interest which is the benefit of multiple modeling.
14. The real-time visual tracking method according to claim 1, wherein the multiple model visual tracking step further comprising:
sensing both the rapid and imperceptible changes of the target at the same time without sacrificing either low or high learning rates.
15. The real-time visual tracking method according to claim 2, wherein the multiple model visual tracking step further comprising:
interacting the multiple filter groups to tolerate the errors of each other in different conditions by using their corresponding learning rates when one of the filter groups starts to give low quality tracking results.
16. The real-time visual tracking method according to claim 6, wherein the target bounding box generation and feedback decision step further comprising:
constructing a target model, which is appropriate to be used in a shape based classifier.
17. The real-time visual tracking method according to claim 1, wherein the method of multiple model visual target tracking step further comprising: a robust tracking adaptation under different temporal variation rates by changing the update parameters.
PCT/TR2014/000117 2014-04-22 2014-04-22 Target localization and size estimation via multiple model learning in visual tracking WO2015163830A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/TR2014/000117 WO2015163830A1 (en) 2014-04-22 2014-04-22 Target localization and size estimation via multiple model learning in visual tracking

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/TR2014/000117 WO2015163830A1 (en) 2014-04-22 2014-04-22 Target localization and size estimation via multiple model learning in visual tracking

Publications (1)

Publication Number Publication Date
WO2015163830A1 true WO2015163830A1 (en) 2015-10-29

Family

ID=50928223

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/TR2014/000117 WO2015163830A1 (en) 2014-04-22 2014-04-22 Target localization and size estimation via multiple model learning in visual tracking

Country Status (1)

Country Link
WO (1) WO2015163830A1 (en)

Cited By (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106204639A (en) * 2016-06-27 2016-12-07 开易(北京)科技有限公司 Based on frequency domain regression model target tracking method, system and senior drive assist system
CN106651913A (en) * 2016-11-29 2017-05-10 开易(北京)科技有限公司 Target tracking method based on correlation filtering and color histogram statistics and ADAS (Advanced Driving Assistance System)
CN107154024A (en) * 2017-05-19 2017-09-12 南京理工大学 Dimension self-adaption method for tracking target based on depth characteristic core correlation filter
CN107316316A (en) * 2017-05-19 2017-11-03 南京理工大学 The method for tracking target that filtering technique is closed with nuclear phase is adaptively merged based on multiple features
CN107424177A (en) * 2017-08-11 2017-12-01 哈尔滨工业大学(威海) Positioning amendment long-range track algorithm based on serial correlation wave filter
CN107590820A (en) * 2017-08-25 2018-01-16 北京飞搜科技有限公司 A kind of object video method for tracing and its intelligent apparatus based on correlation filtering
CN108053425A (en) * 2017-12-25 2018-05-18 北京航空航天大学 A kind of high speed correlation filtering method for tracking target based on multi-channel feature
CN108257150A (en) * 2017-12-20 2018-07-06 北京华航无线电测量研究所 A kind of target correlation filtering tracking based on embedded system
CN108280808A (en) * 2017-12-15 2018-07-13 西安电子科技大学 The method for tracking target of correlation filter is exported based on structuring
CN108694725A (en) * 2018-05-14 2018-10-23 肇庆学院 A kind of the Robotic Dynamic tracking and system of view-based access control model conspicuousness
CN108765452A (en) * 2018-05-11 2018-11-06 西安天和防务技术股份有限公司 A kind of detection of mobile target in complex background and tracking
CN108776975A (en) * 2018-05-29 2018-11-09 安徽大学 A kind of visual tracking method based on semi-supervised feature and filter combination learning
CN109002750A (en) * 2017-12-11 2018-12-14 罗普特(厦门)科技集团有限公司 A kind of correlation filtering tracking based on conspicuousness detection and image segmentation
CN109255799A (en) * 2018-07-26 2019-01-22 华中科技大学 A kind of method for tracking target and system based on spatially adaptive correlation filter
CN109685832A (en) * 2018-12-26 2019-04-26 山东创科自动化科技有限公司 A kind of motion target tracking method, device and computer equipment
CN109766752A (en) * 2018-11-28 2019-05-17 西安电子科技大学 A kind of object matching and localization method and system, computer based on deep learning
CN109859246A (en) * 2019-02-15 2019-06-07 哈尔滨工程大学 A kind of low-altitude low-velocity unmanned plane tracking of combination correlation filtering and vision significance
CN109886994A (en) * 2019-01-11 2019-06-14 上海交通大学 Adaptive sheltering detection system and method in video tracking
CN109919970A (en) * 2017-12-12 2019-06-21 武汉盛捷达电力科技有限责任公司 Based on a kind of improved Vision Tracking of MeanShift principle
CN109949344A (en) * 2019-03-18 2019-06-28 吉林大学 It is a kind of to suggest that the nuclear phase of window closes filter tracking method based on color probability target
CN109993769A (en) * 2019-03-07 2019-07-09 安徽创世科技股份有限公司 A kind of multiple-target system of deep learning SSD algorithm combination KCF algorithm
CN110070563A (en) * 2019-04-30 2019-07-30 山东大学 Correlation filter method for tracking target and system based on joint perception
CN110211150A (en) * 2019-04-25 2019-09-06 南开大学 A kind of real-time vision target identification method with scale coordination mechanism
CN110211157A (en) * 2019-06-04 2019-09-06 重庆邮电大学 A kind of target long time-tracking method based on correlation filtering
CN110717934A (en) * 2019-10-17 2020-01-21 湖南大学 Anti-occlusion target tracking method based on STRCF
CN110807795A (en) * 2019-10-31 2020-02-18 北方工业大学 MDnet-based unmanned aerial vehicle remote sensing target tracking method and device
CN110929620A (en) * 2019-11-15 2020-03-27 浙江大华技术股份有限公司 Target tracking method and device and storage device
CN110991565A (en) * 2019-12-24 2020-04-10 华北理工大学 Target tracking optimization algorithm based on KCF
CN111008585A (en) * 2019-11-29 2020-04-14 西安电子科技大学 Ship target detection method based on self-adaptive layered high-resolution SAR image
CN111008996A (en) * 2019-12-09 2020-04-14 华侨大学 Target tracking method through hierarchical feature response fusion
CN111127519A (en) * 2019-12-25 2020-05-08 中国电子科技集团公司信息科学研究院 Target tracking control system and method for dual-model fusion
TWI695343B (en) * 2019-05-06 2020-06-01 中華電信股份有限公司 Automatic labeling method for detecting moving objects
CN111340846A (en) * 2020-02-25 2020-06-26 重庆邮电大学 Multi-feature fusion anti-occlusion target tracking method
CN111383252A (en) * 2018-12-29 2020-07-07 曜科智能科技(上海)有限公司 Multi-camera target tracking method, system, device and storage medium
CN111915653A (en) * 2020-08-17 2020-11-10 沈阳飞机工业(集团)有限公司 Method for tracking double-station visual target
CN112364891A (en) * 2020-10-20 2021-02-12 西安理工大学 Sample classification-based self-adaptive decontamination method
JP2021071769A (en) * 2019-10-29 2021-05-06 オムロン株式会社 Object tracking device and object tracking method
CN112767450A (en) * 2021-01-25 2021-05-07 开放智能机器(上海)有限公司 Multi-loss learning-based related filtering target tracking method and system
CN112819865A (en) * 2021-02-26 2021-05-18 西安理工大学 Correlation filtering tracking method based on self-adaptive regular feature joint time correlation
CN112987571A (en) * 2021-02-25 2021-06-18 中国人民解放军国防科技大学 High dynamic vision control system and vision measurement performance attenuation fault-tolerant control method thereof
CN113160271A (en) * 2021-03-15 2021-07-23 华侨大学 High-precision infrared target tracking method integrating related filtering and particle filtering
CN113327272A (en) * 2021-05-28 2021-08-31 北京理工大学重庆创新中心 Robustness long-time tracking method based on correlation filtering
CN113379798A (en) * 2021-06-03 2021-09-10 中国电子科技集团公司第二十八研究所 Correlation filtering tracking method based on interactive evaluation model
CN113470118A (en) * 2021-07-15 2021-10-01 中国人民解放军火箭军工程大学 Target size estimation system based on trinocular vision
CN113486820A (en) * 2021-07-09 2021-10-08 厦门理工学院 Bidirectional target tracking method and system based on efficient template updating and selecting mechanism
CN113837200A (en) * 2021-08-31 2021-12-24 中国计量大学 Autonomous learning method in visual saliency detection
CN116563348B (en) * 2023-07-06 2023-11-14 中国科学院国家空间科学中心 Infrared weak small target multi-mode tracking method and system based on dual-feature template

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080304740A1 (en) 2007-06-06 2008-12-11 Microsoft Corporation Salient Object Detection
EP2202671A2 (en) * 2008-12-26 2010-06-30 Canon Kabushiki Kaisha Subject tracking apparatus and control method therefor, image capturing apparatus, and display apparatus
US20120288152A1 (en) * 2011-05-10 2012-11-15 Canon Kabushiki Kaisha Object recognition apparatus, control method for object recognition apparatus and storage medium
US20130101210A1 (en) 2011-10-24 2013-04-25 Hao Tang Auto-cropping
US8477998B1 (en) 2008-06-20 2013-07-02 Google Inc. Object tracking in video with visual constraints
US8520956B2 (en) 2009-06-09 2013-08-27 Colorado State University Research Foundation Optimized correlation filters for signal processing

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080304740A1 (en) 2007-06-06 2008-12-11 Microsoft Corporation Salient Object Detection
US8477998B1 (en) 2008-06-20 2013-07-02 Google Inc. Object tracking in video with visual constraints
EP2202671A2 (en) * 2008-12-26 2010-06-30 Canon Kabushiki Kaisha Subject tracking apparatus and control method therefor, image capturing apparatus, and display apparatus
US8520956B2 (en) 2009-06-09 2013-08-27 Colorado State University Research Foundation Optimized correlation filters for signal processing
US20120288152A1 (en) * 2011-05-10 2012-11-15 Canon Kabushiki Kaisha Object recognition apparatus, control method for object recognition apparatus and storage medium
US20130101210A1 (en) 2011-10-24 2013-04-25 Hao Tang Auto-cropping

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
BAKER S ET AL: "The Template Update Problem", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, IEEE COMPUTER SOCIETY, USA, vol. 26, no. 6, 1 June 2004 (2004-06-01), pages 810 - 815, XP011111532, ISSN: 0162-8828, DOI: 10.1109/TPAMI.2004.77 *
NOBUYOKI OTSU: "A threshold selection method from gray-level histograms", IEEE TRANS. SYS. MAN., CYBER., vol. 9, no. 1, 1979, pages 62 - 66
Y. WEI: "Geodesic saliency using background priors", ECCV, 2012
Y. WEI; F. WEN; W. ZHU; J.SUN: "Geodesic Saliency Using Background Priors", IEEE, ICCV, 2012

Cited By (74)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106204639A (en) * 2016-06-27 2016-12-07 开易(北京)科技有限公司 Based on frequency domain regression model target tracking method, system and senior drive assist system
CN106651913A (en) * 2016-11-29 2017-05-10 开易(北京)科技有限公司 Target tracking method based on correlation filtering and color histogram statistics and ADAS (Advanced Driving Assistance System)
CN107154024A (en) * 2017-05-19 2017-09-12 南京理工大学 Dimension self-adaption method for tracking target based on depth characteristic core correlation filter
CN107316316A (en) * 2017-05-19 2017-11-03 南京理工大学 The method for tracking target that filtering technique is closed with nuclear phase is adaptively merged based on multiple features
CN107424177A (en) * 2017-08-11 2017-12-01 哈尔滨工业大学(威海) Positioning amendment long-range track algorithm based on serial correlation wave filter
CN107424177B (en) * 2017-08-11 2021-10-26 哈尔滨工业大学(威海) Positioning correction long-range tracking method based on continuous correlation filter
CN107590820A (en) * 2017-08-25 2018-01-16 北京飞搜科技有限公司 A kind of object video method for tracing and its intelligent apparatus based on correlation filtering
CN107590820B (en) * 2017-08-25 2020-06-02 兰州飞搜信息科技有限公司 Video object tracking method based on correlation filtering and intelligent device thereof
CN109002750A (en) * 2017-12-11 2018-12-14 罗普特(厦门)科技集团有限公司 A kind of correlation filtering tracking based on conspicuousness detection and image segmentation
CN109002750B (en) * 2017-12-11 2021-03-30 罗普特科技集团股份有限公司 Relevant filtering tracking method based on significance detection and image segmentation
CN109919970A (en) * 2017-12-12 2019-06-21 武汉盛捷达电力科技有限责任公司 Based on a kind of improved Vision Tracking of MeanShift principle
CN108280808A (en) * 2017-12-15 2018-07-13 西安电子科技大学 The method for tracking target of correlation filter is exported based on structuring
CN108257150A (en) * 2017-12-20 2018-07-06 北京华航无线电测量研究所 A kind of target correlation filtering tracking based on embedded system
CN108053425A (en) * 2017-12-25 2018-05-18 北京航空航天大学 A kind of high speed correlation filtering method for tracking target based on multi-channel feature
CN108765452A (en) * 2018-05-11 2018-11-06 西安天和防务技术股份有限公司 A kind of detection of mobile target in complex background and tracking
CN108694725A (en) * 2018-05-14 2018-10-23 肇庆学院 A kind of the Robotic Dynamic tracking and system of view-based access control model conspicuousness
CN108776975B (en) * 2018-05-29 2021-11-05 安徽大学 Visual tracking method based on semi-supervised feature and filter joint learning
CN108776975A (en) * 2018-05-29 2018-11-09 安徽大学 A kind of visual tracking method based on semi-supervised feature and filter combination learning
CN109255799A (en) * 2018-07-26 2019-01-22 华中科技大学 A kind of method for tracking target and system based on spatially adaptive correlation filter
CN109255799B (en) * 2018-07-26 2021-07-27 华中科技大学 Target tracking method and system based on spatial adaptive correlation filter
CN109766752B (en) * 2018-11-28 2023-01-03 西安电子科技大学 Target matching and positioning method and system based on deep learning and computer
CN109766752A (en) * 2018-11-28 2019-05-17 西安电子科技大学 A kind of object matching and localization method and system, computer based on deep learning
CN109685832A (en) * 2018-12-26 2019-04-26 山东创科自动化科技有限公司 A kind of motion target tracking method, device and computer equipment
CN111383252A (en) * 2018-12-29 2020-07-07 曜科智能科技(上海)有限公司 Multi-camera target tracking method, system, device and storage medium
CN111383252B (en) * 2018-12-29 2023-03-24 曜科智能科技(上海)有限公司 Multi-camera target tracking method, system, device and storage medium
CN109886994B (en) * 2019-01-11 2021-07-20 上海交通大学 Self-adaptive occlusion detection system and method in video tracking
CN109886994A (en) * 2019-01-11 2019-06-14 上海交通大学 Adaptive sheltering detection system and method in video tracking
CN109859246A (en) * 2019-02-15 2019-06-07 哈尔滨工程大学 A kind of low-altitude low-velocity unmanned plane tracking of combination correlation filtering and vision significance
CN109859246B (en) * 2019-02-15 2022-06-17 哈尔滨工程大学 Low-altitude slow unmanned aerial vehicle tracking method combining correlation filtering and visual saliency
CN109993769A (en) * 2019-03-07 2019-07-09 安徽创世科技股份有限公司 A kind of multiple-target system of deep learning SSD algorithm combination KCF algorithm
CN109993769B (en) * 2019-03-07 2022-09-13 安徽创世科技股份有限公司 Multi-target tracking system combining deep learning SSD algorithm with KCF algorithm
CN109949344A (en) * 2019-03-18 2019-06-28 吉林大学 It is a kind of to suggest that the nuclear phase of window closes filter tracking method based on color probability target
CN109949344B (en) * 2019-03-18 2022-12-27 吉林大学 Nuclear correlation filtering tracking method based on color probability target suggestion window
CN110211150A (en) * 2019-04-25 2019-09-06 南开大学 A kind of real-time vision target identification method with scale coordination mechanism
CN110211150B (en) * 2019-04-25 2023-04-18 南开大学 Real-time visual target identification method with scale coordination mechanism
CN110070563A (en) * 2019-04-30 2019-07-30 山东大学 Correlation filter method for tracking target and system based on joint perception
TWI695343B (en) * 2019-05-06 2020-06-01 中華電信股份有限公司 Automatic labeling method for detecting moving objects
CN110211157A (en) * 2019-06-04 2019-09-06 重庆邮电大学 A kind of target long time-tracking method based on correlation filtering
CN110717934A (en) * 2019-10-17 2020-01-21 湖南大学 Anti-occlusion target tracking method based on STRCF
CN110717934B (en) * 2019-10-17 2023-04-28 湖南大学 Anti-occlusion target tracking method based on STRCF
WO2021084972A1 (en) * 2019-10-29 2021-05-06 オムロン株式会社 Object tracking device and object tracking method
JP2021071769A (en) * 2019-10-29 2021-05-06 オムロン株式会社 Object tracking device and object tracking method
JP7354767B2 (en) 2019-10-29 2023-10-03 オムロン株式会社 Object tracking device and object tracking method
CN110807795B (en) * 2019-10-31 2022-07-22 北方工业大学 MDnet-based unmanned aerial vehicle remote sensing target tracking method and device
CN110807795A (en) * 2019-10-31 2020-02-18 北方工业大学 MDnet-based unmanned aerial vehicle remote sensing target tracking method and device
CN110929620A (en) * 2019-11-15 2020-03-27 浙江大华技术股份有限公司 Target tracking method and device and storage device
CN110929620B (en) * 2019-11-15 2023-04-07 浙江大华技术股份有限公司 Target tracking method and device and storage device
CN111008585B (en) * 2019-11-29 2023-04-07 西安电子科技大学 Ship target detection method based on self-adaptive layered high-resolution SAR image
CN111008585A (en) * 2019-11-29 2020-04-14 西安电子科技大学 Ship target detection method based on self-adaptive layered high-resolution SAR image
CN111008996A (en) * 2019-12-09 2020-04-14 华侨大学 Target tracking method through hierarchical feature response fusion
CN111008996B (en) * 2019-12-09 2023-04-07 华侨大学 Target tracking method through hierarchical feature response fusion
CN110991565A (en) * 2019-12-24 2020-04-10 华北理工大学 Target tracking optimization algorithm based on KCF
CN111127519A (en) * 2019-12-25 2020-05-08 中国电子科技集团公司信息科学研究院 Target tracking control system and method for dual-model fusion
CN111127519B (en) * 2019-12-25 2024-03-12 中国电子科技集团公司信息科学研究院 Dual-model fusion target tracking control system and method thereof
CN111340846A (en) * 2020-02-25 2020-06-26 重庆邮电大学 Multi-feature fusion anti-occlusion target tracking method
CN111340846B (en) * 2020-02-25 2023-02-17 重庆邮电大学 Multi-feature fusion anti-occlusion target tracking method
CN111915653A (en) * 2020-08-17 2020-11-10 沈阳飞机工业(集团)有限公司 Method for tracking double-station visual target
CN112364891A (en) * 2020-10-20 2021-02-12 西安理工大学 Sample classification-based self-adaptive decontamination method
CN112767450A (en) * 2021-01-25 2021-05-07 开放智能机器(上海)有限公司 Multi-loss learning-based related filtering target tracking method and system
CN112987571A (en) * 2021-02-25 2021-06-18 中国人民解放军国防科技大学 High dynamic vision control system and vision measurement performance attenuation fault-tolerant control method thereof
CN112987571B (en) * 2021-02-25 2022-08-09 中国人民解放军国防科技大学 High dynamic vision control system and vision measurement performance attenuation fault-tolerant control method thereof
CN112819865B (en) * 2021-02-26 2024-02-09 西安理工大学 Correlation filtering tracking method based on self-adaptive regular feature joint time correlation
CN112819865A (en) * 2021-02-26 2021-05-18 西安理工大学 Correlation filtering tracking method based on self-adaptive regular feature joint time correlation
CN113160271A (en) * 2021-03-15 2021-07-23 华侨大学 High-precision infrared target tracking method integrating related filtering and particle filtering
CN113160271B (en) * 2021-03-15 2023-10-31 华侨大学 High-precision infrared target tracking method integrating correlation filtering and particle filtering
CN113327272A (en) * 2021-05-28 2021-08-31 北京理工大学重庆创新中心 Robustness long-time tracking method based on correlation filtering
CN113379798A (en) * 2021-06-03 2021-09-10 中国电子科技集团公司第二十八研究所 Correlation filtering tracking method based on interactive evaluation model
CN113379798B (en) * 2021-06-03 2022-11-22 中国电子科技集团公司第二十八研究所 Correlation filtering tracking method based on interactive evaluation model
CN113486820B (en) * 2021-07-09 2023-06-06 厦门理工学院 Bidirectional target tracking method and system based on efficient template updating and selecting mechanism
CN113486820A (en) * 2021-07-09 2021-10-08 厦门理工学院 Bidirectional target tracking method and system based on efficient template updating and selecting mechanism
CN113470118A (en) * 2021-07-15 2021-10-01 中国人民解放军火箭军工程大学 Target size estimation system based on trinocular vision
CN113470118B (en) * 2021-07-15 2023-12-05 中国人民解放军火箭军工程大学 Target size estimation system based on trinocular vision
CN113837200A (en) * 2021-08-31 2021-12-24 中国计量大学 Autonomous learning method in visual saliency detection
CN116563348B (en) * 2023-07-06 2023-11-14 中国科学院国家空间科学中心 Infrared weak small target multi-mode tracking method and system based on dual-feature template

Similar Documents

Publication Publication Date Title
WO2015163830A1 (en) Target localization and size estimation via multiple model learning in visual tracking
US11854240B2 (en) Vision based target tracking that distinguishes facial feature targets
US11302315B2 (en) Digital video fingerprinting using motion segmentation
Bertinetto et al. Staple: Complementary learners for real-time tracking
Yang et al. Robust superpixel tracking
WO2019218824A1 (en) Method for acquiring motion track and device thereof, storage medium, and terminal
US8582887B2 (en) Image processing system, learning device and method, and program
US7940956B2 (en) Tracking apparatus that tracks a face position in a dynamic picture image using ambient information excluding the face
US7526101B2 (en) Tracking objects in videos with adaptive classifiers
US8873798B2 (en) Methods for tracking objects using random projections, distance learning and a hybrid template library and apparatuses thereof
Vojíř et al. The enhanced flock of trackers
US10896495B2 (en) Method for detecting and tracking target object, target object tracking apparatus, and computer-program product
Wang et al. Robust and efficient fragments-based tracking using mean shift
US20100316298A1 (en) Multiple view face tracking
WO2006087581A1 (en) Method for facial features detection
Khan et al. Joint feature correspondences and appearance similarity for robust visual object tracking
CN112785622B (en) Method and device for tracking unmanned captain on water surface and storage medium
WO2015171084A1 (en) A real- time, semi-automatic method for target tracking window initialization in thermal imagery
Xiao et al. An enhanced adaptive coupled-layer LGTracker++
JP2012190159A (en) Information processing device, information processing method, and program
KR102434397B1 (en) Real time multi-object tracking device and method by using global motion
JP2008003749A (en) Feature point detection device, method, and program
Xiao et al. Single-scale siamese network based RGB-D object tracking with adaptive bounding boxes
Fan et al. Closed-loop adaptation for robust tracking
Sliti et al. Efficient visual tracking via sparse representation and back-projection histogram

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14729488

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14729488

Country of ref document: EP

Kind code of ref document: A1