CN111260738A - Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion - Google Patents

Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion Download PDF

Info

Publication number
CN111260738A
CN111260738A CN202010017064.8A CN202010017064A CN111260738A CN 111260738 A CN111260738 A CN 111260738A CN 202010017064 A CN202010017064 A CN 202010017064A CN 111260738 A CN111260738 A CN 111260738A
Authority
CN
China
Prior art keywords
target
image
scale
hog
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010017064.8A
Other languages
Chinese (zh)
Inventor
唐晨
邱岳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN202010017064.8A priority Critical patent/CN111260738A/en
Publication of CN111260738A publication Critical patent/CN111260738A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • G06V10/464Salient features, e.g. scale invariant feature transforms [SIFT] using a plurality of salient features, e.g. bag-of-words [BoW] representations

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the field of computer vision, and aims to realize target tracking in a complex scene, describe a target by fully utilizing the color characteristic and the gradient characteristic of the target, improve the robustness under the conditions of scale change, shielding and the like and have higher precision. The invention discloses a multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion, which comprises the following steps: step 1: inputting a current t frame image; step 2: extracting gradient features and color features of the sample; and step 3: convolving the filter h obtained from the previous frame with the candidate target image block x of the current frame; and 4, step 4: the weights are adaptively distributed to the two characteristics through the magnitude of the response values of the two characteristics of the HOG and the CN; and 5: extracting HOG characteristics of the multi-scale image block; step 6: marking out the target of the current frame by using a rectangular frame; and 7: and updating the translation filter and the scale filter according to a linear interpolation method, and finally realizing target tracking. The invention is mainly applied to the target tracking occasion.

Description

Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion
Technical Field
The invention belongs to the field of computer vision, and relates to a multi-scale target tracking algorithm based on relevant filtering and self-adaptive feature fusion.
Background
The target tracking, which is essentially to robustly estimate the motion state of a moving target in each frame of an image sequence, is one of the popular research subjects in the field of computer vision, and has wide application in the field of computer vision, for example, in the aspects of man-machine interaction, intelligent traffic monitoring, military guidance, robots, etc., the target tracking technology plays a key role. Although the object tracking problem has been studied for decades, many advances have been made in recent years, but it is still a very challenging problem. Many factors affect the performance of the tracking algorithm, such as occlusion, illumination change, scale change, rapid movement, etc., and finding a suitable method to solve the target tracking problem in these complex scenes becomes a key and difficult point of research.
Methods for solving the problem of target tracking can be divided into two broad categories-generation and discrimination. And a generation method, wherein a target area is modeled in the current frame, and the area most similar to the model is found in the next frame and is the prediction position. Known examples include kalman filter, particle filter, Mean Shift (Mean-Shift), and the like. And the discrimination method is to regard the tracking problem as two classification problems of the target and the background on each frame of image, take the target area as a positive sample and the background area as a negative sample, train the classifier by continuously collecting and updating the positive and negative samples in the tracking process, and select the classifier with the maximum response from the current target to be selected as the tracking result of the frame. Among the discriminant methods, the Correlation-based filtering method is a method widely used in recent years, and is represented by a kernel Correlation filtering and tracking algorithm (KCF).
Although the traditional correlation filtering algorithm has small calculation amount, the traditional correlation filtering algorithm has some significant defects in practical application:
(1) objects are not well described using only a single feature.
(2) The scale of the target cannot be accurately estimated to cope with the scale variation problem.
(3) When the target is blocked, the tracking precision is obviously reduced.
(4) The learning rate is a fixed value, and thus, efficient model updating cannot be performed in individual scenes.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention aims to realize target tracking in a complex scene, describe the target by fully utilizing the color characteristic and the gradient characteristic of the target and introduce the scale filter on the basis of translating the filter, so that the robustness under the conditions of scale change, shielding and the like can be improved, and the tracking result has higher precision. Therefore, the technical scheme adopted by the invention is that the multi-scale target tracking method based on the relevant filtering and the self-adaptive feature fusion comprises the following steps:
step 1: inputting a current t frame image;
step 2: collecting image sample blocks, extracting gradient features and color features of the samples, and adopting a Histogram of Oriented Gradients (HOG) feature xHOGAnd color attribute CN (color name) feature xCNExpressing;
and step 3: and (3) convolving the filter h obtained from the previous frame with the candidate target image block x of the current frame, and then searching response vectors of all test samples, wherein the maximum response position is the prediction position of the target. For a given single target image block x, the response output of the classifier is
Figure BDA0002359290940000021
Image sample x expressing HOG featuresHOGAnd CN feature expressed image sample xCNRespectively substituting the above formula to obtain respective related filter response graphs and obtain respective maximum response value RHOGAnd RCN
And 4, step 4: and (3) adaptively allocating weights to the two characteristics according to the magnitude of the response values of the two characteristics, namely HOG and CN, and calculating as shown in formula (2):
Figure BDA0002359290940000022
and performing weighted fusion on the final target tracking result according to the weight distributed by each feature weight for the CN feature and the HOG feature, and calculating as shown in the formula (3):
R=δRCN+(1-δ)RHOG(3)
wherein delta belongs to [0,1 ]. If δ is 0, it represents a tracking result in which only the HOG feature is used for the tracking result; if δ is 1, it represents a tracking result in which the tracking result uses only color features;
and 5: extracting HOG features of multi-scale image blocks, training a least square classifier of a kernel function by using the HOG features of the multi-scale image blocks to obtain a one-dimensional scale tracker, and finally searching the maximum output response R of the filterSThe prediction of the scale is done with equation (4):
Figure BDA0002359290940000023
wherein the parameters
Figure BDA0002359290940000024
To learn the parameters, kSOf (2) element(s)
Figure BDA0002359290940000025
xSIs the target scale model learned from frame t-1,
Figure BDA0002359290940000026
is a sample taken from a new frame;
step 6: the predicted position R and the predicted scale R of the target are comparedSCombining, marking the target of the current frame by using a rectangular frame;
and 7: updating the translation filter and the scale filter according to a linear interpolation method, as shown in the following formulas (5) to (7):
Figure BDA0002359290940000027
Figure BDA0002359290940000028
Figure BDA0002359290940000029
wherein the parameters
Figure BDA00023592909400000210
In order to learn the parameters, the user may,
Figure BDA00023592909400000211
kjis an element of
Figure BDA00023592909400000212
Gamma is the adaptive learning rate and is,
Figure BDA00023592909400000213
λ is a regular term coefficient as a predicted value of the target.
And 8: and inputting the next frame of image, and processing according to the steps to realize target tracking.
A further technical scheme is that in the step 2 and the step 5, the specific extraction process of the HOG features comprises the following steps:
(1) firstly, graying the target image.
(2) The method of Gamma correction is used for carrying out color space normalization on the input image to be detected, thereby achieving the purposes of adjusting the contrast of the image and reducing the influence caused by local shadow and illumination change, and simultaneously inhibiting the interference caused by noise. Gamma compression formula: i (x, y) ═ I (x, y)GammaHere, Gamma is 1/2;
(3) calculating the gradient of each pixel of the target image, wherein the gradient of pixel points in the image is as follows:
Gx(x,y)=H(x+1,y)-H(x-1,y) (8)
Gy(x,y)=H(x,y+1)-H(x,y-1) (9)
in the formula Gx(x,y)、Gy(x, y) and H (x, y) represent pixel values of horizontal and vertical gradients at a pixel point (x, y) in the input image, respectively. The gradient amplitude and gradient direction at pixel point (x, y) are respectively:
Figure BDA0002359290940000031
Figure BDA0002359290940000032
(4) dividing the image into small cell units each containing 4 x 4 pixel points, and constructing a gradient direction histogram for each cell unit;
(5) forming a block by every 3 × 3 cell units, and connecting the feature vectors of the cell units in the block in series to obtain a gradient histogram of the block;
(6) the feature vectors of the gradient histograms of all blocks in the target image are connected in series to form a directional gradient histogram of the target image, namely the finally obtained feature vector used by a discrimination classifier;
in step 7, the specific process of selecting the adaptive learning rate γ is as follows:
the frame difference method, i.e. the difference between two adjacent frames of pictures, is used as the basis to determine the learning rate in a segmented manner, and for an image block x with the size of m multiplied by n, each pixel point uses xijWherein i is 0. ltoreq. m-1 and j is 0. ltoreq. n-1. For the image of the t-th frame,
Figure BDA0002359290940000033
when the value of q is more than 0 and less than 2.5, the change of the target pictures of two adjacent frames is small, the change of the target appearance model is small, and a low learning rate is set to keep good tracking; when q is more than or equal to 2.5 and less than 8, the change of the target appearance model is normal, and the conventional learning rate is set; when q is more than or equal to 8, the change of the target appearance model is obvious, a larger learning rate is set, and the specific segmented learning rate is shown as the formula (13):
Figure BDA0002359290940000034
the invention has the characteristics and beneficial effects that:
in the target tracking method provided by the invention, the description of the target model is not limited to a single gradient feature due to the introduction of the color feature, the problem of scale change is well solved by introducing the scale filter to predict the scale of the target, and the model can be updated to adapt to different changes of a scene due to the application of the self-adaptive learning rate. Therefore, the target tracking method provided by the invention has stronger robustness and reliability in a complex scene.
Drawings
FIG. 1 is a tracking framework flow diagram of the present invention providing a tracking algorithm;
FIG. 2 is a graph comparing partial tracking results on a BlurBody data set by the present invention and a KCF tracking algorithm.
FIG. 2(a) is a partial tracking result of a KCF tracking algorithm on a BlurBody data set;
FIG. 2(b) is a partial tracking result of the tracking algorithm provided by the present invention on the BlurBody data set;
FIG. 3 is a comparison of partial trace results on a CarScale data set by the present invention and a KCF trace algorithm.
FIG. 3(a) is the partial trace result of the KCF trace algorithm on the CarScale data set;
FIG. 3(b) is a partial trace result of the tracing algorithm provided by the present invention on the CarScale data set;
FIG. 4 is a comparison graph of partial tracking results on a Coke data set by the present invention and a KCF tracking algorithm.
FIG. 4(a) is the partial tracking result of the KCF tracking algorithm on the Coke data set;
FIG. 4(b) is a partial tracking result of the tracking algorithm provided by the present invention on the Coke data set;
Detailed Description
In order to overcome the defects of the prior art, the invention aims to realize target tracking in a complex scene, describe the target by fully utilizing the color characteristic and the gradient characteristic of the target and introduce the scale filter on the basis of translating the filter, so that the robustness under the conditions of scale change, shielding and the like can be improved, and the tracking result has higher precision. Therefore, the technical scheme adopted by the invention is a multi-scale target tracking algorithm based on relevant filtering and self-adaptive feature fusion.
The concrete steps are detailed as follows:
step 1: inputting a current t frame image;
step 2: collecting image sample blocks, extracting gradient feature and color feature of the sample, and adopting HOG feature xHOGAnd CN feature xCNAnd (4) expressing.
The specific extraction process of the HOG features comprises the following steps:
(1) firstly, graying the target image.
(2) The method of Gamma correction is used for carrying out color space normalization on the input image to be detected, thereby achieving the purposes of adjusting the contrast of the image and reducing the influence caused by local shadow and illumination change, and simultaneously inhibiting the interference caused by noise. Gamma compression formula: i (x, y) ═ I (x, y)GammaHere, Gamma is 1/2.
(3) The gradient of each pixel of the target image is calculated, and the main purpose is to capture basic contour information and further weaken the interference caused by illumination. The gradient of the pixel points in the image is:
Gx(x,y)=H(x+1,y)-H(x-1,y) (1)
Gy(x,y)=H(x,y+1)-H(x,y-1) (2)
in the formula Gx(x,y)、Gy(x, y) and H (x, y) represent pixel values of horizontal and vertical gradients at a pixel point (x, y) in the input image, respectively. The gradient amplitude and gradient direction at pixel point (x, y) are respectively:
Figure BDA0002359290940000041
Figure BDA0002359290940000051
(4) the image is divided into small cell units each containing 4 x 4 pixels and a gradient direction histogram is constructed for each cell unit.
(5) By grouping every 3 x 3 cell units into a block, and concatenating the respective feature vectors of the cell units included in that block, a gradient histogram of the block can be obtained.
(6) And (3) connecting the gradient histogram feature vectors of all blocks in the target image in series to form a directional gradient histogram of the target image, namely the finally obtained feature vector used by the discrimination classifier.
And step 3: and (3) convolving the filter h obtained from the previous frame with the candidate target image block x of the current frame, and then searching response vectors of all test samples, wherein the maximum response position is the prediction position of the target. For a given single target image block x, the response output of the classifier is
Figure BDA0002359290940000052
Image sample x expressing HOG featuresHOGAnd CN feature expressed image sample xCNRespectively substituting the above formula to obtain respective related filter response graphs and obtain respective maximum response value RHOGAnd RCN
And 4, step 4: and (3) adaptively allocating weights to the two characteristics according to the magnitude of the response values of the two characteristics, namely the HOG and the CN, and calculating the weight as shown in the formula (6):
Figure BDA0002359290940000053
and performing weighted fusion on the final target tracking result according to the weight distributed by each feature weight for the CN feature and the HOG feature, and calculating as shown in the formula (7):
R=δRCN+(1-δ)RHOG(7)
wherein delta belongs to [0,1 ]. If δ is 0, it represents a tracking result in which only the HOG feature is used for the tracking result; if δ is 1, this represents a tracking result in which only the color feature is used for the tracking result.
And 5: extracting HOG features of multi-scale image blocks, training a least square classifier of a kernel function by using the HOG features of the multi-scale image blocks to obtain a one-dimensional scale tracker, and finally searching the maximum output response R of the filterSThe prediction of the scale is accomplished using equation (8).
Figure BDA0002359290940000054
Wherein the parameters
Figure BDA0002359290940000055
To learn the parameters, kSOf (2) element(s)
Figure BDA0002359290940000056
xSIs the target scale model learned from frame t-1,
Figure BDA0002359290940000057
is the sample taken from the new frame.
Step 6: the predicted position R and the predicted scale R of the target are comparedSIn combination, the target of the current frame is marked with a rectangular box.
And 7: updating the translation filter and the scale filter according to a linear interpolation method, as shown in the following formulas (9) to (11):
Figure BDA0002359290940000058
Figure BDA0002359290940000061
Figure BDA0002359290940000062
wherein the parameters
Figure BDA0002359290940000063
In order to learn the parameters, the user may,
Figure BDA0002359290940000064
kjis an element of
Figure BDA0002359290940000065
Gamma is the adaptive learning rate and is,
Figure BDA0002359290940000066
λ is a regular term coefficient as a predicted value of the target.
The specific process of selecting the self-adaptive learning rate gamma is as follows:
the frame difference method, namely the difference between two adjacent frames of pictures is used as a basis to determine the learning rate in a segmented manner. For a frame of image block x with size of mxn, each pixel point uses xijWherein i is 0. ltoreq. m-1 and j is 0. ltoreq. n-1. For the image of the t-th frame,
Figure BDA0002359290940000067
when the value of q is more than 0 and less than 2.5, the change of the target pictures of two adjacent frames is small, the change of the target appearance model is small, and a low learning rate is set to keep good tracking; when q is more than or equal to 2.5 and less than 8, the change of the target appearance model is normal, and the conventional learning rate is set; when q is more than or equal to 8, the change of the target appearance model is obvious, and a larger learning rate is set. The specific segment learning rate is as shown in equation (13):
Figure BDA0002359290940000068
and 8: and inputting the next frame of image, and processing according to the steps to realize target tracking.
To verify the validity of the method, experimental results are given.
The tracking method provided by the invention is adopted to track the video sequences under 3 groups of complex scenes, and the tracking results are compared with the tracking results of the KCF algorithm under the same condition, and the obtained partial tracking results are shown in figures 2, 3 and 4.
In the video sequence shown in fig. 2, a person and a surrounding area are selected as a tracking target, the target in the video generates a motion blur condition, and the scale of the person is changed continuously in the tracking process. Fig. 2(a) and (b) show partial tracking results obtained by the KCF algorithm and the tracking algorithm provided by the present invention, respectively, and the target areas selected by the two algorithms at the initial frame are the same. From the tracking result of fig. 2, it can be known that the tracking algorithm provided by the present invention can well locate the target, the tracking accuracy is higher, and a good tracking effect is obtained. However, the KCF algorithm starts to shift when the moving target becomes blurred, and the center of the subsequent target and the center of the tracking frame shift greatly, resulting in a tracking failure.
In the video sequence shown in fig. 3, a car and its surrounding area are selected as a tracking target, and the car is from far to near during the tracking process, and the scale changes continuously. Fig. 3(a) and (b) show partial tracking results obtained by the KCF algorithm and the tracking algorithm provided by the present invention, respectively, and the target areas selected by the two algorithms at the initial frame are the same. From the tracking result of fig. 3, it can be known that the tracking algorithm provided by the present invention can well locate the target, the tracking accuracy is higher, and a good tracking effect is obtained. The KCF algorithm cannot detect the scale in the video sequence, and the size of the tracking frame is unchanged all the time, so that the tracking precision is influenced.
In the video sequence shown in fig. 4, the pop can and the area around the pop can are selected as the tracking target, and the target in the video is blocked for several times. Fig. 4(a) and (b) show partial tracking results obtained by the KCF algorithm and the tracking algorithm provided by the present invention, respectively, and the target areas selected by the two algorithms at the initial frame are the same. From the tracking result of fig. 4, it can be known that the tracking algorithm provided by the present invention can well locate the target, the tracking accuracy is higher, and a good tracking effect is obtained. When the moving target meets partial shielding, the tracking frame is shifted by the KCF algorithm, and the target is not re-detected subsequently, so that the tracking fails.
Therefore, the multi-scale target tracking algorithm based on the relevant filtering and the self-adaptive feature fusion has stronger robustness on the tracking of the moving target in complex scenes such as motion blur, scale change, occlusion and the like.
While the present invention has been described with reference to the drawings, the foregoing embodiments are illustrative rather than limiting, and that those skilled in the art, having the benefit of the teachings herein, may make numerous modifications thereto without departing from the spirit or scope of the invention as set forth in the appended claims.
Those skilled in the art will appreciate that the drawings are only schematic illustrations of preferred embodiments, and the above-described embodiments of the present invention are merely provided for description and do not represent the merits of the embodiments.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (3)

1. A multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion is characterized by comprising the following steps:
step 1: inputting a current t frame image;
step 2: collecting image sample blocks, extracting gradient features and color features of the samples, and adopting a Histogram of Oriented Gradients (HOG) feature xHOGAnd color attribute CN (color name) feature xCNExpressing;
and step 3: convolving the filter h obtained from the previous frame with the candidate target image block x of the current frame, then searching all test sample response vectors, setting the maximum response position as the predicted position of the target, and outputting the response of the classifier as
Figure FDA0002359290930000011
Image sample x expressing HOG featuresHOGAnd CN feature expressed image sample xCNRespectively substituting the above formula to obtain respective related filter response graphs and obtain respective maximum response value RHOGAnd RCN
And 4, step 4: and (3) adaptively allocating weights to the two characteristics according to the magnitude of the response values of the two characteristics, namely HOG and CN, and calculating as shown in formula (2):
Figure FDA0002359290930000012
and performing weighted fusion on the final target tracking result according to the weight distributed by each feature weight for the CN feature and the HOG feature, and calculating as shown in the formula (3):
R=δRCN+(1-δ)RHOG(3)
wherein δ belongs to [0,1], if δ is 0, the tracking result represents the tracking result only using the HOG feature; if δ is 1, it represents a tracking result in which the tracking result uses only color features;
and 5: extracting HOG features of multi-scale image blocks, training a least square classifier of a kernel function by using the HOG features of the multi-scale image blocks to obtain a one-dimensional scale tracker, and finally searching the maximum output response R of the filterSThe prediction of the scale is done with equation (4):
Figure FDA0002359290930000013
wherein the parameters
Figure FDA0002359290930000014
To learn the parameters, kSOf (2) element(s)
Figure FDA0002359290930000015
xSIs the target scale model learned from frame t-1,
Figure FDA0002359290930000016
is a sample taken from a new frame;
step 6: the predicted position R and the predicted scale R of the target are comparedSCombining, marking the target of the current frame by using a rectangular frame;
and 7: updating the translation filter and the scale filter according to a linear interpolation method, as shown in the following formulas (5) to (7):
Figure FDA0002359290930000017
Figure FDA0002359290930000018
Figure FDA0002359290930000019
wherein the parameters
Figure FDA00023592909300000110
In order to learn the parameters, the user may,
Figure FDA00023592909300000111
kjis an element of
Figure FDA00023592909300000112
Gamma is the adaptive learning rate and is,
Figure FDA0002359290930000021
the lambda is a regular term coefficient;
and 8: and inputting the next frame of image, processing according to the steps 1-7, and finally realizing target tracking.
2. The multi-scale target tracking method based on correlation filtering and adaptive feature fusion as claimed in claim 1, wherein in the step 2 and the step 5, the specific extraction process of the HOG features is as follows:
(1) firstly, graying a target image;
(2) the method of Gamma correction is used for carrying out color space normalization on an input image to be detected, thereby achieving the purposes of adjusting the contrast of the image and reducing the influence caused by local shadow and illumination change, and simultaneously inhibiting the interference caused by noise, and the Gamma compression formula: i (x, y) ═ I (x, y)GammaHere, Gamma is 1/2;
(3) calculating the gradient of each pixel of the target image, wherein the gradient of pixel points in the image is as follows:
Gx(x,y)=H(x+1,y)-H(x-1,y) (8)
Gy(x,y)=H(x,y+1)-H(x,y-1) (9)
in the formula Gx(x,y)、Gy(x, y) and H (x, y) respectively represent pixel values of gradients in the horizontal direction and the vertical direction at a pixel point (x, y) in the input image, and the gradient magnitude and the gradient direction at the pixel point (x, y) are respectively:
Figure FDA0002359290930000022
Figure FDA0002359290930000023
(4) dividing the image into small cell units each containing 4 x 4 pixel points, and constructing a gradient direction histogram for each cell unit;
(5) forming a block by every 3 × 3 cell units, and connecting the feature vectors of the cell units in the block in series to obtain a gradient histogram of the block;
(6) and (3) connecting the gradient histogram feature vectors of all blocks in the target image in series to form a directional gradient histogram of the target image, namely the finally obtained feature vector used by the discrimination classifier.
3. The multi-scale target tracking method based on correlation filtering and adaptive feature fusion as claimed in claim 1, wherein in step 7, the specific process of adaptive learning rate γ selection adopts a frame difference method, i.e. the difference between two adjacent frames of pictures is used as the basis to determine the learning rate in a segmented manner, and for a frame of image block x with the size of mxn, each pixel point uses xijWhere 0 ≦ i ≦ m-1, 0 ≦ j ≦ n-1, for the t-th frame image:
Figure FDA0002359290930000024
when the value of q is more than 0 and less than 2.5, the change of the target pictures of two adjacent frames is small, the change of the target appearance model is small, and a low learning rate is set to keep good tracking; when q is more than or equal to 2.5 and less than 8, the change of the target appearance model is normal, and the conventional learning rate is set; when q is more than or equal to 8, the change of the target appearance model is obvious, a larger learning rate is set, and the specific segmented learning rate is shown as the formula (13):
Figure FDA0002359290930000031
CN202010017064.8A 2020-01-08 2020-01-08 Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion Pending CN111260738A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010017064.8A CN111260738A (en) 2020-01-08 2020-01-08 Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010017064.8A CN111260738A (en) 2020-01-08 2020-01-08 Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion

Publications (1)

Publication Number Publication Date
CN111260738A true CN111260738A (en) 2020-06-09

Family

ID=70950960

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010017064.8A Pending CN111260738A (en) 2020-01-08 2020-01-08 Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion

Country Status (1)

Country Link
CN (1) CN111260738A (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112164093A (en) * 2020-08-27 2021-01-01 同济大学 Automatic person tracking method based on edge features and related filtering
CN112200833A (en) * 2020-09-17 2021-01-08 天津城建大学 Relevant filtering video tracking algorithm based on residual error network and short-term visual memory
CN112329784A (en) * 2020-11-23 2021-02-05 桂林电子科技大学 Correlation filtering tracking method based on space-time perception and multimodal response
CN112435280A (en) * 2020-11-13 2021-03-02 桂林电子科技大学 Moving target detection and tracking method for unmanned aerial vehicle video
CN112598710A (en) * 2020-12-25 2021-04-02 杭州电子科技大学 Space-time correlation filtering target tracking method based on feature online selection
CN112651999A (en) * 2021-01-19 2021-04-13 滨州学院 Unmanned aerial vehicle ground target real-time tracking method based on space-time context perception
CN112767437A (en) * 2020-12-30 2021-05-07 大连海事大学 Water surface unmanned ship tracking method, system and storage medium based on KCF self-adaptive multi-feature fusion filtering
CN113177970A (en) * 2021-04-29 2021-07-27 燕山大学 Multi-scale filtering target tracking method based on self-adaptive feature fusion
CN113593250A (en) * 2021-07-12 2021-11-02 浙江工贸职业技术学院 Illegal parking detection system based on visual identification
CN115482254A (en) * 2022-06-16 2022-12-16 沈阳建筑大学 KCF method with scale self-adaption and anti-blocking functions
CN116405614A (en) * 2022-06-10 2023-07-07 上海玄戒技术有限公司 Lens shading correction method and device, electronic equipment, chip and medium
CN116912289A (en) * 2023-08-09 2023-10-20 北京航空航天大学 Weak and small target layering visual tracking method oriented to edge intelligence

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107644430A (en) * 2017-07-27 2018-01-30 孙战里 Target following based on self-adaptive features fusion
CN108549839A (en) * 2018-03-13 2018-09-18 华侨大学 The multiple dimensioned correlation filtering visual tracking method of self-adaptive features fusion

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107644430A (en) * 2017-07-27 2018-01-30 孙战里 Target following based on self-adaptive features fusion
CN108549839A (en) * 2018-03-13 2018-09-18 华侨大学 The multiple dimensioned correlation filtering visual tracking method of self-adaptive features fusion

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
赵璐璐: "基于相关滤波的目标跟踪算法研究", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112164093A (en) * 2020-08-27 2021-01-01 同济大学 Automatic person tracking method based on edge features and related filtering
CN112200833A (en) * 2020-09-17 2021-01-08 天津城建大学 Relevant filtering video tracking algorithm based on residual error network and short-term visual memory
CN112435280A (en) * 2020-11-13 2021-03-02 桂林电子科技大学 Moving target detection and tracking method for unmanned aerial vehicle video
CN112329784A (en) * 2020-11-23 2021-02-05 桂林电子科技大学 Correlation filtering tracking method based on space-time perception and multimodal response
CN112598710A (en) * 2020-12-25 2021-04-02 杭州电子科技大学 Space-time correlation filtering target tracking method based on feature online selection
CN112598710B (en) * 2020-12-25 2024-03-12 杭州电子科技大学 Space-time correlation filtering target tracking method based on feature on-line selection
CN112767437A (en) * 2020-12-30 2021-05-07 大连海事大学 Water surface unmanned ship tracking method, system and storage medium based on KCF self-adaptive multi-feature fusion filtering
CN112651999A (en) * 2021-01-19 2021-04-13 滨州学院 Unmanned aerial vehicle ground target real-time tracking method based on space-time context perception
CN113177970A (en) * 2021-04-29 2021-07-27 燕山大学 Multi-scale filtering target tracking method based on self-adaptive feature fusion
CN113593250A (en) * 2021-07-12 2021-11-02 浙江工贸职业技术学院 Illegal parking detection system based on visual identification
CN116405614A (en) * 2022-06-10 2023-07-07 上海玄戒技术有限公司 Lens shading correction method and device, electronic equipment, chip and medium
CN116405614B (en) * 2022-06-10 2024-01-30 上海玄戒技术有限公司 Lens shading correction method and device, electronic equipment, chip and medium
CN115482254A (en) * 2022-06-16 2022-12-16 沈阳建筑大学 KCF method with scale self-adaption and anti-blocking functions
CN116912289A (en) * 2023-08-09 2023-10-20 北京航空航天大学 Weak and small target layering visual tracking method oriented to edge intelligence
CN116912289B (en) * 2023-08-09 2024-01-30 北京航空航天大学 Weak and small target layering visual tracking method oriented to edge intelligence

Similar Documents

Publication Publication Date Title
CN111260738A (en) Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion
CN108549839B (en) Adaptive feature fusion multi-scale correlation filtering visual tracking method
CN110929593B (en) Real-time significance pedestrian detection method based on detail discrimination
CN110120064B (en) Depth-related target tracking algorithm based on mutual reinforcement and multi-attention mechanism learning
CN107169994B (en) Correlation filtering tracking method based on multi-feature fusion
CN110097575B (en) Target tracking method based on local features and scale pool
CN110853074B (en) Video target detection network system for enhancing targets by utilizing optical flow
CN110428450B (en) Scale-adaptive target tracking method applied to mine tunnel mobile inspection image
CN111582349B (en) Improved target tracking algorithm based on YOLOv3 and kernel correlation filtering
CN110298297A (en) Flame identification method and device
CN109993052B (en) Scale-adaptive target tracking method and system under complex scene
CN110472577B (en) Long-term video tracking method based on adaptive correlation filtering
CN105741319B (en) Improvement visual background extracting method based on blindly more new strategy and foreground model
CN107609571B (en) Adaptive target tracking method based on LARK features
CN111046789A (en) Pedestrian re-identification method
CN111402237A (en) Video image anomaly detection method and system based on space-time cascade self-encoder
CN112329656B (en) Feature extraction method for human action key frame in video stream
CN112036381B (en) Visual tracking method, video monitoring method and terminal equipment
CN112464844A (en) Human behavior and action recognition method based on deep learning and moving target detection
CN111340842A (en) Correlation filtering target tracking algorithm based on joint model
CN111626090A (en) Moving target detection method based on depth frame difference convolutional neural network
CN114973112A (en) Scale-adaptive dense crowd counting method based on antagonistic learning network
CN112329784A (en) Correlation filtering tracking method based on space-time perception and multimodal response
Iraei et al. Object tracking with occlusion handling using mean shift, Kalman filter and edge histogram
CN110827262A (en) Weak and small target detection method based on continuous limited frame infrared image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200609

WD01 Invention patent application deemed withdrawn after publication