CN108830878A - A kind of method for tracking target based on FPN neural network - Google Patents
A kind of method for tracking target based on FPN neural network Download PDFInfo
- Publication number
- CN108830878A CN108830878A CN201810329415.1A CN201810329415A CN108830878A CN 108830878 A CN108830878 A CN 108830878A CN 201810329415 A CN201810329415 A CN 201810329415A CN 108830878 A CN108830878 A CN 108830878A
- Authority
- CN
- China
- Prior art keywords
- neural network
- tracking
- fpn
- target
- candidate sample
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 55
- 238000000034 method Methods 0.000 title claims abstract description 26
- 239000011159 matrix material Substances 0.000 claims description 29
- 238000012549 training Methods 0.000 claims description 16
- 238000005070 sampling Methods 0.000 claims description 15
- 230000004044 response Effects 0.000 claims description 12
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 6
- 238000006243 chemical reaction Methods 0.000 claims description 4
- 238000005215 recombination Methods 0.000 claims 1
- 230000006798 recombination Effects 0.000 claims 1
- 230000004927 fusion Effects 0.000 abstract 1
- 239000010410 layer Substances 0.000 description 20
- 230000000694 effects Effects 0.000 description 5
- 238000013527 convolutional neural network Methods 0.000 description 4
- 125000004122 cyclic group Chemical group 0.000 description 3
- 238000000605 extraction Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 241000195940 Bryophyta Species 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000012512 characterization method Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000013139 quantization Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 239000002356 single layer Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20048—Transform domain processing
- G06T2207/20056—Discrete and fast Fourier transform, [DFT, FFT]
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Biophysics (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention proposes a kind of method for tracking target based on FPN neural network.This method does not use traditional VGG deep neural network, but uses FPN deep neural network instead.Using the good fusion faculty of discriminating power possessed by spatial information possessed by the depth characteristic figure in FPN deep neural network to the output of neural network shallow-layer network and the depth characteristic figure of deep layer network output, to improve target tracking accuracy.This method is a kind of track algorithm of realtime robustness, is achieved good results in different tracking scenes.
Description
Technical Field
The invention belongs to the technical field of computer vision, and particularly relates to a target tracking method based on an FPN neural network.
Background
Target tracking is one of the most active researches on computer vision technology because of wide application in many fields such as behavior analysis, vehicle navigation, human-computer interaction, medical imaging, video monitoring and the like. Target tracking refers to the location of a target in the first frame of a given video, with each frame following it, for target localization. The core problem of target tracking follows targets that change over time. Although the target tracking algorithm is rapidly developed under the continuous research of scholars at home and abroad in recent years, a good effect cannot be obtained under the conditions of severe illumination change, rapid target movement, partial shielding and the like.
In recent years, scholars at home and abroad propose various tracking algorithms which can be mainly divided into two types: one is based on generative models describing and characterizing the target itself; another class is discriminant models that aim at separating objects from the background. The generative model focuses on establishing a representation of the target appearance model, and although it is important to construct an effective appearance model to handle various challenging situations in tracking, at the same time, it also adds significant computational complexity and discards useful information around the target area that can be used to better separate the object from the background; the discriminant model converts the tracking problem into a two-classification problem of a target and a background, namely, the tracked target is used as a foreground, and the foreground target and the background are distinguished by using a judger for online learning or offline training, so that the position of the foreground target is obtained. Before judgment, feature extraction is often performed to serve as a judgment basis to improve the judgment accuracy, but this also results in a large number of candidate samples needing feature extraction, so that the real-time performance is difficult to achieve.
Correlation filtering is a conventional signal processing method that describes the degree of similarity between two samples. The 2010 MOSSE algorithm is introduced into target tracking, so that the tracking algorithm speed reaches a high speed state, but the accuracy is low due to the fact that random sampling is adopted, and the number of positive and negative training is insufficient. The CSK algorithm in 2012 based on the MOSSE algorithm carries out dense sampling on the target by establishing a cyclic shift structure, so that the number of positive and negative samples is increased, and the problem of insufficient training samples in target tracking is solved. Besides, the calculation of the target sample is converted into the solution in the frequency domain by processing the samples through cyclic displacement, and the efficiency of target tracking is greatly improved by using a fast Fourier transform method. However, the CSK algorithm adopts a single-channel gray scale feature, and is not robust enough in feature characterization. Aiming at the problems, in 2015, the CN algorithm is changed to be the color characteristic of multiple channels, and the KCF algorithm is changed to be the HOG characteristic of multiple channels, so that the precision is improved. However, the CN algorithm and the KCF algorithm use fixed-size templates in convolution solution, so that the model has no scale self-adaptation function, the DSST algorithm is additionally provided with a scale filter on the basis of an original position filter, the FDSST algorithm is improved on the basis of the DSST algorithm to increase the tracking speed of the DSST algorithm, and the SAMF algorithm obtains candidate samples through multi-scale sampling to enable the model to have scale adaptability. Since the number of positive and negative samples is increased by using the cyclic shift to construct the samples, the image pixels cross the boundary, so that the false samples are generated, and the discrimination force of the classifier is reduced, namely the boundary effect. The SRDCF algorithm proposed in 2015 greatly reduces the boundary effect and improves the tracking accuracy by introducing a regularization weight coefficient that conforms to the spatial constraint. The tracking algorithm based on the correlation filtering belongs to a discriminant tracking algorithm, and features are often extracted before judgment so as to serve as judgment basis to improve the judgment accuracy, and the characterization capability of the features determines the tracking effect to a great extent. Since the first time Hinton took advantage of the AlexNet deep convolutional neural network in the ImageNet image classification competition in 2012, the deep convolutional neural network started to rise, which also exhibited impressive performance in many tasks, in particular, the strong feature extraction capability. In 2015, the deep convolutional neural network of the VGG is proposed to be applied to the SRDCF algorithm by the DeepSRDCF algorithm, so that the precision is further improved, but only a feature map output by a single-layer neural network can be used, and the original potential of the deep convolutional neural network is greatly limited. The 2016C-COT algorithm enables the depth feature map output by the multi-layer neural network to be effectively combined through a method of training a continuous convolution filter. However, the way of combining the depth feature maps output by the multilayer neural network with the C-COT algorithm is only to simply combine the depth feature maps in a high-dimensional space, and does not well integrate the spatial information of the depth feature maps output by the shallow network of the neural network and the discriminative power of the depth feature maps output by the deep network.
Aiming at the problem that the existing tracking algorithm cannot well integrate the spatial information of the depth feature map output by the shallow network of the neural network and the discrimination capability of the depth feature map output by the deep network, a tracking algorithm needs to be designed, so that the spatial information of the depth feature map output by the shallow network of the neural network and the discrimination capability of the depth feature map output by the deep network can be well utilized, and the tracking precision is improved.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a target tracking method based on an FPN neural network.
In order to achieve the purpose, the invention adopts the following technical scheme:
a target tracking method based on an FPN neural network comprises the following specific steps:
step one, for the first frame image t being 1, the central position (x) of the first frame tracking target given by the tracking taskt,yt) And tracking target area size information (l)t,ht) Expanding the tracking target area by a certain ratio (l)p,t,hp,t)=α(lt,ht) According to the target center position (x)t,yt) And the expanded size (l) of the tracking target regionp,t,hp,t) Sampling in the frame image to obtain a training sample; wherein xtFor tracking the abscissa, y, of the central position of the targettFor tracking the ordinate of the central position of the target,/tFor the length of the original tracking target region, htWidth of original tracking target region, expansion ratio of α, lp,tTo extend the length of the trailing target area, hp,tTo expand the width of the target area to be tracked;
step two, inputting training samples T obtained by sampling in the first frame image into the FPN neural network, and extracting P in the FPN neural network2Characteristics of the layerWherein T represents a training sample, ZTP representing the number of channels n in an FPN neural network2The characteristics of the layers are such that,representing P in FPN neural network2Characteristic Z of the layerTA medium nth dimension feature;
step three, extracting P obtained by passing the training sample T through the FPN neural network2Characteristics of the layerFor calculation of the correlation filter parameters;
step four, for the next frame of image t +1, tracking the central position (x) of the target according to the previous framet,yt) And the expanded size (l) of the tracking target regionp,t,hp,t) For the expanded tracking target area (l) of the previous framep,t,hp,t) Carrying out multi-scale scaling to obtain various candidate region sizes { (l)p,t+1,hp,t+1)}={β(lp,t,hp,t) Wherein β is a scaling scale, β ═ {0.985,0.99, 0.995,1,1.005,1.01,1.015}, and then, the target center position (x) is tracked according to the previous framet,yt) And a plurality of candidate region sizes { (l)p,t+1,hp,t+1)}={β(lp,t,hp,t) Sampling the frame image to obtain a candidate sample set X ═ X (X)1X2… X7);
Inputting the candidate sample set X obtained by sampling into the FPN neural network, and extracting P in the FPN neural network2Characteristics of the layerWhereinRepresenting the first candidate sample X1Inputting the data into an FPN neural network and providing P with the number of channels n2Layer characteristics, i.e.
Step six, extracting P obtained by passing each candidate sample of the candidate sample set through an FPN neural network2Characteristic Z of the layerXFor calculating the response map, and finally determining the central position (x) of the tracking target of the framet+1,yt+1) And tracking target area size (l)t+1,ht+1);
And step seven, after the central position and the area size of the tracking target are obtained, continuously repeating the step one to the step six until the video is finished, and completing the tracking of the central position and the area size of the tracking target.
The calculation of the parameters of the relevant filter in the third step specifically includes:
first, for P2Characteristics of the layerPerforming fast Fourier transform to transform the features from time domain to frequency domainTo obtain
Second, for each feature channelk is 1,2, … n, which is vectorized and then reconstructed into a diagonal matrix, i.e., a matrix with a high degree of freedom
Diagonal matrix constructed from all channels in a featurek 1,2, … n are recombined to form a diagonal matrix, i.e.
Using this diagonal matrix DtCalculating to obtain AtAndwherein,
wherein W is a regularization matrix constructed by the spatial regularization coefficients;the method comprises the steps of obtaining a Gaussian label through fast Fourier transform and real quantization;
finally, according to the formulaAndcalculating to obtain related filtering parametersThe specific method comprises the following steps: firstly, matrix A is formed1Decomposed into an upper triangular matrix L1And a lower triangular matrix U1I.e. A1=L1+U1(ii) a Then, Gaussian-Seidel iteration is carried out to obtain relevant filter parameters after real numberFinally by calculatingDeriving correlation filter parametersWhere B is a unitary matrix of orthonormal bases, the function of which is to correlate the filter parametersAnd (5) carrying out real number conversion.
The method comprises the following steps:
firstly, each candidate sample of the candidate sample set is extracted to obtain P after passing through an FPN neural network2Characteristic Z of the layerXPerforming fast Fourier transform to obtain frequency domain characteristicsWhereink is 1, and 2 … 7 is the kth candidate sample P2Characteristics of the layerObtaining frequency domain characteristics through fast Fourier transform
Then, for each candidate sample in the candidate sample set, the characteristic of fast Fourier transform is performedk is 1,2 … 7, which is related to the filter parameters after fast Fourier transformPerform dot multiplication, andthen, performing inverse Fourier transform to obtain a response map corresponding to the candidate samplek =1,2 … 7, i.e.
WhereinFor the kth candidate sample XkK is the corresponding graph of 1,2 … 7, F-1Is an inverse Fourier transform;
finally, response graphs corresponding to all candidate samples in the candidate sample setFind the point (x) where the response value is maximumt+1,yt+1) The corresponding position of the point is the tracking target position of the frame, and the size of the candidate sample corresponding to the point is the enlarged tracking target area size (l)p,t+1,hp,t+1) Then, the target region size is calculated from the enlargement ratio α
Compared with the prior art, the method has the following outstanding advantages:
the method does not use the traditional VGG deep neural network, but uses the FPN deep neural network instead, and utilizes the space information of the depth characteristic diagram output by the shallow network of the neural network in the FPN deep neural network and the good judgment capability of the depth characteristic diagram output by the deep network, thereby improving the target tracking precision. The method is a real-time robust tracking algorithm, and achieves good effect in different tracking scenes.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
Detailed Description
The following further describes a specific embodiment of the present invention with reference to the drawings.
As shown in fig. 1, a target tracking method based on an FPN neural network includes the following specific steps:
step one, for the first frame image t being 1, the central position (x) of the first frame tracking target given by the tracking taskt,yt) And tracking target area size information (l)t,ht) Expanding the tracking target area by a certain ratio (l)p,t,hp,t)=α(lt,ht) According to the target center position (x)t,yt) And the expanded size (l) of the tracking target regionp,t,hp,t) Sampling in the frame image to obtain a training sample; wherein xtFor tracking the abscissa, y, of the central position of the targettFor tracking the ordinate of the central position of the target,/tFor the length of the original tracking target region, htWidth of original tracking target region, expansion ratio of α, lp,tTo extend the length of the trailing target area, hp,tTo expand the width of the target area to be tracked;
step two, inputting training samples T obtained by sampling in the first frame image into the FPN neural network, and extracting P in the FPN neural network2Characteristics of the layerWherein T represents a training sample, ZTP representing the number of channels n in an FPN neural network2The characteristics of the layers are such that,representing P in FPN neural network2Characteristic Z of the layerTA medium nth dimension feature;
step three, extracting P obtained by passing the training sample T through the FPN neural network2Characteristics of the layerFor calculation of the correlation filter parameters;
first, for P2Characteristics of the layerPerforming fast Fourier transform to transform the features from time domain to frequency domainTo obtain
Second, for each feature channelk is 1,2, … n, which is vectorized and then reconstructed into a diagonal matrix, i.e., a matrix with a high degree of freedom
Diagonal matrix constructed from all channels in a featurek 1,2, … n are recombined to form a diagonal matrix, i.e.
Using this diagonal matrix DtCalculating to obtain AtAndwherein,
wherein W is a regularization matrix constructed by the spatial regularization coefficients;the method comprises the steps of obtaining a Gaussian label through fast Fourier transform and real quantization;
finally, according to the formulaAndcalculating to obtain related filtering parametersThe specific method comprises the following steps: firstly, matrix A is formed1Decomposed into an upper triangular matrix L1And a lower triangular matrix U1I.e. A1=L1+U1(ii) a Then, Gaussian-Seidel iteration is carried out to obtain relevant filter parameters after real numberFinally by calculatingDeriving correlation filter parametersWhere B is a unitary matrix of orthonormal bases, the function of which is to correlate the filter parametersAnd (5) carrying out real number conversion.
Step four, for the next frame of image t +1, tracking the central position (x) of the target according to the previous framet,yt) And the expanded size (l) of the tracking target regionp,t,hp,t) For the expanded tracking target area (l) of the previous framep,t,hp,t) Carrying out multi-scale scaling to obtain various candidate region sizes { (l)p,t+1,hp,t+1)}={β(lp,t,hp,t) Wherein β is a scaling scale, β ═ {0.985,0.99, 0.995,1,1.005,1.01,1.015}, and then, the target center position (x) is tracked according to the previous framet,yt) And a plurality of candidate region sizes { (l)p,t+1,hp,t+1)}={β(lp,t,hp,t) Sampling the frame image to obtain a candidate sample set X ═ X (X)1X2… X7);
Inputting the candidate sample set X obtained by sampling into the FPN neural network, and extracting P in the FPN neural network2Characteristics of the layerWhereinRepresenting the first candidate sample X1Inputting the data into an FPN neural network and providing P with the number of channels n2Layer characteristics, i.e.
Step six, extracting P obtained by passing each candidate sample of the candidate sample set through an FPN neural network2Characteristics of the layerZXFor calculating the response map, and finally determining the central position (x) of the tracking target of the framet+1,yt+1) And tracking target area size (l)t+1,ht+1);
Firstly, each candidate sample of the candidate sample set is extracted to obtain P after passing through an FPN neural network2Characteristic Z of the layerXPerforming fast Fourier transform to obtain frequency domain characteristicsWhereink is 1, and 2 … 7 is the kth candidate sample P2Characteristics of the layerObtaining frequency domain characteristics through fast Fourier transform
Then, for each candidate sample in the candidate sample set, the characteristic of fast Fourier transform is performedk is 1,2 … 7, which is related to the filter parameters after fast Fourier transformPerforming dot multiplication and then performing inverse Fourier transform to obtain a response map corresponding to the candidate samplek is 1,2 … 7, i.e.
WhereinFor the kth candidate sample XkK is the corresponding graph of 1,2 … 7, F-1Is an inverse Fourier transform;
finally, response graphs corresponding to all candidate samples in the candidate sample setFind the point (x) where the response value is maximumt+1,yt+1) The corresponding position of the point is the tracking target position of the frame, and the size of the candidate sample corresponding to the point is the enlarged tracking target area size (l)p,t+1,hp,t+1) Then, the target region size is calculated from the enlargement ratio α
And step seven, after the central position and the area size of the tracking target are obtained, continuously repeating the step one to the step six until the video is finished, and completing the tracking of the central position and the area size of the tracking target.
Claims (4)
1. A target tracking method based on an FPN neural network is characterized by comprising the following specific steps:
step one, for the first frame image t being 1, the central position (x) of the first frame tracking target given by the tracking taskt,yt) And tracking target area size information (l)t,ht) Expanding the tracking target area by a certain ratio (l)p,t,hp,t)=α(lt,ht) According to the target center position (x)t,yt) And the expanded size (l) of the tracking target regionp,t,hp,t) Sampling in the frame image to obtain a training sample; wherein xtFor tracking the abscissa, y, of the central position of the targettFor tracking the ordinate of the central position of the target,/tFor the length of the original tracking target region, htWidth of original tracking target region, expansion ratio of α, lp,tTo extend the length of the trailing target area, hp,tTo expand the width of the target area to be tracked;
step two, inputting training samples T obtained by sampling in the first frame image into the FPN neural network, and extracting P in the FPN neural network2Characteristics of the layerWherein T represents a training sample, ZTP representing the number of channels n in an FPN neural network2The characteristics of the layers are such that,representing P in FPN neural network2Characteristic Z of the layerTA medium nth dimension feature;
step three, extracting P obtained by passing the training sample T through the FPN neural network2Characteristics of the layerFor calculation of the correlation filter parameters;
step four, for the next frame of image t +1, tracking the central position (x) of the target according to the previous framet,yt) And the expanded size (l) of the tracking target regionp,t,hp,t) For the expanded tracking target area (l) of the previous framep,t,hp,t) Carrying out multi-scale scaling to obtain various candidate region sizes { (l)p,t+1,hp,t+1)}={β(lp,t,hp,t) Wherein β is a scaling scale, β ═ {0.985,0.99, 0.995,1,1.005,1.01,1.015}, and then, the target center position (x) is tracked according to the previous framet,yt) And a plurality of candidate region sizes { (l)p,t+1,hp,t+1)}={β(lp,t,hp,t) Sampling the frame image to obtain a candidate sample set X ═ X (X)1X2… X7);
Inputting the candidate sample set X obtained by sampling into the FPN neural network, and extracting P in the FPN neural network2Characteristics of the layerWhereinRepresenting the first candidate sample X1Inputting the data into an FPN neural network and providing P with the number of channels n2Layer characteristics, i.e.
Step six, extracting P obtained by passing each candidate sample of the candidate sample set through an FPN neural network2Characteristic Z of the layerXFor calculating the response map, and finally determining the central position (x) of the tracking target of the framet+1,yt+1) And tracking target area size (l)t+1,ht+1);
And step seven, after the central position and the area size of the tracking target are obtained, continuously repeating the step one to the step six until the video is finished, and completing the tracking of the central position and the area size of the tracking target.
2. The target tracking method based on the FPN neural network of claim 1, wherein the calculation of the relevant filter parameters in step three is specifically:
first, for P2Characteristics of the layerPerforming fast Fourier transform to transform the features from time domain to frequency domainTo obtain
Second, for each feature channelVectorised and then reconstructed into a diagonal matrix, i.e.
Diagonal matrix constructed from all channels in a featureThe recombination is configured as a diagonal matrix, i.e.
Using this diagonal matrix DtCalculating to obtain AtAndwherein,
wherein W is a regularization matrix constructed by the spatial regularization coefficients;obtained by fast Fourier transform and real number of Gaussian labelThe label of (1);
finally, according to the formulaAndcalculating to obtain related filtering parameters after fast Fourier transform
3. The target tracking method based on the FPN neural network of claim 2, wherein the calculation is performed to obtain the relevant filtering parameters by the following specific method: firstly, matrix A is formedtDecomposed into an upper triangular matrix LtAnd a lower triangular matrix UtI.e. At=Lt+Ut(ii) a Then, Gaussian-Seidel iteration is carried out to obtain related filter parameters which are subjected to fast Fourier transform and real number conversionFinally by calculatingObtaining the parameters of the related filter after fast Fourier transformWherein B is unitary matrix composed of orthonormal basis and is used for fast Fourier transform of related filter parametersAnd (5) carrying out real number conversion.
4. The target tracking method based on the FPN neural network of claim 1, wherein the sixth specific method of the steps is as follows:
firstly, each candidate sample of the candidate sample set is extracted to obtain P after passing through an FPN neural network2Characteristic Z of the layerXPerforming fast Fourier transform to obtain frequency domain characteristicsWhereinFor the kth candidate sample P2Characteristics of the layerObtaining frequency domain characteristics through fast Fourier transform
Then, for each candidate sample in the candidate sample set, the characteristic of fast Fourier transform is performedCorrelating it with fast Fourier transformed correlation filter parametersPerforming dot multiplication and then performing inverse Fourier transform to obtain a response map corresponding to the candidate sampleNamely, it is
WhereinFor the kth candidate sample XkK is the corresponding graph of 1,2 … 7, F-1Is an inverse Fourier transform;
finally, response graphs corresponding to all candidate samples in the candidate sample setFind the point (x) where the response value is maximumt+1,yt+1) The corresponding position of the point is the tracking target position of the frame, and the size of the candidate sample corresponding to the point is the enlarged tracking target area size (l)p,t+1,hp,t+1) Then, the target region size is calculated from the enlargement ratio α
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810329415.1A CN108830878B (en) | 2018-04-13 | 2018-04-13 | Target tracking method based on FPN neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810329415.1A CN108830878B (en) | 2018-04-13 | 2018-04-13 | Target tracking method based on FPN neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108830878A true CN108830878A (en) | 2018-11-16 |
CN108830878B CN108830878B (en) | 2021-02-23 |
Family
ID=64155460
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810329415.1A Active CN108830878B (en) | 2018-04-13 | 2018-04-13 | Target tracking method based on FPN neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108830878B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109740552A (en) * | 2019-01-09 | 2019-05-10 | 上海大学 | A kind of method for tracking target based on Parallel Signature pyramid neural network |
CN109741366A (en) * | 2018-11-27 | 2019-05-10 | 昆明理工大学 | A kind of correlation filtering method for tracking target merging multilayer convolution feature |
CN109767456A (en) * | 2019-01-09 | 2019-05-17 | 上海大学 | A kind of method for tracking target based on SiameseFC frame and PFP neural network |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105741316A (en) * | 2016-01-20 | 2016-07-06 | 西北工业大学 | Robust target tracking method based on deep learning and multi-scale correlation filtering |
US20160342837A1 (en) * | 2015-05-19 | 2016-11-24 | Toyota Motor Engineering & Manufacturing North America, Inc. | Apparatus and method for object tracking |
CN106506999A (en) * | 2016-10-18 | 2017-03-15 | 天津大学 | TDI cmos image sensor FPN bearing calibrations based on match by moment |
CN106887011A (en) * | 2017-01-20 | 2017-06-23 | 北京理工大学 | A kind of multi-template method for tracking target based on CNN and CF |
CN107016689A (en) * | 2017-02-04 | 2017-08-04 | 中国人民解放军理工大学 | A kind of correlation filtering of dimension self-adaption liquidates method for tracking target |
CN107154024A (en) * | 2017-05-19 | 2017-09-12 | 南京理工大学 | Dimension self-adaption method for tracking target based on depth characteristic core correlation filter |
CN107767405A (en) * | 2017-09-29 | 2018-03-06 | 华中科技大学 | A kind of nuclear phase for merging convolutional neural networks closes filtered target tracking |
-
2018
- 2018-04-13 CN CN201810329415.1A patent/CN108830878B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160342837A1 (en) * | 2015-05-19 | 2016-11-24 | Toyota Motor Engineering & Manufacturing North America, Inc. | Apparatus and method for object tracking |
CN105741316A (en) * | 2016-01-20 | 2016-07-06 | 西北工业大学 | Robust target tracking method based on deep learning and multi-scale correlation filtering |
CN106506999A (en) * | 2016-10-18 | 2017-03-15 | 天津大学 | TDI cmos image sensor FPN bearing calibrations based on match by moment |
CN106887011A (en) * | 2017-01-20 | 2017-06-23 | 北京理工大学 | A kind of multi-template method for tracking target based on CNN and CF |
CN107016689A (en) * | 2017-02-04 | 2017-08-04 | 中国人民解放军理工大学 | A kind of correlation filtering of dimension self-adaption liquidates method for tracking target |
CN107154024A (en) * | 2017-05-19 | 2017-09-12 | 南京理工大学 | Dimension self-adaption method for tracking target based on depth characteristic core correlation filter |
CN107767405A (en) * | 2017-09-29 | 2018-03-06 | 华中科技大学 | A kind of nuclear phase for merging convolutional neural networks closes filtered target tracking |
Non-Patent Citations (3)
Title |
---|
MING TANG ET AL.: "Multi-kernel Correlation Filter for Visual Tracking", 《2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION》 * |
TSUNG-YI LIN ET AL.: "Feature Pyramid Networks for Object Detection", 《ARXIV》 * |
王春平 等: "基于金字塔特征的核相关滤波跟踪算法", 《探测与控制学报》 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109741366A (en) * | 2018-11-27 | 2019-05-10 | 昆明理工大学 | A kind of correlation filtering method for tracking target merging multilayer convolution feature |
CN109740552A (en) * | 2019-01-09 | 2019-05-10 | 上海大学 | A kind of method for tracking target based on Parallel Signature pyramid neural network |
CN109767456A (en) * | 2019-01-09 | 2019-05-17 | 上海大学 | A kind of method for tracking target based on SiameseFC frame and PFP neural network |
Also Published As
Publication number | Publication date |
---|---|
CN108830878B (en) | 2021-02-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112184752A (en) | Video target tracking method based on pyramid convolution | |
CN110020606B (en) | Crowd density estimation method based on multi-scale convolutional neural network | |
CN108830157B (en) | Human behavior identification method based on attention mechanism and 3D convolutional neural network | |
CN109271960B (en) | People counting method based on convolutional neural network | |
CN108734151B (en) | Robust long-range target tracking method based on correlation filtering and depth twin network | |
CN109767456A (en) | A kind of method for tracking target based on SiameseFC frame and PFP neural network | |
CN106845374A (en) | Pedestrian detection method and detection means based on deep learning | |
CN109461172A (en) | Manually with the united correlation filtering video adaptive tracking method of depth characteristic | |
CN112163498B (en) | Method for establishing pedestrian re-identification model with foreground guiding and texture focusing functions and application of method | |
CN109191493B (en) | Target tracking method based on RefineNet neural network and sparse optical flow | |
CN108830878B (en) | Target tracking method based on FPN neural network | |
CN104484890A (en) | Video target tracking method based on compound sparse model | |
CN109410249B (en) | Self-adaptive target tracking method combining depth characteristic and hand-drawn characteristic | |
CN108898619B (en) | Target tracking method based on PVANET neural network | |
CN109740552A (en) | A kind of method for tracking target based on Parallel Signature pyramid neural network | |
CN110111369A (en) | A kind of dimension self-adaption sea-surface target tracking based on edge detection | |
CN112785626A (en) | Twin network small target tracking method based on multi-scale feature fusion | |
CN106529441A (en) | Fuzzy boundary fragmentation-based depth motion map human body action recognition method | |
CN112396036A (en) | Method for re-identifying blocked pedestrians by combining space transformation network and multi-scale feature extraction | |
Gao et al. | Background subtraction via 3D convolutional neural networks | |
CN116051601A (en) | Depth space-time associated video target tracking method and system | |
CN113033342A (en) | Crowd scene pedestrian target detection and counting method based on density estimation | |
CN111951298B (en) | Target tracking method integrating time sequence information | |
Zheng et al. | Badminton action recognition based on improved I3D convolutional neural network | |
CN116645718A (en) | Micro-expression recognition method and system based on multi-stream architecture |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |