CN105894008A - Target motion track method through combination of feature point matching and deep nerve network detection - Google Patents

Target motion track method through combination of feature point matching and deep nerve network detection Download PDF

Info

Publication number
CN105894008A
CN105894008A CN201410767363.8A CN201410767363A CN105894008A CN 105894008 A CN105894008 A CN 105894008A CN 201410767363 A CN201410767363 A CN 201410767363A CN 105894008 A CN105894008 A CN 105894008A
Authority
CN
China
Prior art keywords
target
tracking
feature point
point matching
nerve network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410767363.8A
Other languages
Chinese (zh)
Inventor
陈姝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangxi Kasite Cartoon Co Ltd
Original Assignee
Guangxi Kasite Cartoon Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangxi Kasite Cartoon Co Ltd filed Critical Guangxi Kasite Cartoon Co Ltd
Priority to CN201410767363.8A priority Critical patent/CN105894008A/en
Publication of CN105894008A publication Critical patent/CN105894008A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The present invention discloses a target motion track method through combination of feature point matching and deep nerve network detection. The method comprises: firstly, learning the visual prior of a learning target through samples by the deep nerve network, performing tracking under the Bayes inference framework, taking the target visual prior as a target appearance expression in the tracking process, and obtaining the tracking result through the particle filter sequence. In order to avoid tracking drift, a system state model is built by the feature point matching, and the target is divided into subtargets to perform similar measurement to improve the local shielding capability resistance of the algorithm. The target motion track method through combination of feature point matching and deep nerve network detection is able to accurately track the motion target in the video, is widely applicable to the fields of man-machine interaction, interactive entertainment, intelligent monitoring, medical diagnosis and the like.

Description

The target travel tracking detected in conjunction with Feature Points Matching and deep neural network
[technical field]
The present invention relates to computer vision and field of video processing, particularly to moving target based on video with Track method.
[background technology]
Motion target tracking based on video can be widely applied for man-machine interaction, interaction entertainment, intelligent monitoring Etc. multiple fields.At present, conventional motion target tracking method mainly uses discriminant track algorithm or generation Formula track algorithm.This two classes track algorithm is respectively arranged with its pluses and minuses, it determines formula track algorithm is owing to have employed effectively Target visual represent, can obtain preferable tracking effect under simple tracking environmental, but complicated Under background, tracking effect is poor.Production track algorithm tracking effect under the complex environment such as blocking is preferable, but It is that then tracking effect is poor when target appearance drastically changes.
[summary of the invention]
In consideration of it, it is an object of the invention to utilize the advantage of this two classes track algorithm, propose a kind of in complexity Environment carries out the method for target travel accurate tracking.
For reaching above-mentioned purpose, the present invention by the following technical solutions:
1, stack sparse own coding neutral net is trained to obtain the priori vision table of target on public data collection Show.
2, on the basis of the SIFT matching double points extracted, set up object function, and optimize this object function and must transport The motion model of upper and lower two interframe of moving-target.
3, on the basis of above two steps, particle filter is utilized to carry out motion target tracking.
Compared with prior art, the present invention has a following significant advantage:
1, this invention utilizes the priori visual representation of deep neural network learning target, can improve the precision of tracking.
2, this invention utilizes Feature Points Matching and deep neural network detection to be possible to prevent follow the tracks of drift and follow the tracks of loss.
[accompanying drawing explanation]
Fig. 1 is target following flow chart;
Fig. 2 is stack sparse own coding neural network structure figure;
Fig. 3 is characteristic layer 1 sparse own coding structure chart;
Fig. 4 is that the present invention is applied to video tracking result figure;
[detailed description of the invention]
Below in conjunction with the accompanying drawings and the present invention is described in further detail by detailed description of the invention.
Target travel track algorithm proposed by the invention is as it is shown in figure 1, now specifically introduce the realization of each step Details:
1, off-line training.VOC2010 and Caltech101 data set is utilized to carry out stack own coding nerve net Network training obtains the vision priori of target.The stack own coding neural network structure used is as in figure 2 it is shown, be total to Dividing 5 layers, the 1st layer is input layer, and last layer is softmax grader, and middle three layers is sparse own coding Device.Training uses successively greedy coaching method to be undertaken in two steps, each layer of first step training network successively, the Two steps utilize back-propagation algorithm to carry out whole network on the basis of every layer of initial weight that the first step calculates Fine setting.
The first step is trained: sparse own coding device is attempted approaching an identity function, so that outputClose to Input x.Structure (see Fig. 3) and the training process of sparse own coding device are described as a example by characteristic layer 1.
If the i-th sample is xi, W, W ', b, b ' are respectively between input layer and hidden layer and hidden layer and output layer Weight matrix and bias vector.Input layer (xi=[x1 (l), x2 (l)..., x1024 (l), 1]T), hidden layer ( It is the excitation output of hidden layer neuron j under the i-th sample) and defeated Go out layer () there is following relation
Wherein f () is logistic sigmoid function.
The object function setting up sparse own coding device is
Wherein β controls the weight of openness penalty factor, and J (W, W ', b, b ') is cost function, is defined as
For sparse constraint item, wherein m is the quantity of hidden layer neuron, and ρ is openness Parameter, it is common that one close to 0 less value, β be used for control constraints item weight.ForBetween cross entropy.
Sparse own coding device is to solve the parameter making formula (2) minimize, i.e.
W , W ′ , b , b ′ = arg min W , W ′ , b , b ′ ( J sparse ( W , W ′ , b , b ′ ) ) - - - ( 4 )
Second step is finely tuned: during the first step is trained, and during owing to training each layer parameter, can fix other Each layer parameter keeps constant, thus if it is desired to obtain more preferable result, after above-mentioned pre-training process completes, The parameter of all layers can be adjusted to improve result by back-propagation algorithm simultaneously.
In order to improve the ability of algorithm opposing partial occlusion, except one overall goals own coding degree of depth god of training Through network, target is divided into four nonoverlapping sub-goals (top half, the latter half, left-half simultaneously And right half part) be respectively trained an own coding deep neural network, due to size reduce half, thus this four The size of the input layer of the own coding deep neural network that individual sub-goal is corresponding is all 512.Getting well of do so is in Under at partial occlusion, target part region is visible, utilizes partially visible region to carry out target similarity measurement The reliability of algorithm can be improved.
2, target following.The present invention uses particle filter as the basic skills followed the tracks of.Particle filter tracking Precision depends on the reliability of state model and observation model, and the structure of the two model is described below.
2.1 state models build
We are with vector x t = ( x t 1 , y t 1 , x t 2 , y t 2 , x t 3 , y t 3 , x t 4 , y t 4 ) T Represent dbjective state, It is respectively the target upper left corner in the picture, the lower left corner, the upper right corner, the coordinate in the lower right corner. The state model used is as follows
p ( x t | x t - 1 ) = g ( x t - 1 ) + v t - 1 , v t - 1 ~ N ( 0 , Σ ) - - - ( 5 )
Wherein N (0, ∑) is null vector average multivariate Gaussian function, and ∑ is diagonal matrix.g(xt-1) it is motion mould Type, uses six parameter affine transform to be defined as follows
x t = a 1 x t - 1 + a 2 y t - 1 + a 0 y t = a 4 x t - 1 + a 5 y t - 1 + a 3 - - - ( 6 )
Wherein a0, a1, a2, a3, a4, a5For motion model parameters, following methods is used to be calculated.
Being defined error function by above formula is
E ( p ; x ) = x t - a 1 x t - 1 - a 2 y t - 1 - a 0 y t - a 4 x t - 1 - a 5 y t - 1 - a 3 - - - ( 7 )
Wherein p=(a0, a1, a2, a3, a4, a5)T, x=(xt, yt, xt+1, yt+1) it is that 2 coupling SIFT feature points are at image On coordinate.Minimizing following formula can be in the hope of motion model parameters
Gauss-Newton iterative method is used to optimize above formula.
2.2 observation models build
Using the output of deep neural network as the similarity of particle, owing to having 5 deep neural network, Therefore the output of the deep neural network of the similarity combining target of particle and sub-goal, it is defined as follows
p ( z t | x t ) ∝ θ 1 · c t f + θ 2 · max ( c t t , c t b , c t l , c t r ) - - - ( 9 )
Wherein θ1, θ2For coefficient of similarity, meet θ12=1, its value adjusts according to experiment and arranges;For mesh The output of mark deep neural network,It is respectively the output of four sub-target depth neutral nets.
2.3 online target followings
Target to be tracked is marked in first frame by user and obtains, and extracts in object region to be tracked SIFT feature.Extract, in first frame, the sparse own coding of stack that off-line training is obtained by the positive negative sample of target Deep neural network carries out retraining and obtains the vision special table representation model of target.Positive sample set is by selected target Region obtains according to following transformation equation,
x ′ y ′ = cos θ sin θ - sin θ cos θ 1 s s 1 x y - - - ( 10 )
Wherein (x, y)TFor the pixel coordinate in selected target region, (x ', y ')TFor the area pixel after conversion Point coordinates, θ, s are transformation parameter, and θ is translation-angle, and span is θ ∈ [-π/4, π/4], and s is for becoming Changing yardstick, span is s ∈ [0.8,1.2].Randomly choose one group (θ, s) value according to formula (10) by To Template Carrying out converting and train positive sample as one, negative sample collection is then made with the image-region outside selected target region For data source, from these data sources, extract negative sample with different yardsticks and position.
Target on-line tracking flow process is as shown in Figure 1.Two threshold values a are also set in this algorithm1, a2, and a1> a2, wherein a1It is used for judging currently to follow the tracks of result the most reliable, if credibility exceedes this threshold value, then Using tracking result as a new positive sample, and the image-region beyond this region selects negative sample, weight Newly train deep neural network.a2It is used for judging whether particle filter tracking drifts about, if all particles Similarity be below this threshold value, then show that tracker lost efficacy, need scanned whole figure by deep neural network As carrying out target detection, and carry out particle filter according to testing result and reinitialize.
The target following result that this method is applied to video is shown in Fig. 4.

Claims (2)

1. the target travel tracking combining Feature Points Matching and deep neural network detection.Its feature It is to comprise the following steps:
A) carry out stack own coding neural metwork training at VOC2010 and Caltech101 data set and obtain mesh Target vision priori.
B) the SIFT matching double points that extracts and on set up object function, optimize this object function and obtain moving target The motion model of upper and lower two interframe.
C) on the basis of above two steps, particle filter is utilized to carry out motion target tracking.
A kind of combination Feature Points Matching the most according to claim 1 and the target of deep neural network detection Motion tracking method, it is characterised in that: described step b) sets up following error function.
E ( p ; x ) = x t - a 1 x t - 1 - a 2 y t - 1 - a 0 y t - a 4 x t - 1 - a 5 y t - 1 - a 3
CN201410767363.8A 2015-01-16 2015-01-16 Target motion track method through combination of feature point matching and deep nerve network detection Pending CN105894008A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410767363.8A CN105894008A (en) 2015-01-16 2015-01-16 Target motion track method through combination of feature point matching and deep nerve network detection

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410767363.8A CN105894008A (en) 2015-01-16 2015-01-16 Target motion track method through combination of feature point matching and deep nerve network detection

Publications (1)

Publication Number Publication Date
CN105894008A true CN105894008A (en) 2016-08-24

Family

ID=56701463

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410767363.8A Pending CN105894008A (en) 2015-01-16 2015-01-16 Target motion track method through combination of feature point matching and deep nerve network detection

Country Status (1)

Country Link
CN (1) CN105894008A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106651917A (en) * 2016-12-30 2017-05-10 天津大学 Image target tracking algorithm based on neural network
CN107169117A (en) * 2017-05-25 2017-09-15 西安工业大学 A kind of manual draw human motion search method based on autocoder and DTW
CN108257156A (en) * 2018-01-24 2018-07-06 清华大学深圳研究生院 A kind of method of the automatic tracing target object from video
CN108447080A (en) * 2018-03-02 2018-08-24 哈尔滨工业大学深圳研究生院 Method for tracking target, system and storage medium based on individual-layer data association and convolutional neural networks
CN109559329A (en) * 2018-11-28 2019-04-02 陕西师范大学 A kind of particle filter tracking method based on depth denoising autocoder
CN109684953A (en) * 2018-12-13 2019-04-26 北京小龙潜行科技有限公司 The method and device of pig tracking is carried out based on target detection and particle filter algorithm

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106651917A (en) * 2016-12-30 2017-05-10 天津大学 Image target tracking algorithm based on neural network
CN107169117A (en) * 2017-05-25 2017-09-15 西安工业大学 A kind of manual draw human motion search method based on autocoder and DTW
CN107169117B (en) * 2017-05-25 2020-11-10 西安工业大学 Hand-drawn human motion retrieval method based on automatic encoder and DTW
CN108257156A (en) * 2018-01-24 2018-07-06 清华大学深圳研究生院 A kind of method of the automatic tracing target object from video
CN108257156B (en) * 2018-01-24 2021-05-04 清华大学深圳研究生院 Method for automatically tracking target object from video
CN108447080A (en) * 2018-03-02 2018-08-24 哈尔滨工业大学深圳研究生院 Method for tracking target, system and storage medium based on individual-layer data association and convolutional neural networks
CN108447080B (en) * 2018-03-02 2023-05-23 哈尔滨工业大学深圳研究生院 Target tracking method, system and storage medium based on hierarchical data association and convolutional neural network
CN109559329A (en) * 2018-11-28 2019-04-02 陕西师范大学 A kind of particle filter tracking method based on depth denoising autocoder
CN109684953A (en) * 2018-12-13 2019-04-26 北京小龙潜行科技有限公司 The method and device of pig tracking is carried out based on target detection and particle filter algorithm

Similar Documents

Publication Publication Date Title
CN109597087B (en) Point cloud data-based 3D target detection method
CN105894008A (en) Target motion track method through combination of feature point matching and deep nerve network detection
CN105741267B (en) The multi-source image change detecting method of cluster guidance deep neural network classification
CN108549876A (en) The sitting posture detecting method estimated based on target detection and human body attitude
CN107481188A (en) A kind of image super-resolution reconstructing method
CN107330357A (en) Vision SLAM closed loop detection methods based on deep neural network
CN105046717A (en) Robust video object tracking method
CN104281853A (en) Behavior identification method based on 3D convolution neural network
CN104463191A (en) Robot visual processing method based on attention mechanism
Wang et al. Distilling knowledge from an ensemble of convolutional neural networks for seismic fault detection
CN109978882A (en) A kind of medical imaging object detection method based on multi-modal fusion
CN106408030A (en) SAR image classification method based on middle lamella semantic attribute and convolution neural network
Lu et al. Learning transform-aware attentive network for object tracking
Qi et al. FTC-Net: Fusion of transformer and CNN features for infrared small target detection
CN107689052A (en) Visual target tracking method based on multi-model fusion and structuring depth characteristic
CN106778687A (en) Method for viewing points detecting based on local evaluation and global optimization
CN106611420A (en) SAR image segmentation method based on deconvolution network and sketch direction constraint
CN107808376A (en) A kind of detection method of raising one's hand based on deep learning
CN104318246A (en) Depth self-adaption ridgelet network based polarimetric SAR (Synthetic Aperture Radar) image classification
CN114937066A (en) Point cloud registration system and method based on cross offset features and space consistency
CN113158943A (en) Cross-domain infrared target detection method
CN106611423A (en) SAR image segmentation method based on ridge wave filter and deconvolution structural model
CN103985143A (en) Discriminative online target tracking method based on videos in dictionary learning
Shu et al. LVC-Net: Medical image segmentation with noisy label based on local visual cues
CN108320306A (en) Merge the video target tracking method of TLD and KCF

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160824

WD01 Invention patent application deemed withdrawn after publication