CN105894008A - Target motion track method through combination of feature point matching and deep nerve network detection - Google Patents
Target motion track method through combination of feature point matching and deep nerve network detection Download PDFInfo
- Publication number
- CN105894008A CN105894008A CN201410767363.8A CN201410767363A CN105894008A CN 105894008 A CN105894008 A CN 105894008A CN 201410767363 A CN201410767363 A CN 201410767363A CN 105894008 A CN105894008 A CN 105894008A
- Authority
- CN
- China
- Prior art keywords
- target
- tracking
- feature point
- point matching
- nerve network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Image Analysis (AREA)
Abstract
The present invention discloses a target motion track method through combination of feature point matching and deep nerve network detection. The method comprises: firstly, learning the visual prior of a learning target through samples by the deep nerve network, performing tracking under the Bayes inference framework, taking the target visual prior as a target appearance expression in the tracking process, and obtaining the tracking result through the particle filter sequence. In order to avoid tracking drift, a system state model is built by the feature point matching, and the target is divided into subtargets to perform similar measurement to improve the local shielding capability resistance of the algorithm. The target motion track method through combination of feature point matching and deep nerve network detection is able to accurately track the motion target in the video, is widely applicable to the fields of man-machine interaction, interactive entertainment, intelligent monitoring, medical diagnosis and the like.
Description
[technical field]
The present invention relates to computer vision and field of video processing, particularly to moving target based on video with
Track method.
[background technology]
Motion target tracking based on video can be widely applied for man-machine interaction, interaction entertainment, intelligent monitoring
Etc. multiple fields.At present, conventional motion target tracking method mainly uses discriminant track algorithm or generation
Formula track algorithm.This two classes track algorithm is respectively arranged with its pluses and minuses, it determines formula track algorithm is owing to have employed effectively
Target visual represent, can obtain preferable tracking effect under simple tracking environmental, but complicated
Under background, tracking effect is poor.Production track algorithm tracking effect under the complex environment such as blocking is preferable, but
It is that then tracking effect is poor when target appearance drastically changes.
[summary of the invention]
In consideration of it, it is an object of the invention to utilize the advantage of this two classes track algorithm, propose a kind of in complexity
Environment carries out the method for target travel accurate tracking.
For reaching above-mentioned purpose, the present invention by the following technical solutions:
1, stack sparse own coding neutral net is trained to obtain the priori vision table of target on public data collection
Show.
2, on the basis of the SIFT matching double points extracted, set up object function, and optimize this object function and must transport
The motion model of upper and lower two interframe of moving-target.
3, on the basis of above two steps, particle filter is utilized to carry out motion target tracking.
Compared with prior art, the present invention has a following significant advantage:
1, this invention utilizes the priori visual representation of deep neural network learning target, can improve the precision of tracking.
2, this invention utilizes Feature Points Matching and deep neural network detection to be possible to prevent follow the tracks of drift and follow the tracks of loss.
[accompanying drawing explanation]
Fig. 1 is target following flow chart;
Fig. 2 is stack sparse own coding neural network structure figure;
Fig. 3 is characteristic layer 1 sparse own coding structure chart;
Fig. 4 is that the present invention is applied to video tracking result figure;
[detailed description of the invention]
Below in conjunction with the accompanying drawings and the present invention is described in further detail by detailed description of the invention.
Target travel track algorithm proposed by the invention is as it is shown in figure 1, now specifically introduce the realization of each step
Details:
1, off-line training.VOC2010 and Caltech101 data set is utilized to carry out stack own coding nerve net
Network training obtains the vision priori of target.The stack own coding neural network structure used is as in figure 2 it is shown, be total to
Dividing 5 layers, the 1st layer is input layer, and last layer is softmax grader, and middle three layers is sparse own coding
Device.Training uses successively greedy coaching method to be undertaken in two steps, each layer of first step training network successively, the
Two steps utilize back-propagation algorithm to carry out whole network on the basis of every layer of initial weight that the first step calculates
Fine setting.
The first step is trained: sparse own coding device is attempted approaching an identity function, so that outputClose to
Input x.Structure (see Fig. 3) and the training process of sparse own coding device are described as a example by characteristic layer 1.
If the i-th sample is xi, W, W ', b, b ' are respectively between input layer and hidden layer and hidden layer and output layer
Weight matrix and bias vector.Input layer (xi=[x1 (l), x2 (l)..., x1024 (l), 1]T), hidden layer
( It is the excitation output of hidden layer neuron j under the i-th sample) and defeated
Go out layer () there is following relation
Wherein f () is logistic sigmoid function.
The object function setting up sparse own coding device is
Wherein β controls the weight of openness penalty factor, and J (W, W ', b, b ') is cost function, is defined as
For sparse constraint item, wherein m is the quantity of hidden layer neuron, and ρ is openness
Parameter, it is common that one close to 0 less value, β be used for control constraints item weight.ForBetween cross entropy.
Sparse own coding device is to solve the parameter making formula (2) minimize, i.e.
Second step is finely tuned: during the first step is trained, and during owing to training each layer parameter, can fix other
Each layer parameter keeps constant, thus if it is desired to obtain more preferable result, after above-mentioned pre-training process completes,
The parameter of all layers can be adjusted to improve result by back-propagation algorithm simultaneously.
In order to improve the ability of algorithm opposing partial occlusion, except one overall goals own coding degree of depth god of training
Through network, target is divided into four nonoverlapping sub-goals (top half, the latter half, left-half simultaneously
And right half part) be respectively trained an own coding deep neural network, due to size reduce half, thus this four
The size of the input layer of the own coding deep neural network that individual sub-goal is corresponding is all 512.Getting well of do so is in
Under at partial occlusion, target part region is visible, utilizes partially visible region to carry out target similarity measurement
The reliability of algorithm can be improved.
2, target following.The present invention uses particle filter as the basic skills followed the tracks of.Particle filter tracking
Precision depends on the reliability of state model and observation model, and the structure of the two model is described below.
2.1 state models build
We are with vector Represent dbjective state, It is respectively the target upper left corner in the picture, the lower left corner, the upper right corner, the coordinate in the lower right corner.
The state model used is as follows
Wherein N (0, ∑) is null vector average multivariate Gaussian function, and ∑ is diagonal matrix.g(xt-1) it is motion mould
Type, uses six parameter affine transform to be defined as follows
Wherein a0, a1, a2, a3, a4, a5For motion model parameters, following methods is used to be calculated.
Being defined error function by above formula is
Wherein p=(a0, a1, a2, a3, a4, a5)T, x=(xt, yt, xt+1, yt+1) it is that 2 coupling SIFT feature points are at image
On coordinate.Minimizing following formula can be in the hope of motion model parameters
Gauss-Newton iterative method is used to optimize above formula.
2.2 observation models build
Using the output of deep neural network as the similarity of particle, owing to having 5 deep neural network,
Therefore the output of the deep neural network of the similarity combining target of particle and sub-goal, it is defined as follows
Wherein θ1, θ2For coefficient of similarity, meet θ1+θ2=1, its value adjusts according to experiment and arranges;For mesh
The output of mark deep neural network,It is respectively the output of four sub-target depth neutral nets.
2.3 online target followings
Target to be tracked is marked in first frame by user and obtains, and extracts in object region to be tracked
SIFT feature.Extract, in first frame, the sparse own coding of stack that off-line training is obtained by the positive negative sample of target
Deep neural network carries out retraining and obtains the vision special table representation model of target.Positive sample set is by selected target
Region obtains according to following transformation equation,
Wherein (x, y)TFor the pixel coordinate in selected target region, (x ', y ')TFor the area pixel after conversion
Point coordinates, θ, s are transformation parameter, and θ is translation-angle, and span is θ ∈ [-π/4, π/4], and s is for becoming
Changing yardstick, span is s ∈ [0.8,1.2].Randomly choose one group (θ, s) value according to formula (10) by To Template
Carrying out converting and train positive sample as one, negative sample collection is then made with the image-region outside selected target region
For data source, from these data sources, extract negative sample with different yardsticks and position.
Target on-line tracking flow process is as shown in Figure 1.Two threshold values a are also set in this algorithm1, a2, and
a1> a2, wherein a1It is used for judging currently to follow the tracks of result the most reliable, if credibility exceedes this threshold value, then
Using tracking result as a new positive sample, and the image-region beyond this region selects negative sample, weight
Newly train deep neural network.a2It is used for judging whether particle filter tracking drifts about, if all particles
Similarity be below this threshold value, then show that tracker lost efficacy, need scanned whole figure by deep neural network
As carrying out target detection, and carry out particle filter according to testing result and reinitialize.
The target following result that this method is applied to video is shown in Fig. 4.
Claims (2)
1. the target travel tracking combining Feature Points Matching and deep neural network detection.Its feature
It is to comprise the following steps:
A) carry out stack own coding neural metwork training at VOC2010 and Caltech101 data set and obtain mesh
Target vision priori.
B) the SIFT matching double points that extracts and on set up object function, optimize this object function and obtain moving target
The motion model of upper and lower two interframe.
C) on the basis of above two steps, particle filter is utilized to carry out motion target tracking.
A kind of combination Feature Points Matching the most according to claim 1 and the target of deep neural network detection
Motion tracking method, it is characterised in that: described step b) sets up following error function.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410767363.8A CN105894008A (en) | 2015-01-16 | 2015-01-16 | Target motion track method through combination of feature point matching and deep nerve network detection |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410767363.8A CN105894008A (en) | 2015-01-16 | 2015-01-16 | Target motion track method through combination of feature point matching and deep nerve network detection |
Publications (1)
Publication Number | Publication Date |
---|---|
CN105894008A true CN105894008A (en) | 2016-08-24 |
Family
ID=56701463
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410767363.8A Pending CN105894008A (en) | 2015-01-16 | 2015-01-16 | Target motion track method through combination of feature point matching and deep nerve network detection |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105894008A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106651917A (en) * | 2016-12-30 | 2017-05-10 | 天津大学 | Image target tracking algorithm based on neural network |
CN107169117A (en) * | 2017-05-25 | 2017-09-15 | 西安工业大学 | A kind of manual draw human motion search method based on autocoder and DTW |
CN108257156A (en) * | 2018-01-24 | 2018-07-06 | 清华大学深圳研究生院 | A kind of method of the automatic tracing target object from video |
CN108447080A (en) * | 2018-03-02 | 2018-08-24 | 哈尔滨工业大学深圳研究生院 | Method for tracking target, system and storage medium based on individual-layer data association and convolutional neural networks |
CN109559329A (en) * | 2018-11-28 | 2019-04-02 | 陕西师范大学 | A kind of particle filter tracking method based on depth denoising autocoder |
CN109684953A (en) * | 2018-12-13 | 2019-04-26 | 北京小龙潜行科技有限公司 | The method and device of pig tracking is carried out based on target detection and particle filter algorithm |
-
2015
- 2015-01-16 CN CN201410767363.8A patent/CN105894008A/en active Pending
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106651917A (en) * | 2016-12-30 | 2017-05-10 | 天津大学 | Image target tracking algorithm based on neural network |
CN107169117A (en) * | 2017-05-25 | 2017-09-15 | 西安工业大学 | A kind of manual draw human motion search method based on autocoder and DTW |
CN107169117B (en) * | 2017-05-25 | 2020-11-10 | 西安工业大学 | Hand-drawn human motion retrieval method based on automatic encoder and DTW |
CN108257156A (en) * | 2018-01-24 | 2018-07-06 | 清华大学深圳研究生院 | A kind of method of the automatic tracing target object from video |
CN108257156B (en) * | 2018-01-24 | 2021-05-04 | 清华大学深圳研究生院 | Method for automatically tracking target object from video |
CN108447080A (en) * | 2018-03-02 | 2018-08-24 | 哈尔滨工业大学深圳研究生院 | Method for tracking target, system and storage medium based on individual-layer data association and convolutional neural networks |
CN108447080B (en) * | 2018-03-02 | 2023-05-23 | 哈尔滨工业大学深圳研究生院 | Target tracking method, system and storage medium based on hierarchical data association and convolutional neural network |
CN109559329A (en) * | 2018-11-28 | 2019-04-02 | 陕西师范大学 | A kind of particle filter tracking method based on depth denoising autocoder |
CN109684953A (en) * | 2018-12-13 | 2019-04-26 | 北京小龙潜行科技有限公司 | The method and device of pig tracking is carried out based on target detection and particle filter algorithm |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109597087B (en) | Point cloud data-based 3D target detection method | |
CN105894008A (en) | Target motion track method through combination of feature point matching and deep nerve network detection | |
CN105741267B (en) | The multi-source image change detecting method of cluster guidance deep neural network classification | |
CN108549876A (en) | The sitting posture detecting method estimated based on target detection and human body attitude | |
CN107481188A (en) | A kind of image super-resolution reconstructing method | |
CN107330357A (en) | Vision SLAM closed loop detection methods based on deep neural network | |
CN105046717A (en) | Robust video object tracking method | |
CN104281853A (en) | Behavior identification method based on 3D convolution neural network | |
CN104463191A (en) | Robot visual processing method based on attention mechanism | |
Wang et al. | Distilling knowledge from an ensemble of convolutional neural networks for seismic fault detection | |
CN109978882A (en) | A kind of medical imaging object detection method based on multi-modal fusion | |
CN106408030A (en) | SAR image classification method based on middle lamella semantic attribute and convolution neural network | |
Lu et al. | Learning transform-aware attentive network for object tracking | |
Qi et al. | FTC-Net: Fusion of transformer and CNN features for infrared small target detection | |
CN107689052A (en) | Visual target tracking method based on multi-model fusion and structuring depth characteristic | |
CN106778687A (en) | Method for viewing points detecting based on local evaluation and global optimization | |
CN106611420A (en) | SAR image segmentation method based on deconvolution network and sketch direction constraint | |
CN107808376A (en) | A kind of detection method of raising one's hand based on deep learning | |
CN104318246A (en) | Depth self-adaption ridgelet network based polarimetric SAR (Synthetic Aperture Radar) image classification | |
CN114937066A (en) | Point cloud registration system and method based on cross offset features and space consistency | |
CN113158943A (en) | Cross-domain infrared target detection method | |
CN106611423A (en) | SAR image segmentation method based on ridge wave filter and deconvolution structural model | |
CN103985143A (en) | Discriminative online target tracking method based on videos in dictionary learning | |
Shu et al. | LVC-Net: Medical image segmentation with noisy label based on local visual cues | |
CN108320306A (en) | Merge the video target tracking method of TLD and KCF |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20160824 |
|
WD01 | Invention patent application deemed withdrawn after publication |