CN109559329A - A kind of particle filter tracking method based on depth denoising autocoder - Google Patents

A kind of particle filter tracking method based on depth denoising autocoder Download PDF

Info

Publication number
CN109559329A
CN109559329A CN201811433093.1A CN201811433093A CN109559329A CN 109559329 A CN109559329 A CN 109559329A CN 201811433093 A CN201811433093 A CN 201811433093A CN 109559329 A CN109559329 A CN 109559329A
Authority
CN
China
Prior art keywords
encoder
particle
weight
training
automatic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811433093.1A
Other languages
Chinese (zh)
Other versions
CN109559329B (en
Inventor
李良福
宋睿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shaanxi Normal University
Original Assignee
Shaanxi Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shaanxi Normal University filed Critical Shaanxi Normal University
Priority to CN201811433093.1A priority Critical patent/CN109559329B/en
Publication of CN109559329A publication Critical patent/CN109559329A/en
Application granted granted Critical
Publication of CN109559329B publication Critical patent/CN109559329B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to computer vision analysis technical fields, more particularly to a kind of particle filter tracking method based on depth denoising autocoder, including initial phase, using the manual spotting position of video first frame, it is tracked in process and tracking processing in first frame, it needs respectively in target background and prospect a certain number of positive negative samples selected around, initialize trained network model, second step, carry out importance sampling, third step, calculate observation probability, 4th step, update weight, 5th step, judge and selects the maximum particle of weight in weight information, think the particle be we next the target to be tracked, tracking new samples are updated for next frame, process of the circulation second step to the 5th step, until video playing finishes, this method can efficiently differentiate target signature and background, improve track algorithm Precision.

Description

Particle filter tracking method based on depth denoising automatic encoder
Technical Field
The invention belongs to the technical field of computer vision analysis, and particularly relates to a particle filter tracking method based on a depth denoising automatic encoder.
Background
Visual target tracking is an important research direction for computer vision and visual analysis. Typical visual analysis requires consistent and stable tracking of the object of interest. For monocular visual target tracking, numerous scholars propose theories and algorithms worthy of reference. In practical applications, the problem still faces a huge challenge due to the influence of factors such as complex background, target occlusion, rapid target motion, illumination change and the like.
The deep neural network has strong learning ability in the aspects of target detection and target classification. The deep learning architecture is more suitable for learning classification features than specific targets. In addition, the deep neural network algorithm usually requires a long iterative training process to converge, and the real-time requirement of online learning is difficult to meet. Therefore, it is difficult to extend the current deep learning network architecture to the field of target tracking.
Disclosure of Invention
The invention provides a particle filter tracking method based on a depth denoising automatic encoder, aiming at solving the problems of interference such as complex background, light change, target shielding and the like in the target tracking process and the problem of poor anti-jamming capability of the existing tracking algorithm. The technical problem to be solved by the invention is realized by the following technical scheme:
a particle filter tracking method based on a depth denoising automatic encoder comprises the following steps: step 1: training a deep network model, carrying out unsupervised greedy training on each layer of network layer by layer, adding noise into training data to obtain more stable feature expression, and carrying out supervised learning on the features through a classification neural network to further optimize parameters of the network;
step 2: manually calibrating the target position by adopting a first frame of a video, selecting positive and negative training samples from the sequence, and initializing the deep network model in the step 1;
and step 3: sampling particle sets by adopting importance, then propagating each particle forwards through a trained network model, and calculating the confidence coefficient of each particle in the online tracking process through a classification neural network;
and 4, step 4: calculating the observation probability of the particles according to the confidence coefficient of the particles in the step 3;
and 5: and (5) updating the weight of the particles according to the observation probability in the step (4) to determine the target position, updating and tracking a new sample for the next frame, and circulating the processes from the step (3) to the step (5) until the video is played.
Further, the depth network model in the step 1 is overlapped by an automatic noise reduction coder, and the output of the next layer is used as the input of the upper layer; the automatic noise reduction encoder comprises an encoder, a decoder and an implicit layer, wherein the decoder needs to predict original undamaged data according to noise characteristics and finally outputs the closest original input, Gaussian noise is usually used as an attenuation vector, and the expression of the Gaussian noise is as follows:
where x is the original input without noise interference,is data contaminated by noise, and σ represents the degree of regularization of the auto-encoder.
Further, the training process in step 1 is as follows: assume a training sample set x ∈ R for unlabeled classesdMapping the input x to the hidden layer by the activation function f to obtain z ∈ Rd
z∈fθ(x)=σ(Wx+b) (1)
Where θ is { W, b }, W is a weight matrix, b is a coding layer vector, σ is a non-linear activation function, and the decoder remaps the input coded representation to form reconstructed y
y∈fθ′(h)=σ(W′h+b′) (2)
Where θ '{ W', b '}, W' is the transpose of the weight matrix W, and σ is the activation function of decoding; the automatic noise reduction encoder enables y to be approximately equal to x through the process;
hypothesis training set { (x)(1),y(1)),...,(x(m),y(m)) Contains m training samples, x represents a single sample feature, y represents the input to which the sample corresponds, and a single sample (x, y) is taken to define its cost function;
wherein h isW,b(x) Is corresponding to a networkThe output value of sample x, and thus the cost function of the training set of m samples, is:
λ is the weight loss coefficient, controlling the relative importance of the two parts; the process of training an automatic noise reduction encoder is to adjust the minimum reconstruction error J (W, b) of the parameters { theta, theta' } in the training sample set, where J (W, b) is a convex function, usually optimized by an iterative method.
Further, the classification neural network comprises an automatic noise reduction encoder coding part and k classification layers connected with sparse constraints.
Further, the classification neural network learning method in the step 1 is as follows, and Z is an activation function of a hidden layer of the self-encoder. In the forward propagation phase, the activation function Z is:
z=WTx+b (6)
wherein x is an input vector; w is a weight; b is a bias (bias).
The first K maxima of the activation function are maintained and all others are set to zero:
wherein, (gamma)cIs complementary to z, (Γ)c=sup pk(z). The sparsity z is used to calculate the network reconstruction error:
where x is the training sample set, W represents the weight, b' represents the transpose of the bias (bias), the weight is back-propagated by the first K maxima of the activation function output to reconstruct the error iteration adjustment.
Further, the confidence coefficient algorithm in step 3 is as follows: let oiIs corresponding to class kiThe output of the neural network, the expectation of the output value is the posterior probability.
E{oi}=P(ki|x) (9)
Where x is the network input. Typically, the corresponding class of the largest output is taken as the decision, so the confidence can be obtained from the posterior probability of the neural network, and the largest output of the classified neural network is taken as the confidence:
c(x)=E{max oi} (10)
further, the importance sampling method in step 3 is as follows:
when a new frame image arrives, q(s) is distributed according to the importancet|st-1,y1:t) And a motion model, from the set of particles at time t-1Obtaining n particles at time tWherein the importance weight corresponding to the set of particlesSum of (S)tIs 1; target state stS is represented by six affine parameters, horizontal translation, vertical translation, scaling, width/height ratio, rotation and skewt(tx, ty, sxy, ra, ar, sa); distribution per dimension of state transitionsIn the motion model, the model is an independent zero-mean normal distribution model.
Further, the observation probability calculation method in step 4 is as follows:
each particle is propagated forward through the classification neural network to obtain its confidenceAnd will maximize confidenceComparing with the set threshold value tau ifReselecting the positive and negative training samples, and initializing a classification neural network; if it is notThe probability of observation of the particle is calculated as follows:
wherein y istIt means that the sample at time t corresponds to the input,refers to the ith particle at time t.
Further, the method for updating the weight of the particle in step 5 includes:
wherein,each dimension representing a state transition in the importance sample is distributed,i.e. the calculated probability distribution of the particles, the distribution q(s) in generalt|st-1,y1:t) Using a first order Markov process q(s)t|st-1) I.e. the state transition is independent of the model observations, then the weights are updated as:
whereinIndicating the weight at the time immediately before the update,representing the observation probability of the particles obtained by the previous step, wherein for each frame, the particle with the maximum weight is a tracking result; updating a positive sample for each tracking frame, and then tracking the next positive sample; determining a state corresponding to the largest particle as a frame target position outside the current vehicle
The invention has the beneficial effects that:
the automatic noise reduction encoder obtains distributed characteristic representation of high-dimensional complex input through unsupervised greedy training layer by layer and parameter optimization multi-layer network structure, and only needs to adjust network parameters for different tasks; the method can effectively distinguish the target characteristics from the background through the automatic noise reduction encoder for deep noise reduction; and a classification neural network is introduced, so that the classification capability of the network is improved, the precision of a tracking algorithm is improved, and finally, particle filtering is used for tracking the target.
Drawings
FIG. 1 is a schematic diagram of an automatic noise reduction encoder.
Fig. 2 is a schematic diagram of a classification neural network structure.
Fig. 3 is a diagram illustrating the tracking result of the indoor occlusion phenomenon.
Fig. 4 is a diagram illustrating the tracking result of the outdoor occlusion phenomenon.
Fig. 5 is a diagram illustrating the tracking result of the illumination change target.
Fig. 6 is a schematic diagram of a target fuzzy target tracking result.
Detailed Description
The present invention will be described in further detail with reference to specific examples, but the embodiments of the present invention are not limited thereto.
A particle filter tracking method based on a depth denoising automatic encoder comprises the following steps:
step 1: training a deep network model, carrying out unsupervised greedy training on each layer of network layer by layer, adding noise into training data to obtain more stable feature expression, and carrying out supervised learning on the features through a classification neural network to further optimize parameters of the network;
step 2: manually calibrating the target position by adopting a first frame of a video, selecting positive and negative training samples from the sequence, and initializing the deep network model in the step 1;
and step 3: sampling particle sets by adopting importance, then propagating each particle forwards through a trained network model, and calculating the confidence coefficient of each particle in the online tracking process through a classification neural network;
and 4, step 4: calculating the observation probability of the particles according to the confidence coefficient of the particles in the step 3;
and 5: and (5) updating the weight of the particles according to the observation probability in the step (4) to determine the target position, updating and tracking a new sample for the next frame, and circulating the processes from the step (3) to the step (5) until the video is played.
As shown in fig. 1, in step 1, the depth network model is overlapped by an automatic noise reduction encoder, the depth automatic encoder is a typical unsupervised learning network, which is a depth network model, overlapped by a self-encoder, and uses the output of the next layer as the input of the upper layer, the essence of the automatic encoder is to learn the same function, i.e. the input of the network is equal to the reconstructed output, and the training and parameter optimization process is to realize output and reproduction input; the automatic noise reduction encoder comprises an encoder, a decoder and an implicit layer; the auto noise reduction encoder accepts the corrupted data as input and predicts the original uncorrupted data as output by training. The purpose of the noise reduction automatic encoder is to allow the use of a very large encoder and simultaneously prevent useless constant functions between the encoder and the decoder, and based on a statistical theory, the core idea of the automatic noise reduction encoder is to disturb original input and noise according to a certain rule, so that the original input is damaged, damaged data is input into a network, and the representation of a hidden layer is obtained. The decoder needs to predict the original uncorrupted data according to the noise characteristics, and finally outputs the closest original input, which is the effect of removing interference, and gaussian noise is usually used as an attenuation vector, and its expression is:
where x is the original input without noise interference,which is noise contaminated data, and sigma represents the degree of regularization of the auto-encoder, for the problem of generation of impairment data, the binomial random numbers are not only simple but also easy to calculate, i.e., binomial distribution random numbers are generated using the same input shape and then multiplied by the input. We use the squared error function as the reconstruction error and train it in exactly the same way as other feed forward networks.
The training process in step 1 is as follows: assume a training sample set x ∈ R for unlabeled classesdMapping the input x to the hidden layer by the activation function f to obtain z ∈ Rd
z∈fθ(x)=σ(Wx+b) (1)
Where θ is { W, b }, W is a weight matrix, b is a coding layer vector, σ is a non-linear activation function, and the decoder remaps the input coded representation to form reconstructed y
y∈fθ′(h)=σ(W′h+b′) (2)
Where θ '{ W', b '}, W' is the transpose of the weight matrix W, and σ is the activation function of decoding; the automatic noise reduction encoder enables y to be approximately equal to x through the process;
hypothesis training set { (x)(1),y(1),...,(x(m),y(m)) Contains m training samples, x represents a single sample feature, y represents the input to which the sample corresponds, and a single sample (x, y) is taken to define its cost function;
wherein h isW,b(x) Is the output value of sample x corresponding to the network, so the cost function of the training set of m samples is:
it can be seen that the first part of the equation is the mean variance term of the cost function. The second part is a weight attenuation term which can prevent the weight from changing too much so as to prevent overfitting, and lambda is a weight reduction coefficient which controls the relative importance of the two parts; the process of training an automatic noise reduction encoder is to adjust the minimum reconstruction error J (W, b) of the parameters { theta, theta' } in the training sample set, where J (W, b) is a convex function, usually optimized by an iterative method.
The classification neural network comprises an automatic noise reduction encoder coding part and classification layers connected with k sparse constraints.
The purpose of constructing a classification neural network is to calculate the confidence of each particle in the online tracking process. The automatic noise reduction encoder consists of an encoding part of the automatic noise reduction encoder and k sparse constraint connected classification layers, and a schematic diagram of a classification neural network structure is shown in FIG. 2; the invariant characteristic of the target can be effectively learned by introducing k sparse constraint, the linear discrimination capability of the classification neural network is improved, and the overfitting problem is solved to a certain extent. Neuroscience research shows that the response of visual signals in cerebral cortex is sparse, so that the expression of original signals can be more meaningful by introducing sparse restriction in a deep neural network, particularly for a classification task, the idea is verified in principal component analysis and sparse coding, K sparse restriction keeps K maximum activation functions of hidden layers, and the rest are set to be zero.
The classified neural network learning method in the step 1 is as follows, and Z is an activation function of a hidden layer of a self-encoder. In the forward propagation phase, the activation function Z is:
z=WTx+b (6)
wherein x is an input vector; w is a weight; b is a bias (bias).
The first K maxima of the activation function are maintained and all others are set to zero:
wherein, (gamma)cIs complementary to z, (Γ)c=sup pk(z). The sparsity z is used to calculate the network reconstruction error:
where x is the training sample set, W represents the weight, b' represents the transpose of the bias (bias), the weight is back-propagated by the first K maxima of the activation function output to reconstruct the error iteration adjustment. The confidence level of the classification neural network output is the confidence level, which reflects its decision confidence at some point in the feature vector space.
The confidence coefficient algorithm in the step 3 is as follows: let oiIs corresponding to class kiThe output of the neural network, the expectation of the output value is the posterior probability.
E{oi}=P(ki|x) (9)
Where x is the network input. Typically, the corresponding class of the largest output is taken as the decision, so the confidence can be obtained from the posterior probability of the neural network, and the largest output of the classified neural network is taken as the confidence:
c(x)=E{max oi} (10)
the importance sampling method in step 3 is as follows:
when a new frame image arrives, q(s) is distributed according to the importancet|st-1,y1:t) And a motion model, from the set of particles at time t-1Obtaining n particles at time tWherein the importance weight corresponding to the set of particlesSum of (S)tIs 1; target state stS is represented by six affine parameters, horizontal translation, vertical translation, scaling, width/height ratio, rotation and skewt(tx, ty, sxy, ra, ar, sa); distribution per dimension of state transitionsIs an independent zero-mean normal score in the motion modelAnd (4) cloth modeling.
The observation probability calculation method in step 4 is as follows:
each particle is propagated forward through the classification neural network to obtain its confidenceAnd will maximize confidenceComparing with the set threshold value tau ifReselecting the positive and negative training samples, and initializing a classification neural network; if it is notThe probability of observation of the particle is calculated as follows:
wherein y istIt means that the sample at time t corresponds to the input,refers to the ith particle at time t.
The method for updating the weight of the particles in the step 5 comprises the following steps:
wherein,each dimension representing a state transition in the importance sample is distributed,i.e. the calculated probability distribution of the particles, the distribution q(s) in generalt|st-1,y1:t) Using a first order Markov process q(s)t|st-1) I.e. the state transition is independent of the model observations, then the weights are updated as:
whereinIndicating the weight at the time immediately before the update,representing the observation probability of the particles obtained by the previous step, wherein for each frame, the particle with the maximum weight is a tracking result; updating a positive sample for each tracking frame, and then tracking the next positive sample; the state corresponding to the largest particle is determined as the frame target position outside the current vehicle.
The experimental operating environment is as follows: 3.8GHz, quad-core AMD processor, 8GB memory. Video sequences in various environments are used for verification, including illumination change, target occlusion and target rapid motion.
Fig. 3 and 4 show occlusion of an object. The shielding phenomenon is the phenomenon that a target is partially shielded due to the complexity of the surrounding environment and the interference of other surrounding objects, and the target cannot be lost in the whole tracking process of the tracker; outdoor photography often produces strong light changes. When the light changes greatly, the performance of target tracking is affected, and as can be seen from fig. 5, when a target enters a tunnel, huge illumination changes exist in an image, but from the tracking result, the algorithm accurately completes the tracking task.
The problem of target blurring appears in fig. 6, the target blurring is caused by too fast speed of the target in the moving process or unstable photography, the image of the target in the image is not clear, the target is blurred, the tracking effect is affected, the tracking algorithm in the text accurately completes tracking, and no target is lost.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several simple deductions or substitutions can be made without departing from the spirit of the invention, and all shall be considered as belonging to the protection scope of the invention.

Claims (9)

1. A particle filter tracking method based on a depth denoising automatic encoder is characterized in that:
step 1: training a deep network model, carrying out unsupervised greedy training on each layer of network layer by layer, adding noise into training data to obtain more stable feature expression, and carrying out supervised learning on the features through a classification neural network to further optimize parameters of the network;
step 2: manually calibrating the target position by adopting a first frame of a video, selecting positive and negative training samples from the sequence, and initializing the deep network model in the step 1;
and step 3: sampling particle sets by adopting importance, then propagating each particle forwards through a trained network model, and calculating the confidence coefficient of each particle in the online tracking process through a classification neural network;
and 4, step 4: calculating the observation probability of the particles according to the confidence coefficient of the particles in the step 3;
and 5: and (5) updating the weight of the particles according to the observation probability in the step (4) to determine the target position, updating and tracking a new sample for the next frame, and circulating the processes from the step (3) to the step (5) until the video is played.
2. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: in the step 1, the deep network models are overlapped by an automatic noise reduction coder, and the output of the next layer is used as the input of the upper layer; the automatic noise reduction encoder comprises an encoder, a decoder and an implicit layer, wherein the decoder needs to predict original undamaged data according to noise characteristics and finally outputs the closest original input, Gaussian noise is usually used as an attenuation vector, and the expression of the Gaussian noise is as follows:
where x is the original input without noise interference,is data contaminated by noise, and σ represents the degree of regularization of the auto-encoder.
3. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the training process in the step 1 is as follows: assume a training sample set x ∈ R for unlabeled classesdMapping the input x to the hidden layer by the activation function f to obtain z ∈ Rd
z∈fθ(x)=σ(Wx+b) (1)
Where θ is { W, b }, W is a weight matrix, b is a coding layer vector, σ is a non-linear activation function, and the decoder remaps the input coded representation to form reconstructed y
y∈fθ′(h)=σ(W′h+b′) (2)
Where θ '{ W', b '}, W' is the transpose of the weight matrix W, and σ is the activation function of decoding; the automatic noise reduction encoder enables y to be approximately equal to x through the process;
hypothesis training set { (x)(1),y(1)),...,(x(m),y(m)) Contains m training samples, x represents a single sample feature, y represents the input to which the sample corresponds, and a single sample (x, y) is taken to define its cost function;
wherein h isW,b(x) Is the output value of sample x corresponding to the network, so the cost function of the training set of m samples is:
λ is the weight loss coefficient, controlling the relative importance of the two parts; the process of training an automatic noise reduction encoder is to adjust the minimum reconstruction error J (W, b) of the parameters { theta, theta' } in the training sample set, where J (W, b) is a convex function, usually optimized by an iterative method.
4. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the classification neural network comprises an automatic noise reduction encoder coding part and k classification layers connected with sparse constraints.
5. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the classified neural network learning method in the step 1 is as follows, and Z is an activation function of a hidden layer of a self-encoder. In the forward propagation phase, the activation function Z is:
z=WTx+b (6)
wherein x is an input vector; w is a weight; b is a bias (bias).
The first K maxima of the activation function are maintained and all others are set to zero:
wherein, (gamma)cIs complementary to z, (Γ)c=sup pk(z). The sparsity z is used to calculate the network reconstruction error:
where x is the training sample set, W represents the weight, b' represents the transpose of the bias (bias), the weight is back-propagated by the first K maxima of the activation function output to reconstruct the error iteration adjustment.
6. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the confidence coefficient algorithm in the step 3 is as follows: let oiIs corresponding to class kiThe output of the neural network, the expectation of the output value is the posterior probability.
E{oi}=P(ki|x) (9)
Where x is the network input. Typically, the corresponding class of the largest output is taken as the decision, so the confidence can be obtained from the posterior probability of the neural network, and the largest output of the classified neural network is taken as the confidence:
c(x)=E{maxoi} (10)
7. the particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the importance sampling method in the step 3 is as follows:
when a new frame image arrives, q(s) is distributed according to the importancet|st-1,y1∶t) And a motion model, from the set of particles at time t-1Obtaining n particles at time tWherein the importance weight corresponding to the set of particlesSum of (S)tIs 1; target state StS is represented by six affine parameters, horizontal translation, vertical translation, scaling, width/height ratio, rotation and skewt(tx, ty, sxy, ra, ar, sa); distribution per dimension of state transitionsIn the motion model, the model is an independent zero-mean normal distribution model.
8. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the observation probability calculation method in the step 4 is as follows:
each particle is propagated forward through the classification neural network to obtain its confidenceAnd will maximize confidenceComparing with the set threshold value tau ifReselecting the positive and negative training samples, and initializing a classification neural network; if it is notThe probability of observation of the particle is calculated as follows:
wherein y istIt means that the sample at time t corresponds to the input,refers to the ith particle at time t.
9. The particle filter tracking method based on the automatic deep denoising encoder as claimed in claim 1, wherein: the method for updating the weight of the particles in the step 5 comprises the following steps:
wherein,each dimension representing a state transition in the importance sample is distributed,i.e. the calculated probability distribution of the particles, the distribution q(s) in generalt|st-1,y1:t) Using a first order Markov process q(s)t|st-1) I.e. the state transition is independent of the model observations, then the weights are updated as:
whereinIndicating the weight at the time immediately before the update,representing the observation probability of the particles obtained by the previous step, wherein for each frame, the particle with the maximum weight is a tracking result; updating a positive sample for each tracking frame, and then tracking the next positive sample; the state corresponding to the largest particle is determined as the frame target position outside the current vehicle.
CN201811433093.1A 2018-11-28 2018-11-28 Particle filter tracking method based on depth denoising automatic encoder Active CN109559329B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811433093.1A CN109559329B (en) 2018-11-28 2018-11-28 Particle filter tracking method based on depth denoising automatic encoder

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811433093.1A CN109559329B (en) 2018-11-28 2018-11-28 Particle filter tracking method based on depth denoising automatic encoder

Publications (2)

Publication Number Publication Date
CN109559329A true CN109559329A (en) 2019-04-02
CN109559329B CN109559329B (en) 2023-04-07

Family

ID=65867657

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811433093.1A Active CN109559329B (en) 2018-11-28 2018-11-28 Particle filter tracking method based on depth denoising automatic encoder

Country Status (1)

Country Link
CN (1) CN109559329B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110473557A (en) * 2019-08-22 2019-11-19 杭州派尼澳电子科技有限公司 A kind of voice signal decoding method based on depth self-encoding encoder
CN110825123A (en) * 2019-10-21 2020-02-21 哈尔滨理工大学 Control system and method for automatic following loading vehicle based on motion algorithm
CN110889459A (en) * 2019-12-06 2020-03-17 北京深境智能科技有限公司 Learning method based on edge and Fisher criterion
CN111552322A (en) * 2020-04-29 2020-08-18 东南大学 Unmanned aerial vehicle tracking method based on LSTM-particle filter coupling model
CN111563423A (en) * 2020-04-17 2020-08-21 西北工业大学 Unmanned aerial vehicle image target detection method and system based on depth denoising automatic encoder
CN111735458A (en) * 2020-08-04 2020-10-02 西南石油大学 Navigation and positioning method of petrochemical inspection robot based on GPS, 5G and vision
CN111931368A (en) * 2020-08-03 2020-11-13 哈尔滨工程大学 UUV target state estimation method based on GRU particle filter
CN111950503A (en) * 2020-06-16 2020-11-17 中国科学院地质与地球物理研究所 Aviation transient electromagnetic data processing method and device and computing equipment
CN112396635A (en) * 2020-11-30 2021-02-23 深圳职业技术学院 Multi-target detection method based on multiple devices in complex environment

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002171436A (en) * 2000-12-04 2002-06-14 Sony Corp Image processing apparatus
CN103211563A (en) * 2005-05-16 2013-07-24 直观外科手术操作公司 Methods and system for performing 3-D tool tracking by fusion of sensor and/or camera derived data during minimally invasive robotic surgery
CN105654509A (en) * 2015-12-25 2016-06-08 燕山大学 Motion tracking method based on composite deep neural network
CN105894008A (en) * 2015-01-16 2016-08-24 广西卡斯特动漫有限公司 Target motion track method through combination of feature point matching and deep nerve network detection
CN106127804A (en) * 2016-06-17 2016-11-16 淮阴工学院 The method for tracking target of RGB D data cross-module formula feature learning based on sparse depth denoising own coding device
CN106203350A (en) * 2016-07-12 2016-12-07 北京邮电大学 A kind of moving target is across yardstick tracking and device
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check
CN108460760A (en) * 2018-03-06 2018-08-28 陕西师范大学 A kind of Bridge Crack image discriminating restorative procedure fighting network based on production

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002171436A (en) * 2000-12-04 2002-06-14 Sony Corp Image processing apparatus
CN103211563A (en) * 2005-05-16 2013-07-24 直观外科手术操作公司 Methods and system for performing 3-D tool tracking by fusion of sensor and/or camera derived data during minimally invasive robotic surgery
CN105894008A (en) * 2015-01-16 2016-08-24 广西卡斯特动漫有限公司 Target motion track method through combination of feature point matching and deep nerve network detection
CN105654509A (en) * 2015-12-25 2016-06-08 燕山大学 Motion tracking method based on composite deep neural network
CN106127804A (en) * 2016-06-17 2016-11-16 淮阴工学院 The method for tracking target of RGB D data cross-module formula feature learning based on sparse depth denoising own coding device
CN106203350A (en) * 2016-07-12 2016-12-07 北京邮电大学 A kind of moving target is across yardstick tracking and device
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check
CN108460760A (en) * 2018-03-06 2018-08-28 陕西师范大学 A kind of Bridge Crack image discriminating restorative procedure fighting network based on production

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
季顺平 等: "基于HSV颜色特征和贡献度重构的行人跟踪", 《激光与光电子学进展》 *
杨红红等: "基于稀疏约束深度学习的交通目标跟踪", 《中国公路学报》 *
焦婷 等: "存在运动目标时的图像镶嵌方法研究", 《计算机应用研究》 *
程帅 等: "多示例深度学习目标跟踪", 《电子与信息学报》 *
邓俊锋: "基于稀疏自动编码器和边缘降噪自动编码器的深度学习算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110473557A (en) * 2019-08-22 2019-11-19 杭州派尼澳电子科技有限公司 A kind of voice signal decoding method based on depth self-encoding encoder
CN110473557B (en) * 2019-08-22 2021-05-28 浙江树人学院(浙江树人大学) Speech signal coding and decoding method based on depth self-encoder
CN110825123A (en) * 2019-10-21 2020-02-21 哈尔滨理工大学 Control system and method for automatic following loading vehicle based on motion algorithm
CN110889459A (en) * 2019-12-06 2020-03-17 北京深境智能科技有限公司 Learning method based on edge and Fisher criterion
CN110889459B (en) * 2019-12-06 2023-04-28 北京深境智能科技有限公司 Learning method based on edge and Fisher criteria
CN111563423A (en) * 2020-04-17 2020-08-21 西北工业大学 Unmanned aerial vehicle image target detection method and system based on depth denoising automatic encoder
CN111552322A (en) * 2020-04-29 2020-08-18 东南大学 Unmanned aerial vehicle tracking method based on LSTM-particle filter coupling model
CN111950503A (en) * 2020-06-16 2020-11-17 中国科学院地质与地球物理研究所 Aviation transient electromagnetic data processing method and device and computing equipment
CN111950503B (en) * 2020-06-16 2024-01-30 中国科学院地质与地球物理研究所 Aviation transient electromagnetic data processing method and device and computing equipment
CN111931368A (en) * 2020-08-03 2020-11-13 哈尔滨工程大学 UUV target state estimation method based on GRU particle filter
CN111735458A (en) * 2020-08-04 2020-10-02 西南石油大学 Navigation and positioning method of petrochemical inspection robot based on GPS, 5G and vision
CN112396635A (en) * 2020-11-30 2021-02-23 深圳职业技术学院 Multi-target detection method based on multiple devices in complex environment

Also Published As

Publication number Publication date
CN109559329B (en) 2023-04-07

Similar Documents

Publication Publication Date Title
CN109559329B (en) Particle filter tracking method based on depth denoising automatic encoder
CN108133188B (en) Behavior identification method based on motion history image and convolutional neural network
Mane et al. Moving object detection and tracking using convolutional neural networks
Ma et al. Variational implicit processes
CN106127804B (en) The method for tracking target of RGB-D data cross-module formula feature learnings based on sparse depth denoising self-encoding encoder
US20220019805A1 (en) Video watermark identification method and apparatus, device, and storage medium
Jena et al. Augmenting gail with bc for sample efficient imitation learning
Civicioglu Using uncorrupted neighborhoods of the pixels for impulsive noise suppression with ANFIS
CN110047096B (en) A kind of multi-object tracking method and system based on depth conditions random field models
CN113793359B (en) Target tracking method integrating twin network and related filtering
CN114861838B (en) Intelligent classification method for pulsatile neural brains based on neuron complex dynamics
CN107945210A (en) Target tracking algorism based on deep learning and environment self-adaption
Wiggers et al. Predictive sampling with forecasting autoregressive models
CN115731396A (en) Continuous learning method based on Bayesian variation inference
CN118295029B (en) Seismic data denoising method integrating self-attention and Mamba architecture
CN112766339A (en) Trajectory recognition model training method and trajectory recognition method
Subudhi et al. Kernelized fuzzy modal variation for local change detection from video scenes
EP4392935A1 (en) Robustifying nerf model novel view synthesis to sparse data
CN112837342B (en) Target tracking method, terminal equipment and storage medium
Huang Auto-attentional mechanism in multi-domain convolutional neural networks for improving object tracking
Kalirajan et al. Deep learning for moving object detection and tracking
Lu et al. Cdvae: Co-embedding deep variational auto encoder for conditional variational generation
CN113256685B (en) Target tracking method and system based on convolutional neural network dictionary pair learning
Mehrkanoon et al. Learning from partially labeled data.
Klaiber et al. A Systematic Literature Review on SOTA Machine learning-supported Computer Vision Approaches to Image Enhancement

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant