CN109493370A - A kind of method for tracking target based on spatial offset study - Google Patents

A kind of method for tracking target based on spatial offset study Download PDF

Info

Publication number
CN109493370A
CN109493370A CN201811186951.7A CN201811186951A CN109493370A CN 109493370 A CN109493370 A CN 109493370A CN 201811186951 A CN201811186951 A CN 201811186951A CN 109493370 A CN109493370 A CN 109493370A
Authority
CN
China
Prior art keywords
target
spatial offset
image block
image
particle
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811186951.7A
Other languages
Chinese (zh)
Other versions
CN109493370B (en
Inventor
权伟
李天瑞
高仕斌
赵丽平
陈金强
陈锦雄
刘跃平
卢学民
王晔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southwest Jiaotong University
Original Assignee
Southwest Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southwest Jiaotong University filed Critical Southwest Jiaotong University
Priority to CN201811186951.7A priority Critical patent/CN109493370B/en
Publication of CN109493370A publication Critical patent/CN109493370A/en
Application granted granted Critical
Publication of CN109493370B publication Critical patent/CN109493370B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of method for tracking target based on spatial offset study, are related to technical field of computer vision.The target object to be tracked is selected and determined, Object selection process is automatically extracted or is manually specified by moving target detecting method;Spatial offset learning network includes that image data extraction, deep neural network, multi-layer perception (MLP) MLP and spatial offset export four parts: under real-time disposition, the video image that memory block is acquired and be stored in by camera is extracted, as the input picture to be tracked;In processed offline, the video file acquired is decomposed into the image sequence of multiple frame compositions;Tracking uses particle filter method in short-term, a particle in particle filter represents a possible target image block, using corresponding target region-of-interest ROI as the on-line training collection of spatial offset learning network, on-line training is carried out to spatial offset learning network using stochastic gradient descent method SGD, updates network parameter;Target positioning and update.

Description

A kind of method for tracking target based on spatial offset study
Technical field
The present invention relates to computer vision, graph image, pattern-recognition, machine learning techniques fields.
Background technique
Visual target tracking is the important subject of computer vision field, and main task is that acquisition target is continuous The information such as position, appearance and movement, and then base is provided for further semantic layer analysis (such as Activity recognition, scene understanding) Plinth.Target following research is widely used in the fields such as intelligent monitoring, human-computer interaction, automatic control system, has very strong reality With value.Currently, method for tracking target mainly includes classical method for tracking target and deep learning method for tracking target.
Classical method for tracking target is broadly divided into production method (Generative Methods) and discriminate method (Discriminative Methods) two classes.Production method assume target can by certain generating process or model into Row expression, such as principal component analysis (PCA), then tracking problem is considered as interested by sparse coding (Sparse Coding) etc. Region in find most probable candidate item.These methods are intended to design a kind of image expression side conducive to robust target following Method.And in terms of motion modeling, being normally based on certain assumes and verifies to carry out, typical method such as Kalman filtering, mean value Drift, particle filter (PF, Particle Filter) etc., wherein particle filter technology is the solution of nonlinear and non-Gaussian problem A kind of highly effective means are provided, it has good robustness to partial occlusion and background interference.Different from production side Tracking is considered as a classification or a kind of continuous object detection problem by method, discriminate method, and task is by target from figure As being distinguished in background.Such methods utilize target and background information simultaneously, are a kind of methods mainly studied at present.Differentiate Formula method generally comprises two main steps, and the first step is to be capable of the visual signature training of discrimination objective and background by selection A classifier and its decision rule are obtained, second step is that the classifier is used for each in visual field during tracking Position evaluate and determines most possible target position.Target frame is then moved to the position and repeats such mistake Journey, and then realize tracking, which be used to design various forms of track algorithms.All in all, classical tracking Main advantage is the speed of service and the dependence less to auxiliary data, while they are also required to the accuracy and in real time in tracking Tradeoff is made between property.
Deep learning (Deep Learning) is the hot spot of the research of machine learning in recent years, due to its powerful mark sheet The data set and hardware supported of Danone power and continuous development, deep learning achieve surprising success in many aspects, such as Speech recognition, image recognition, target detection, visual classification etc..Deep learning target following research and development is also very rapid, but by The requirement of the shortage and real-time of priori knowledge in target following, so that based on needing a large amount of training datas and parameter to calculate Depth learning technology be difficult to adequately be put to good use in this respect, have very big exploration space.From current research achievement From the point of view of, deep learning tracking mainly applies self-encoding encoder network and convolutional neural networks, and there are mainly two types of think for research Road, one is carrying out transfer learning to network to carry out on-line fine again, another kind be the structure of transformation depth network with adapt to The requirement of track.Self-encoding encoder network (AE) is typical non-supervisory deep learning network, because of its feature learning ability and antinoise Performance is applied first in target following.In general, self-encoding encoder network is relatively more intuitive and the scale of construction is moderate, is a kind of outstanding Non-supervisory deep learning model, be able to apply and achieve preferable effect in the track at first.Not with self-encoding encoder network Together, convolutional neural networks (CNN) are a kind of feedforward neural networks of supervision type, its convolution comprising the progress of multiple cycle alternations, Nonlinear transformation and down-sampled operation, embody very powerful performance in pattern-recognition especially Computer Vision Task. All in all, deep learning has more powerful feature representation ability, related training in tracking compared to classical way The selection of collection, the selection of network and the improvement of structure, real-time of algorithm, and application recurrent neural network etc. there is still a need for Further research.
Therefore, the high robust in view of particle filter PF and the powerful feature representation ability of deep neural network, the present invention It is proposed a kind of method for tracking target based on spatial offset study.This method uses 152 layers of deep neural network GoogleNet Feature representation is carried out to input picture, the image block selected by target region-of-interest ROI in frame image and its in the area is made For the input data of spatial offset learning network, the sky of selected image block and target is then exported through multi-layer perception (MLP) MLP study Between deviant, using the particle in particle filter as the image block selected in target region-of-interest ROI during tracking, even With target region-of-interest ROI, the corresponding spatial offset value of each particle can be obtained after the processing of network forward direction, and then be calculated Its corresponding future position is finally realized the position for having prediction most multiple as new target position and is determined target Position, and then realize tracking.Feature representation is carried out because on the one hand this method uses deep neural network, on the other hand simultaneously will The image data of target and background is as network inputs, and studying space deviant, so that target position fixing process is more accurate, and By effective combination with particle filter during tracking, takes full advantage of the high robust of particle filter and simplify network Learning process, and then can be realized the target following of long-time real-time stabilization.In addition, the method for the present invention can be not only used for list Target following can also extend the tracking for multiple target by increasing and adjusting sample label.
Summary of the invention
The object of the present invention is to provide a kind of method for tracking target based on spatial offset study, it can be efficiently solved The technical issues of tracking of long-time real-time stabilization is carried out to general target object under no constraint environment.
The purpose of the present invention is achieved through the following technical solutions: a kind of target following based on spatial offset study Method includes the following steps:
Step 1: Object selection:
The target object to be tracked is selected and determined from initial pictures, Object selection process passes through moving object detection side Method automatically extracts, or is manually specified by man-machine interaction method;
Step 2: the building of spatial offset learning network and initialization:
Spatial offset learning network includes image data extraction, deep neural network, multi-layer perception (MLP) MLP and spatial offset Export four parts:
Image data extraction part extracts two kinds of image blocks to original image, one is the region-of-interest ROI of target, That is goal-orientation, and the image block of 9 times of target sizes, another kind are random in the region-of-interest ROI of the target The image block with target with same size of selection;For deep neural network part, using the pre-training that can disclose acquisition Network G oogleNet carries out feature representation to image, and it is comprising the big of a training images up to a million which, which shares 154 layers, The deep neural network that training obtains on scale data collection ImageNet, input picture pass through the scale of 224 × 224 pixel sizes Input data after normalization as GoogleNet network;It is used as feature representation layer by 152 layers of GoogleNet, it has 1024 A value output;The image block selected in using target region-of-interest ROI and the region is defeated as the image of GoogleNet Enter, after the processing of GoogleNet forward direction, exports two characteristic values, each characteristic value includes 1024 values, by the two characteristic values Data after connection merging as the part multi-layer perception (MLP) MLP input;Multi-layer perception (MLP) MLP include three full articulamentums, first Node layer number is 2048, and second layer number of nodes is 1024, and third layer number of nodes is 512, and multi-layer perception (MLP) MLP's is last One layer of connection space deviates output par, c;Spatial offset output par, c includes 4 values, this 4 values are respectively that selected image block is left The difference Dx of upper angle abscissa and target upper left corner abscissal, the vertical seat of selected image block upper left corner ordinate and the target upper left corner Target difference Dyl, the difference Dx of selected image block lower right corner abscissa and target lower right corner abscissar, the selected image block lower right corner The difference Dy of ordinate and target lower right corner ordinater
Target region-of-interest ROI and the image block selected in the region are pairs of input space offset learning networks , for same frame image, target region-of-interest ROI is constant, and in the region select image block have it is multiple;It is paid close attention in target It randomly chooses 1000 image blocks in the ROI of region, and records the spatial offset value of each selected image block and target, i.e., it is selected The coordinate difference Dx of image block and targetl, Dyl, Dxr, Dyr, by being obtained centered on the target image block determined in step 1 Target region-of-interest ROI and the region in the image block that selects generate initial training collection, using SGD pairs of stochastic gradient descent method Spatial offset learning network carries out off-line training, so that it is determined that the parameter of multi-layer perception (MLP), and then complete spatial offset and learn net The initialization of network;
Step 3: image inputs:
It under real-time disposition, extracts and the video image of memory block is acquired and be stored in by camera, as will be into The input picture of line trace;In processed offline, the video file acquired is decomposed into the image sequence of multiple frame compositions Column extract frame image as input picture sequentially in time one by one;If input picture is sky, whole flow process stops;
Step 4: tracking in short-term:
Tracking uses particle filter method (PF, Particle Filter) in short-term, and a particle in particle filter represents One possible target image block, particle filter include 1000 particles, and each particle is random in target region-of-interest ROI Selection obtains, and target region-of-interest ROI is and the region of 9 times of target sizes centered on the target that the last time determines; The similarity value of the prediction target image block of particle filter output and target image block is calculated as between the two image blocks Regularization intersects cross correlation value NCC (Normalized Cross-Correlation), then if value > 0.9 during tracking, Then indicate that the target for tracking output in short-term is credible, target positioning is completed, and the spatial offset of each particle Yu new definition target is recorded Value, i.e. the coordinate difference Dx of particle and targetl, Dyl, Dxr, Dyr, step 5 is jumped to, otherwise indicates that target is insincere, jumps To step 6;
Step 5: network on-line training:
The particle of particle filter and its corresponding target region-of-interest ROI are as spatial offset learning network using in step 4 On-line training collection, using stochastic gradient descent method SGD to spatial offset learning network carry out on-line training, update network ginseng Number;
Step 6: target positioning and update:
Each particle of particle filter in current goal region-of-interest ROI and step 4 is separately input to spatial offset Practise network, through network forward direction processing after, export the corresponding spatial offset value of each particle, according to where particle position and the grain The corresponding spatial offset value of son, is calculated the target position of the particle prediction;The target position for counting all particle predictions, With most multiple position is predicted as new target position, target positioning is completed;Calculate the target image block and grain of new definition The similarity value between target image block that son filtering tracks in short-term, i.e. their regularization intersect cross correlation value NCC, if should Value > 0.9 then updates the target image block that particle filter tracks in short-term with the target image block of new definition;Jump to step 3.
During tracking, when the objective result that particle filter tracks output in short-term is credible, tracking can carry out reality in short-term When target following, while to spatial offset learning network carry out on-line training, and when its output objective result it is insincere when, then Target is positioned by spatial offset learning network, at the same according to the network determine target come update particle filter in short-term with The target image block of track.Feature representation is carried out because on the one hand this method uses deep neural network, on the other hand simultaneously will The image data of target and background is as network inputs, and studying space deviant, so that target position fixing process is more accurate, and By effective combination with particle filter during tracking, takes full advantage of the high robust of particle filter and simplify network Learning process, and then can be realized the target following of long-time real-time stabilization.
Compared with prior art the advantages of and good effect: the present invention propose it is a kind of based on spatial offset study target with Track method.This method carries out feature representation to input picture using 152 layers of deep neural network GoogleNet, by frame image Middle target region-of-interest ROI and its image block selected in the area are inputted as spatial offset learning network, then through more Layer perceptron MLP study exports the spatial offset value of selected image block and target, by the particle in particle filter during tracking It can after the processing of network forward direction together with target region-of-interest ROI as the image block selected in target region-of-interest ROI The corresponding spatial offset value of each particle is obtained, and then its corresponding future position is calculated, finally with prediction Most multiple position can realize the positioning to target as new target position, and then realize tracking.On the one hand due to this method Feature representation is carried out using deep neural network, it is on the other hand simultaneously that the image data of target and background is defeated as network Enter, and studying space deviant, so that target position fixing process is more accurate, and by having with particle filter during tracking Effect combines, and takes full advantage of the high robust of particle filter and simplifies the learning process of network, and then can be realized for a long time The target following of real-time stabilization.In addition, the method for the present invention can be not only used for monotrack, by increasing and adjusting sample mark Note, can also extend the tracking for multiple target.
Detailed description of the invention
Fig. 1 is spatial offset learning network structure composition schematic diagram of the present invention
Fig. 2 is method for tracking target flow chart of the present invention
Specific embodiment
Embodiment:
Method of the invention can be used for target object tracking various occasions, such as intelligent video analysis, automatic human-computer interaction, Logical video monitoring, vehicle drive, and biocenose analysis and flow surface test the speed.
By taking intelligent video analysis as an example: intelligent video analysis includes many important to automatically analyze task, such as object behavior Analysis, video compress etc., and the basis of these work is then the target following for being able to carry out long-time stable.This hair can be used The tracking realization of bright proposition, specifically, first simultaneously according to picture construction spatial offset learning network where target selection Initialization training is completed, as shown in the spatial offset learning network structure composition of Fig. 1;Then it is filtered during tracking using particle Wave method PF is tracked in short-term, when particle filter track in short-term determining target it is credible when, according in short-term tracking determined by Target region-of-interest ROI is extracted in target position, constitutes online training set and training spatial offset together with the particle of particle filter Practise network, and when particle filter track in short-term determining target it is insincere when, then by spatial offset learning network to target carry out It positions, while updating the target image block of particle filter according to the target that the network determines.Because on the one hand this method uses Deep neural network carries out feature representation, on the other hand simultaneously using the image data of target and background as network inputs, and Studying space deviant so that target position fixing process is more accurate, and passes through effective knot with particle filter during tracking It closes, takes full advantage of the high robust of particle filter and simplify the learning process of network, and then can be realized for a long time in real time Stable target following.
The method of the present invention can be programmed by any computer programming language (such as C language) and be realized, based on this method Tracking system software can realize real-time modeling method application in any PC or embedded system.

Claims (1)

1. a kind of method for tracking target based on spatial offset study, includes the following steps:
Step 1: Object selection:
The target object to be tracked is selected and determined from initial pictures, Object selection process passes through moving target detecting method certainly It is dynamic to extract, or be manually specified by man-machine interaction method;
Step 2: the building of spatial offset learning network and initialization:
Spatial offset learning network includes image data extraction, deep neural network, multi-layer perception (MLP) MLP and spatial offset output Four parts:
Image data extraction part extracts two kinds of image blocks to original image, and one is the region-of-interest ROI of target, i.e., with Centered on target, and the image block of 9 times of target sizes, another kind are randomly choosed in the region-of-interest ROI of the target With target have same size image block;For deep neural network part, using the pre-training network that can disclose acquisition GoogleNet carries out feature representation to image, and it is comprising the extensive of a training images up to a million which, which shares 154 layers, The deep neural network that training obtains on data set ImageNet, input picture pass through the scale normalizing of 224 × 224 pixel sizes Input data after change as GoogleNet network;It is used as feature representation layer by 152 layers of GoogleNet, it there are 1024 values Output;The image block selected in using target region-of-interest ROI and the region is inputted as the image of GoogleNet, After the processing of GoogleNet forward direction, two characteristic values are exported, each characteristic value includes 1024 values, the two characteristic values are connected It engages and the rear data as the part multi-layer perception (MLP) MLP inputs;Multi-layer perception (MLP) MLP includes three full articulamentums, first layer Number of nodes be 2048, second layer number of nodes be 1024, third layer number of nodes be 512, multi-layer perception (MLP) MLP last Layer connection space deviates output par, c;Spatial offset output par, c includes 4 values, this 4 values are respectively selected image block upper left The difference Dx of angle abscissa and target upper left corner abscissal, selected image block upper left corner ordinate and target upper left corner ordinate Difference Dyl, the difference Dx of selected image block lower right corner abscissa and target lower right corner abscissar, the selected image block lower right corner is indulged The difference Dy of coordinate and target lower right corner ordinater
Target region-of-interest ROI and the image block selected in the region are pairs of input space offset learning networks, right In same frame image, target region-of-interest ROI is constant, and in the region select image block have it is multiple;In target region-of-interest It randomly chooses 1000 image blocks in ROI, and records the spatial offset value of each selected image block and target, i.e., selected image The coordinate difference Dx of block and targetl, Dyl, Dxr, Dyr, pass through the mesh obtained centered on the target image block determined in step 1 It marks the image block selected in region-of-interest ROI and the region and generates initial training collection, using stochastic gradient descent method SGD to space It deviates learning network and carries out off-line training, so that it is determined that the parameter of multi-layer perception (MLP), and then complete spatial offset learning network Initialization;
Step 3: image inputs:
Under real-time disposition, extract and the video image of memory block acquired and be stored in by camera, as to carry out with The input picture of track;In processed offline, the video file acquired is decomposed into the image sequence of multiple frame compositions, is pressed According to time sequencing, frame image is extracted one by one as input picture;If input picture is sky, whole flow process stops;
Step 4: tracking in short-term:
Tracking in short-term uses particle filter method, and a particle in particle filter represents a possible target image block, grain Son filtering includes 1000 particles, and each particle is to randomly choose to obtain in target region-of-interest ROI, the target region-of-interest ROI is and the region of 9 times of target sizes centered on the target that the last time determines;The prediction mesh that particle filter is exported The regularization that logo image block and the similarity value of target image block are calculated as between the two image blocks intersects cross correlation value NCC, If value > 0.9 during then tracking, then it represents that the target for tracking output in short-term is credible, and target positioning is completed, and records each grain The spatial offset value of son and new definition target, i.e. the coordinate difference Dx of particle and targetl, Dyl, Dxr, Dyr, step 5 is jumped to, Otherwise indicate that target is insincere, jumps to step 6;
Step 5: network on-line training:
The particle of particle filter and its corresponding target region-of-interest ROI exist as spatial offset learning network using in step 4 Line training set carries out on-line training to spatial offset learning network using stochastic gradient descent method SGD, updates network parameter;
Step 6: target positioning and update:
Each particle of particle filter in current goal region-of-interest ROI and step 4 is separately input to spatial offset study net Network, through network forward direction processing after, export the corresponding spatial offset value of each particle, according to where particle position and the particle pair The spatial offset value answered, is calculated the target position of the particle prediction;The target position for counting all particle predictions, having Most multiple position is predicted as new target position, target positioning is completed;The target image block and particle for calculating new definition are filtered The similarity value between target image block that wave tracks in short-term, i.e. their regularization intersect cross correlation value NCC, if the value > 0.9, then the target image block that particle filter tracks in short-term is updated with the target image block of new definition;Jump to step 3.
CN201811186951.7A 2018-10-12 2018-10-12 Target tracking method based on space offset learning Active CN109493370B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811186951.7A CN109493370B (en) 2018-10-12 2018-10-12 Target tracking method based on space offset learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811186951.7A CN109493370B (en) 2018-10-12 2018-10-12 Target tracking method based on space offset learning

Publications (2)

Publication Number Publication Date
CN109493370A true CN109493370A (en) 2019-03-19
CN109493370B CN109493370B (en) 2021-07-02

Family

ID=65689745

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811186951.7A Active CN109493370B (en) 2018-10-12 2018-10-12 Target tracking method based on space offset learning

Country Status (1)

Country Link
CN (1) CN109493370B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993770A (en) * 2019-04-09 2019-07-09 西南交通大学 A kind of method for tracking target of adaptive space-time study and state recognition
CN110007675A (en) * 2019-04-12 2019-07-12 北京航空航天大学 A kind of Vehicular automatic driving decision system based on driving situation map and the training set preparation method based on unmanned plane
CN110472588A (en) * 2019-08-19 2019-11-19 上海眼控科技股份有限公司 Anchor point frame determines method, apparatus, computer equipment and storage medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160127641A1 (en) * 2014-11-03 2016-05-05 Robert John Gove Autonomous media capturing
CN106169188A (en) * 2016-07-11 2016-11-30 西南交通大学 A kind of method for tracing object based on the search of Monte Carlo tree
CN106980843A (en) * 2017-04-05 2017-07-25 南京航空航天大学 The method and device of target following
CN107146237A (en) * 2017-04-24 2017-09-08 西南交通大学 A kind of method for tracking target learnt based on presence with estimating
CN107169998A (en) * 2017-06-09 2017-09-15 西南交通大学 A kind of real-time tracking and quantitative analysis method based on hepatic ultrasound contrast enhancement image
CN107316058A (en) * 2017-06-15 2017-11-03 国家新闻出版广电总局广播科学研究院 Improve the method for target detection performance by improving target classification and positional accuracy
CN107369169A (en) * 2017-06-08 2017-11-21 温州大学 The approximate most like image block matching method transmitted based on direction alignment with matching that a kind of GPU accelerates
CN108154118A (en) * 2017-12-25 2018-06-12 北京航空航天大学 A kind of target detection system and method based on adaptive combined filter with multistage detection
CN108470355A (en) * 2018-04-04 2018-08-31 中山大学 Merge the method for tracking target of convolutional network feature and discriminate correlation filter
CN108537825A (en) * 2018-03-26 2018-09-14 西南交通大学 A kind of method for tracking target based on transfer learning Recurrent networks
CN108573246A (en) * 2018-05-08 2018-09-25 北京工业大学 A kind of sequential action identification method based on deep learning

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160127641A1 (en) * 2014-11-03 2016-05-05 Robert John Gove Autonomous media capturing
CN106169188A (en) * 2016-07-11 2016-11-30 西南交通大学 A kind of method for tracing object based on the search of Monte Carlo tree
CN106980843A (en) * 2017-04-05 2017-07-25 南京航空航天大学 The method and device of target following
CN107146237A (en) * 2017-04-24 2017-09-08 西南交通大学 A kind of method for tracking target learnt based on presence with estimating
CN107369169A (en) * 2017-06-08 2017-11-21 温州大学 The approximate most like image block matching method transmitted based on direction alignment with matching that a kind of GPU accelerates
CN107169998A (en) * 2017-06-09 2017-09-15 西南交通大学 A kind of real-time tracking and quantitative analysis method based on hepatic ultrasound contrast enhancement image
CN107316058A (en) * 2017-06-15 2017-11-03 国家新闻出版广电总局广播科学研究院 Improve the method for target detection performance by improving target classification and positional accuracy
CN108154118A (en) * 2017-12-25 2018-06-12 北京航空航天大学 A kind of target detection system and method based on adaptive combined filter with multistage detection
CN108537825A (en) * 2018-03-26 2018-09-14 西南交通大学 A kind of method for tracking target based on transfer learning Recurrent networks
CN108470355A (en) * 2018-04-04 2018-08-31 中山大学 Merge the method for tracking target of convolutional network feature and discriminate correlation filter
CN108573246A (en) * 2018-05-08 2018-09-25 北京工业大学 A kind of sequential action identification method based on deep learning

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
NAOKI AKAI ET AL: "Simultaneous pose and reliability estimation using convolutional neural network and Rao–Blackwellized particle filter", 《ADVANCED ROBOTICS》 *
X KONG ET AL: "Attentional convolutional neural networks for object tracking", 《2018 INTEGRATED COMMUNICATIONS, NAVIGATION, SURVEILLANCE CONFERENCE (ICNS)》 *
习文星: "移动背景下视觉的行人检测、识别与跟踪技术研究", 《中国博士学位论文全文数据库 信息科技辑》 *
常发亮等: "复杂环境下基于自适应粒子滤波器的目标跟踪", 《电子学报》 *
张毅锋等: "基于基完备化理论和嵌入多层感知机的深度网络结构设计", 《东南大学学报(自然科学版)》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993770A (en) * 2019-04-09 2019-07-09 西南交通大学 A kind of method for tracking target of adaptive space-time study and state recognition
CN109993770B (en) * 2019-04-09 2022-07-15 西南交通大学 Target tracking method for adaptive space-time learning and state recognition
CN110007675A (en) * 2019-04-12 2019-07-12 北京航空航天大学 A kind of Vehicular automatic driving decision system based on driving situation map and the training set preparation method based on unmanned plane
CN110472588A (en) * 2019-08-19 2019-11-19 上海眼控科技股份有限公司 Anchor point frame determines method, apparatus, computer equipment and storage medium
CN110472588B (en) * 2019-08-19 2020-11-24 上海眼控科技股份有限公司 Anchor point frame determining method and device, computer equipment and storage medium

Also Published As

Publication number Publication date
CN109493370B (en) 2021-07-02

Similar Documents

Publication Publication Date Title
CN109800689B (en) Target tracking method based on space-time feature fusion learning
CN107967451B (en) Method for counting crowd of still image
Molchanov et al. Online detection and classification of dynamic hand gestures with recurrent 3d convolutional neural network
CN109344725B (en) Multi-pedestrian online tracking method based on space-time attention mechanism
Gurghian et al. Deeplanes: End-to-end lane position estimation using deep neural networksa
CN106169188B (en) A kind of method for tracing object based on the search of Monte Carlo tree
CN112842690B (en) Machine vision with dimension data reduction
CN107146237B (en) Target tracking method based on online state learning and estimation
CN109816689A (en) A kind of motion target tracking method that multilayer convolution feature adaptively merges
CN110660082A (en) Target tracking method based on graph convolution and trajectory convolution network learning
CN105809718B (en) A kind of method for tracing object of track entropy minimization
CN114241548A (en) Small target detection algorithm based on improved YOLOv5
CN111862145B (en) Target tracking method based on multi-scale pedestrian detection
CN112836640A (en) Single-camera multi-target pedestrian tracking method
CN109993770A (en) A kind of method for tracking target of adaptive space-time study and state recognition
CN113763424B (en) Real-time intelligent target detection method and system based on embedded platform
CN109493370A (en) A kind of method for tracking target based on spatial offset study
CN110728694A (en) Long-term visual target tracking method based on continuous learning
Munir et al. LDNet: End-to-end lane marking detection approach using a dynamic vision sensor
CN106529441B (en) Depth motion figure Human bodys' response method based on smeared out boundary fragment
CN108537825B (en) Target tracking method based on transfer learning regression network
CN114283355A (en) Multi-target endangered animal tracking method based on small sample learning
CN110197123A (en) A kind of human posture recognition method based on Mask R-CNN
CN110111358B (en) Target tracking method based on multilayer time sequence filtering
CN109272036A (en) A kind of random fern method for tracking target based on depth residual error network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant