CN110298868B - High-instantaneity multi-scale target tracking method - Google Patents

High-instantaneity multi-scale target tracking method Download PDF

Info

Publication number
CN110298868B
CN110298868B CN201910559301.0A CN201910559301A CN110298868B CN 110298868 B CN110298868 B CN 110298868B CN 201910559301 A CN201910559301 A CN 201910559301A CN 110298868 B CN110298868 B CN 110298868B
Authority
CN
China
Prior art keywords
scale
target
frame
frames
node
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910559301.0A
Other languages
Chinese (zh)
Other versions
CN110298868A (en
Inventor
王波涛
石梦华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201910559301.0A priority Critical patent/CN110298868B/en
Publication of CN110298868A publication Critical patent/CN110298868A/en
Application granted granted Critical
Publication of CN110298868B publication Critical patent/CN110298868B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/262Analysis of motion using transform domain methods, e.g. Fourier domain methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20056Discrete and fast Fourier transform, [DFT, FFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Abstract

The invention discloses a multi-scale target tracking method with high real-time performance. The method comprises the steps of extracting fast gradient direction histogram (fhog) features of a target and a surrounding background region of the target, generating positive and negative samples through cyclic shift simulation, using a two-dimensional Gaussian function as a sample label, and training a correlation filter by a ridge regression method; the subsequent frame utilizes the response characteristic of a filter to obtain a target position, and a method based on combination of scale prediction and a scale pool is adopted to calculate the target scale; and repeating the training process to update the interpolation of the related filter. The method realizes the optimization of frequency domain operation and scale estimation modes by utilizing the historical motion information of the target, can optimize the operation efficiency of a Kernel Correlation Filter (KCF) method by about 43 percent while ensuring the tracking precision, provides guarantee for the transplantation of a correlation filtering target tracking method to development boards with low operation capability such as an embedded system and the like due to high real-time property, and can be applied to the fields of intelligent monitoring, aerospace, unmanned driving and the like.

Description

High-instantaneity multi-scale target tracking method
Technical Field
The invention belongs to the field of single-target tracking of computer vision, and particularly provides a high-instantaneity multi-scale target tracking method.
Background
The target tracking is based on image processing, integrates technologies such as automatic control, information science and the like, can position a target in each frame of image, and obtains information such as the size of the target, so that the target area is divided from a background area, and the task of tracking the target in the whole video sequence image is completed.
In recent years, the focus of research in the field of discriminant tracking has mainly focused on the improvement of tracking methods based on correlation filters. Correlation filtering is originally a signal processing method, and the correlation of an input signal is judged through the output response of the input signal after passing through a filter. In target tracking, the center position of a tracking target can be determined by finding the maximum response. Fast and accurate scale estimation methods are necessary to achieve an efficient update of the target model. At present, the method for solving the problem of scale calculation is mainly to use a translation filter to perform target detection on an image block subjected to multi-scale scaling in a scale pool mode, and to obtain the translation position with the largest response and the scale where the translation position is located. The method is simple and easy to understand, but is too coarse, so that the overall complexity of the method can increase in multiple levels along with the increase of the number of the detected scales, and the superiority of high-speed performance of the nuclear correlation filtering tracking method is damaged, so that the real-time requirement cannot be met when the method runs on development boards with low computing power such as an embedded processor, and the real landing requirement of the tracking technology cannot be met.
In order to optimize the target scale estimation method based on the scale pool, the existing redundancy of the target scale estimation method is reduced, and the operation efficiency of the tracking method is improved. Since the increase of the calculation scale has strong destructive power, how to reduce the number of scale calculations is considered, and by incorporating the mechanism of scale prediction therein, the speed advantage can be improved while the high accuracy of the tracking method is ensured.
Disclosure of Invention
The invention provides a multi-scale target tracking method with high real-time property, aiming at solving the problem of low scale estimation efficiency in the target tracking process, and the tracking speed of the related filtering type target tracking method is improved while the high tracking precision is ensured.
In order to achieve the technical purpose, the technical scheme adopted by the invention mainly comprises the following steps:
step 1, determining a region of interest. Selecting a target to be tracked by using a rectangular frame in an initial image frame, processing a 2.5-time pixel area around the target, adjusting the size of a long edge to 96, and scaling a shorter edge according to the changed proportion of the long edge;
and 2, extracting features. Extracting 31-dimensional fast gradient direction histogram (fhog) features from the region of interest;
and 3, training a relevant filtering tracker. Converting the characteristic vector and the two-dimensional Gaussian peak function into a frequency domain for calculation to obtain a correlation filter;
and 4, detecting the target position. Extracting the region of interest in the next frame of image in the same way as the step 1, and obtaining the position of the target by utilizing the correlated filtering tracker obtained by training, wherein the maximum response value is;
and 5, estimating the scale. And combining the scale prediction with the scale pool to determine the target scale.
In scale estimation, the prioritization is first calculated for the scales within the scale pool. The change of the target scale per frame can be divided into three cases: 0.95,1,1.05, i.e. small scale, normal, large scale. This indicates that the target is smaller, unchanged, and larger, respectively.
The traditional scale estimation method is that image blocks with different scales are respectively taken to extract features, response values are obtained by the image blocks and relevant filters, the response values are compared, and the largest image block is the optimal scale. The overall complexity and the number of scales of the method are increased by multiple levels, so that the influence on the speed is large. The prioritization method provided by the invention is based on the response characteristic of the correlation filter, and by taking the situation that the target actually becomes larger as an example, the obtained response value is taken as a variable R, and R (0.95) < R (1) < R (1.05) can be deduced. Therefore, if R (1) and R (1.05) are calculated and compared, the actual maximum value can be obtained, and the response to a small scale does not need to be calculated again. And on the contrary, if the target actually becomes smaller, the response with large scale does not need to be calculated again. For the case that the target actual scale is not changed, the calculation can be performed by randomly dividing into one of the above cases, and then the response of the third scale is calculated and compared.
The present invention uses a statistical-based approach, using the actual motion information of the target, to confirm that a response of a certain scale is first calculated. And counting the scale change conditions of the past 10 frames, wherein the scale change is more, namely, the scale is preferentially calculated in the current frame. In either case, R (1), the response when the scale is unchanged, needs to be calculated before comparison.
The statistical method adopts a mode of creating a special circular linked list, and completes the statistics of the scale change condition of the past frame in constant time complexity so as to avoid the influence of high time complexity on the tracking speed.
In order to further improve the tracking speed, the invention tracks the target with almost unchanged scale by adopting a frame-by-frame multi-scale mode after counting multi-frame information. Namely one frame with multiple scales and multiple frames with single scale are alternately carried out. The target scale information can be updated in time, and time waste caused by multi-scale tracking of each frame can be effectively avoided.
The invention also performs relevant optimization on the redundant calculation, stores and utilizes the frequency domain operation of the characteristic vector, reduces the operation times of the Fast Fourier Transform (FFT) of the nuclear correlation filter, and improves the operation efficiency by 20 percent through optimization.
The frequency domain operation and scale estimation method provided by the invention can be applied to all correlation filtering tracking methods, and is not limited to a Kernel Correlation Filter (KCF) tracking method.
Drawings
FIG. 1: method overall flow chart
FIG. 2: circular linked list schematic diagram
Detailed Description
The following detailed description of embodiments of the invention refers to the accompanying drawings.
As shown in the overall flowchart of fig. 1, the present invention discloses a high-real-time multi-scale target tracking method, which specifically includes the following steps:
step 1, determining a region of interest. Selecting a target to be tracked by using a rectangular frame in an initial image frame, processing a 2.5-time pixel area around the target, adjusting the size of a long edge to 96, and scaling a shorter edge according to the changed proportion of the long edge;
and 2, extracting features. Extracting 31-dimensional fast gradient direction histogram (fhog) features from the region of interest;
and 3, training a relevant filtering tracker and optimizing redundancy. Converting the characteristic vector and the two-dimensional Gaussian peak function into a frequency domain for calculation to obtain a correlation filter, wherein the training formula is as follows (1):
where α is the implicit coefficient of the correlation filter, k is the gaussian kernel function variable k (x, x),
Figure BDA0002107791330000031
the calculation formula is shown as (2), x is a feature vector of the training image block, y is a label vector formed by a two-dimensional Gaussian function, and lambda is a regularization factor.
Figure BDA0002107791330000041
In the formula (2), F-1For the purpose of the inverse transform of the FFT,
Figure BDA0002107791330000042
as a result of the conversion of x into the frequency domain, an operation is a dot product, indicates that a conjugate is taken, and σ is the bandwidth. Two inputs of the kernel function in the training are both x, so that only one FFT needs to be actually performed on x, which can reduce the FFT calculation times of 1/2 in the training.
And 4, detecting the target position. Extracting the region of interest in the next frame of image as in step 1, and using the trained correlation filtering tracker, wherein the detection formula is as follows (3):
Figure BDA0002107791330000043
wherein the content of the first and second substances,
Figure BDA0002107791330000044
and outputs the response values for all positions, alpha is obtained in step 3,
Figure BDA0002107791330000048
for the gaussian kernel function variable k (x, z), the formula is calculated as (4). Wherein x is the training image block feature vector and z is the testing image block feature vector. The maximum response value is the position of the target.
Figure BDA0002107791330000045
The two inputs to the kernel function under test are x and z respectively,
Figure BDA0002107791330000046
and
Figure BDA0002107791330000047
in the multi-scale calculation link, the detection process is repeated for image blocks with different scales, namely only the feature vector z of the test image block is changed, so that the FFT (x) in the first scale operation is stored and is directly called in the subsequent operation.
And 5, estimating the scale. And combining the scale prediction with the scale pool to determine the target scale.
In this step scale estimation, the scaling calculations within the scale pool are first prioritized. The change of the target scale of each frame is divided into three cases: 0.95,1,1.05, i.e. small scale, normal, large scale. This indicates that the target is smaller, unchanged, and larger, respectively. The prioritization method provided by the invention is based on the response characteristic of the correlation filter, and by taking the situation that the target actually becomes larger as an example, the obtained response value is taken as a variable R, and R (0.95) < R (1) < R (1.05) can be deduced. Therefore, if R (1) and R (1.05) are calculated and compared, the actual maximum value can be obtained, and the response to a small scale does not need to be calculated again. Similarly, if the target actually becomes smaller, it is not necessary to calculate the response of large scale again. For the case that the target actual scale is not changed, the calculation can be performed by randomly dividing into one of the above cases, and then the response of the third scale is calculated and compared.
The present invention uses a statistical-based approach, using the actual motion information of the target, to confirm that a response of a certain scale is first calculated. And counting the scale change conditions of the past 10 frames, wherein the scale change is more, namely, the scale is preferentially calculated in the current frame. In either case, R (1), the response when the scale is unchanged, needs to be calculated before comparison. The statistical method adopts a mode of creating a special circular linked list, and completes the statistics of the scale change condition of the past frame in constant time complexity so as to avoid the influence of high time complexity on the tracking speed. The circular chain representation is intended to be as shown in fig. 2, and specifically includes the following:
firstly, defining nodes, wherein each node is provided with a bidirectional pointer which points to front and rear nodes respectively, defining an mcache variable inside the node and used for recording the current scale state, and the state is divided into a large state (with the value of 1), a constant state (with the value of 0), a small state (with the value of-1) and a default state (with the value of 2).
After the circular linked list is created, a pointer is used for pointing to any node, the pointer moves down one node every time when one frame passes, and when the number of frames is larger than the length of the queue, the pointer points to the initial node in the linked list, so that the node information is covered, and the recording of the latest 10 frames of information is realized.
In order to reduce the time complexity of the statistical scale information, a mSum variable is maintained for recording the sum of all frame scale states mScale, and the formula is as follows (5):
mSum=∑mScale(mScale≠2) (5)
the default case of mSchale of 2 indicates that the node is not yet covered and cannot contain it.
The mSchale design is a statistical variable of a 0-mean value, if the sum is greater than 0, the result shows that the number of frames with the scale of the first 10 frames increased is large, if the sum is equal to 0, the number of frames with the scale unchanged is large, if the sum is less than 0, the number of frames with the scale reduced is large, the modification of the mSum variable is positioned at the moment of setting the information of the frame each time, and the formula is as follows (6):
mSum=mSum-mScale(i)+mScale(j) (6)
the scale information mSCAle (i) of the current pointing node is subtracted, and the scale information mSCAle (j) of the current frame is added, so that the time complexity occupied by the statistics of the scale change information is constant complexity, and the tracking speed cannot be reduced.
Meanwhile, the head node information is used for judging the full state and the empty state of the team, if the scale state of the head node is the default, the head node is not used, and the empty state of the team can be judged; if the previous node scale state of the head node is not default, it can be determined that the queue is full. The time complexity of the query in the two states is constant complexity, and the tracking speed cannot be reduced.
For the target with almost unchanged scale, the invention adopts a mode of frame-by-frame multi-scale tracking. Namely one frame with multiple scales and multiple frames with single scale are alternately carried out. The multiscale is performed every 5 frames, namely one frame multiscale and 4 frames single scale. The target scale information can be updated in time, and time waste caused by multi-scale tracking of each frame can be effectively avoided.
In conclusion, the multi-scale target tracking method with high real-time performance provided by the invention is explained in detail. The description should not be construed as limiting the invention, which is intended to be limited only by the appended claims.

Claims (3)

1. A multi-scale target tracking method with high real-time performance is characterized by comprising the following steps:
step 1, determining a region of interest; selecting a target to be tracked by using a rectangular frame in an initial image frame, and extracting the target and a surrounding background area thereof;
step 2, extracting characteristics; extracting the fast gradient direction histogram characteristics of the region of interest;
step 3, training a correlation filter; converting the characteristic vector and the two-dimensional Gaussian function into a frequency domain for calculation to obtain a correlation filter;
step 4, detecting the target position; extracting the region of interest in the next frame of image in the same way as the step 1, and obtaining the position of the target with the maximum response value by using the relevant filter obtained by training;
step 5, scale estimation; combining the scale prediction with a scale pool to determine a target scale;
the step 5 specifically comprises the following steps: based on the response characteristics of the relevant filters, calculating and dividing priorities for the scales in the scale pool; the step of dividing the scale priority specifically comprises the following steps:
first, the change of target scale per frame is divided into three cases: 0.95,1,1.05, namely small-scale, normal and large-scale; thereby respectively representing the situations that the target becomes smaller, unchanged and larger;
when the target actually becomes large, the obtained response value is set as a variable R, and R (0.95) < R (1) < R (1.05) is deduced; therefore, the R (1) and the R (1.05) are calculated and compared to obtain the actual maximum value, and the response to the small scale is not required to be calculated again;
if the target actually becomes smaller, it is deduced that R (0.95) > R (1) > R (1.05); therefore, the R (1) and the R (0.95) are calculated and compared to obtain the actual maximum value, and the large-scale response does not need to be calculated again;
if the actual target scale is not changed, randomly dividing the actual target scale into smaller ones or larger ones to calculate the actual maximum value, and calculating the response of the third scale to compare with the actual maximum value;
based on a statistical method, by utilizing the actual motion information of the target, confirming that the response of a certain scale is calculated at first; counting the scale change conditions of the past 10 frames, and determining the most scale change conditions, namely preferentially calculating the response of the scale corresponding to the most scale change conditions in the current frame; no matter which scale is changed most, R (1) needs to be calculated, namely the response when the scale is not changed;
the statistical-based method is as follows:
firstly, defining nodes, wherein each node is provided with a bidirectional pointer which respectively points to front and rear nodes, and an mcache variable is defined in each node and is used for recording the current scale state, and the state is divided into a variable-length value of 1, a constant-length value of 0, a variable-length value of-1 and a default value of 2;
after a circular linked list is created, a pointer is used for pointing to any node, the pointer moves down one node every time when one frame passes, and when the number of frames is larger than the length of a queue, the pointer points to the initial node in the linked list, so that the node information is covered, and the recording of the latest 10 frames of information is realized;
in order to reduce the time complexity of statistical scale information, maintaining an mSum variable for recording the sum of all frame scale states mSule, wherein the mSum is designed to be a statistical variable with a mean value of 0, if the sum is greater than 0, the number of frames with the scale of the first 10 frames being increased is more, if the sum is equal to 0, the number of frames with the scale being unchanged is more, if the sum is less than 0, the number of frames with the scale being decreased is more, the modification of the mSum variable is positioned at the time of setting current frame information each time, and the mSum subtracts the scale information of the current pointing node and the scale information of the current;
meanwhile, the head node information is used for judging the full state and the empty state of the team, and if the scale state of the head node is default, the head node is not used, and the empty state of the team is judged; and if the previous node scale state of the head node is not default, judging that the queue is full.
2. The multi-scale target tracking method with high real-time performance according to claim 1, characterized in that for a target with unchanged statistical scale, the tracking method adopts a frame-by-frame multi-scale mode; namely one frame with multiple scales and multiple frames with single scale are alternately carried out.
3. The method for multi-scale target tracking with high real-time performance as claimed in claim 2, wherein multi-scale is performed every 5 frames, namely, one frame multi-scale and 4 frames single-scale.
CN201910559301.0A 2019-06-26 2019-06-26 High-instantaneity multi-scale target tracking method Active CN110298868B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910559301.0A CN110298868B (en) 2019-06-26 2019-06-26 High-instantaneity multi-scale target tracking method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910559301.0A CN110298868B (en) 2019-06-26 2019-06-26 High-instantaneity multi-scale target tracking method

Publications (2)

Publication Number Publication Date
CN110298868A CN110298868A (en) 2019-10-01
CN110298868B true CN110298868B (en) 2021-06-25

Family

ID=68028836

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910559301.0A Active CN110298868B (en) 2019-06-26 2019-06-26 High-instantaneity multi-scale target tracking method

Country Status (1)

Country Link
CN (1) CN110298868B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112686609B (en) * 2020-12-31 2021-08-13 江苏佳利达国际物流股份有限公司 Intelligent unmanned logistics transportation method and system based on optimization efficiency evaluation algorithm
CN112991394B (en) * 2021-04-16 2024-01-19 北京京航计算通讯研究所 KCF target tracking method based on cubic spline interpolation and Markov chain

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107103616A (en) * 2017-04-27 2017-08-29 中国科学院长春光学精密机械与物理研究所 A kind of method for tracking target and system
CN108010067A (en) * 2017-12-25 2018-05-08 北京航空航天大学 A kind of visual target tracking method based on combination determination strategy
CN108346159A (en) * 2018-01-28 2018-07-31 北京工业大学 A kind of visual target tracking method based on tracking-study-detection
CN108550161A (en) * 2018-03-20 2018-09-18 南京邮电大学 A kind of dimension self-adaption core correlation filtering fast-moving target tracking method
CN109685073A (en) * 2018-12-28 2019-04-26 南京工程学院 A kind of dimension self-adaption target tracking algorism based on core correlation filtering

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106204638B (en) * 2016-06-29 2019-04-19 西安电子科技大学 It is a kind of based on dimension self-adaption and the method for tracking target of taking photo by plane for blocking processing
CN109410247A (en) * 2018-10-16 2019-03-01 中国石油大学(华东) A kind of video tracking algorithm of multi-template and adaptive features select

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107103616A (en) * 2017-04-27 2017-08-29 中国科学院长春光学精密机械与物理研究所 A kind of method for tracking target and system
CN108010067A (en) * 2017-12-25 2018-05-08 北京航空航天大学 A kind of visual target tracking method based on combination determination strategy
CN108346159A (en) * 2018-01-28 2018-07-31 北京工业大学 A kind of visual target tracking method based on tracking-study-detection
CN108550161A (en) * 2018-03-20 2018-09-18 南京邮电大学 A kind of dimension self-adaption core correlation filtering fast-moving target tracking method
CN109685073A (en) * 2018-12-28 2019-04-26 南京工程学院 A kind of dimension self-adaption target tracking algorism based on core correlation filtering

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
ROBUST VISUAL TRACKING USING JOINT SCALE-SPATIAL CORRELATION FILTERS;Mengdan Zhang et al;《2015 IEEE international conference on image processing》;20151231;第1468-1472页 *
基于可变形部件改进模型的夜间车辆检测方法;孙营 等;《计算机工程》;20190331;第45卷(第3期);第202-206页 *
快速尺度估计的核相关滤波目标跟踪;丁建伟 等;《科学技术与工程》;20170531;第17卷(第15期);第111-114页 *

Also Published As

Publication number Publication date
CN110298868A (en) 2019-10-01

Similar Documents

Publication Publication Date Title
Jiang et al. WeSamBE: A weight-sample-based method for background subtraction
CN110135500B (en) Target tracking method under multiple scenes based on self-adaptive depth characteristic filter
CN110991272B (en) Multi-target vehicle track recognition method based on video tracking
Chen et al. Embedded system real-time vehicle detection based on improved YOLO network
Mao et al. A delay metric for video object detection: What average precision fails to tell
CN112836639A (en) Pedestrian multi-target tracking video identification method based on improved YOLOv3 model
CN110120064A (en) A kind of depth related objective track algorithm based on mutual reinforcing with the study of more attention mechanisms
CN110298868B (en) High-instantaneity multi-scale target tracking method
CN108986142A (en) Shelter target tracking based on the optimization of confidence map peak sidelobe ratio
CN113177640B (en) Discrete asynchronous event data enhancement method
CN112634316B (en) Target tracking method, device, equipment and storage medium
CN111968153A (en) Long-time target tracking method and system based on correlation filtering and particle filtering
CN111178261A (en) Face detection acceleration method based on video coding technology
CN116402850A (en) Multi-target tracking method for intelligent driving
CN109978923B (en) Target tracking method and system based on double-template scale self-adaptive correlation filtering
CN111931722A (en) Correlated filtering tracking method combining color ratio characteristics
Zheng et al. Learning Robust Gaussian Process Regression for Visual Tracking.
Liu et al. DP-FishNet: Dual-path Pyramid Vision Transformer-based underwater fish detection network
CN111161323B (en) Complex scene target tracking method and system based on correlation filtering
CN117274314A (en) Feature fusion video target tracking method and system
Mohamed et al. PAE: Portable appearance extension for multiple object detection and tracking in traffic Scenes
CN110659724A (en) Target detection convolutional neural network construction method based on target scale range
CN108093153B (en) Target tracking method and device, electronic equipment and storage medium
CN115953431A (en) Multi-target tracking method and system for aerial video of unmanned aerial vehicle
Pillai et al. Fine-Tuned EfficientNetB4 Transfer Learning Model for Weather Classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant