CN112991395B - Vision tracking method based on foreground condition probability optimization scale and angle - Google Patents

Vision tracking method based on foreground condition probability optimization scale and angle Download PDF

Info

Publication number
CN112991395B
CN112991395B CN202110462758.7A CN202110462758A CN112991395B CN 112991395 B CN112991395 B CN 112991395B CN 202110462758 A CN202110462758 A CN 202110462758A CN 112991395 B CN112991395 B CN 112991395B
Authority
CN
China
Prior art keywords
frame
foreground
angle
tracking
scale
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110462758.7A
Other languages
Chinese (zh)
Other versions
CN112991395A (en
Inventor
安志勇
刘晓庆
原达
赵峰
王彦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong Technology and Business University
Original Assignee
Shandong Technology and Business University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong Technology and Business University filed Critical Shandong Technology and Business University
Priority to CN202110462758.7A priority Critical patent/CN112991395B/en
Publication of CN112991395A publication Critical patent/CN112991395A/en
Application granted granted Critical
Publication of CN112991395B publication Critical patent/CN112991395B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Geometry (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a visual tracking method based on foreground condition probability optimization scale and angle, and belongs to the field of computer vision. The method comprises the following implementation steps: 1) reading a video frame sequence, and calculating a regression frame, a segmentation mask (foreground) and a minimum external frame of the frame by using a SimMask method; 2) calculating the proportion of the foreground area in the minimum external frame; 3) when the proportion is smaller than a set threshold value, calculating the reliability of the minimum external frame of the frame; 4) selecting different strategies according to the reliability to optimize the minimum bounding box dimension; 5) performing offset setting aiming at the angle of the tracking frame after the scale optimization; 6) IoU values of each offset angle rotation frame and the foreground are calculated; 7) the tracker adapts the rotation box that outputs the largest value of the foreground IoU. The visual tracking method effectively improves the overall performance of target tracking under the complex conditions of target motion, rotation, scale change and the like.

Description

Vision tracking method based on foreground condition probability optimization scale and angle
Technical Field
The invention relates to the field of computer vision, in particular to a vision tracking method based on foreground condition probability optimization scale and angle.
Background
Target tracking is a hot point problem of computer vision, is a premise and a basis for completing higher-level image understanding, and is widely applied to the fields of intelligent video monitoring, human-computer interaction, visual navigation, medical diagnosis and the like. In the process of tracking the target, background interference (shielding, illumination change and the like) and interference of target change (rotation, scale change, deformation and the like) are often encountered, and the problem of tracking offset is caused in subsequent frames, so that the target tracking is a very challenging problem in the field of computer vision.
In recent years, tracking methods are gradually shifted to discriminant by generative methods, which are represented by correlation filtering and deep learning. At present, the twin network series based on deep learning combines the advantages of relevant filtering and deep learning, and excellent performance is achieved in recent years. The twin network is a special neural network structure, the template image and the search image share the weight of a neural network, and the target position is determined after the characteristic diagrams of the two branches are mutually correlated.
The siamf fc network proposed by Bertinetto et al, which is called the mountain-opening operation of the twin network, is very fast in tracking speed, but requires multi-scale detection, and thus tracking accuracy is not ideal. The SiamRPN proposed by Li and the like is added into a regional suggestion network, and compared with the SiamFC, the network uses boundary box regression to replace multi-scale detection, so that the obtained tracking box is more accurate. However, the tracking frames obtained by the above network are all horizontal, and a large loss is generated when the target rotates. The SiamMask proposed by Wang et al adds a segmentation branch on the basis of SiamRPN, and realizes the combination of image segmentation and target tracking. The network first segments the target and then computes the minimum bounding box of the segmentation mask, which is used as the tracking box. However, the calculation method has a great defect that when the target moves, the tracking frame contains more backgrounds, and the direct calculation of the minimum circumscribing frame angle generates great deviation, which results in accuracy reduction and tracking drift.
Aiming at the problems, the invention provides a visual tracking method based on the foreground condition probability optimization scale and angle, which optimizes the tracking frame from the scale and the angle, improves the foreground proportion in the tracking frame and inhibits background interference.
Disclosure of Invention
The invention relates to a visual tracking method based on foreground condition probability optimization scale and angle aiming at inaccurate scale and angle of a tracking frame, aiming at solving the instability and inaccuracy of the scale and angle of the tracking frame caused by complex scenes such as movement, rotation, scale change and the like of a target.
The technical scheme of the invention is as follows:
a visual tracking method based on foreground condition probability optimization scale and angle comprises the following steps:
(1) reading a video frame sequence, and calculating a regression frame, a segmentation mask (foreground) and a minimum external frame of the frame by using a SimMask method;
(2) calculating the proportion of the foreground area in the minimum external frame;
(3) when the proportion is smaller than a set threshold value, calculating the reliability of the minimum external frame of the frame;
(4) selecting different strategies according to the reliability to optimize the minimum bounding box dimension;
(5) performing offset setting aiming at the angle of the tracking frame after the scale optimization;
(6) calculating IoU (Intersection over Union) values of each offset angle rotation frame and the foreground;
(7) the tracker adapts the rotation box that outputs the largest value of the foreground IoU.
In step 1, first, the vertex coordinates of the minimum bounding box M are obtained and recorded as:
M=[(xA,yA),(xB,yB),(xC,yC),(xD,yD)] (1)
in step 2, in the minimum external frame M, the ratio of the foreground F is:
Figure BDA0003042929110000021
f _ area is the area of foreground F, M _ area is the area of minimum bounding box M, MFIs the proportion of the foreground area to the minimum circumscribing frame.
Setting a threshold value rho to evaluate the foreground proportion when MFWhen the ratio of the foreground in the minimum external frame M is higher than the threshold value rho, the original scale of the minimum external frame M is directly output without scale optimization; otherwise, the next dimension optimization is needed.
In step 3, calculating the reliability of the minimum bounding box M of the frame can be divided into the following steps:
s3.1 the regression frame R is rotated to an angle consistent with the minimum external frame M and is marked as R'. The vertex coordinates of R' are noted as:
R′=[(xa,ya),(xb,yb),(xc,yc),(xd,yd)] (3)
s3.2, searching the target searching area of the current frame { (x)1,y1),(x2,y2),.....,(xn,yn) Denoted sample space S. When a certain sample point appears in the rectangle M, the trace frame A occurs, and when the sample point appears in the rectangle R', the trace frame B occurs.
First, under the condition that the tracking frame a occurs, the conditional probability of the tracking frame B occurrence is denoted as P1, specifically:
Figure BDA0003042929110000022
meanwhile, under the condition that the tracking frame B occurs, the conditional probability that the tracking frame a does not occur is recorded as P0, specifically:
Figure BDA0003042929110000031
wherein P (A), P (B), P (A &) B),
Figure BDA0003042929110000032
The probability of the occurrence of the tracking frame A, the probability of the occurrence of the tracking frame B, the probability of the simultaneous occurrence of the tracking frames A and B, and the probability of the occurrence of the tracking frame B but the non-occurrence of the tracking frame A are respectively represented by the following probability calculation formula:
Figure BDA0003042929110000033
the conditional probability P1 reflects the similarity of the two, and P0 reflects the difference of the two. When the P1 value is larger and the P0 value is smaller, the similarity between the M frame and the R' frame is larger, and the reliability of the M frame is higher; conversely, when the smaller the P1 value is or the larger the P0 value is, the greater the difference between the two values is, the lower the reliability of M at this time is, the thresholds β and α are set to be compared with the conditional probabilities P1 and P0, respectively, and the reliability of the minimum bounding box M is evaluated.
In step 4, two strategies are used to optimize the dimension of the minimum bounding box M:
case 1: when P1> beta and P0< alpha are simultaneously satisfied, the reliability of the minimum bounding box M is higher, and the intersection part of the two is taken as a new tracking box T:
T=M∩R′ (7)
case 2: when P1< β or P0> α, the reliability of the minimum bounding box M is low, redefining the scale as:
Figure BDA0003042929110000034
r ' _ w and R ' _ h are the width and height of R ', respectively, and M _ w and M _ h are the width and height of M, respectively, (in any rectangle, width < height). Setting the minimum value of the widths of M and R' as the width of a new tracking frame T, and recording the minimum value as T _ w; and setting the average value of the height of the two as the height thereof, and marking the height as T _ h, wherein the central coordinate and the angle are unchanged. Finally, outputting a tracking frame T after scale optimization:
T=[(x1,y1),(x2,y2),(x3,y3),(x4,y4)] (9)
further angle optimization is performed for T at steps 5-7.
In step 5, n offset thresholds with intervals μ are set clockwise and counterclockwise for the angle θ of the tracking frame T, i.e. a set of tracking frames { T of different angles is generated1(θ-nμ),...,Tn(θ-μ),Tn+1(θ),Tn+2(θ+μ),...,T2n+1(θ + n μ) }, from Ti(i ═ 1,2, …, n, …,2n +1) indicates that the next calculation is performed for a rotating frame at any angle in the set.
In step 6, each angular rotation frame is calculatedTiThe IoU value with foreground includes the following steps:
and S6.1, carrying out noise reduction on the frame of picture by adopting a mean shift algorithm, and carrying out smooth filtering on the image color level.
S6.2 rotating frames TiDrawing the edges of the rectangle and filling colors, setting the colors to be (255, 255, 255), and forming a distinguishing degree with other pixels;
s6.3, converting the RGB image of the frame into a gray-scale image, and giving different weights to three channels:
Gray=R*0.299+G*0.587+B*0.114 (10)
s6.4, carrying out binarization processing on the gray level image, converting the gray level image into a binary matrix consisting of 0 and 1, and recording the binary matrix as
Figure BDA0003042929110000047
1 is a rotating frame TiAll pixels in the pixel array, 0 is the other pixels.
S6.5 pairs of rotating frames TiAnd performing dot multiplication operation on the binary matrix of the foreground to obtain an intersection matrix:
Figure BDA0003042929110000041
MFPIa binary matrix of the foreground pixel map, which is the dot product of the corresponding element in the matrix, MIPIIs the intersection matrix of the two.
S6.6 the area of the matrix is:
Figure BDA0003042929110000042
h (X, Y) is the value at coordinate (X, Y), and X, Y are the rows and columns of the matrix.
S6.7, after the intersection is known, the IoU values of the two are:
Figure BDA0003042929110000043
arbitrary TiThe area is equal to T, denoted as T _ area,f _ area is the foreground area.
In step 7, the specific steps are as follows:
the rotating box with the maximum IoU value of the tracker output and the target foreground is recorded as
Figure BDA0003042929110000044
The formula is as follows:
Figure BDA0003042929110000045
Figure BDA0003042929110000046
namely, the final tracking frame obtained after the tracker is jointly optimized in scale and angle.
Compared with the prior art, the method provided by the invention has the advantages that:
(1) the defects existing in the combination of image segmentation and target tracking are analyzed and discovered.
(2) The accuracy of the tracking frame and the probability of the foreground are improved in the aspect of scale optimization.
(3) And for complex scenes such as severe target motion and the like, the angle of the tracking frame is still accurate.
Drawings
FIG. 1 is a schematic flow chart of the algorithm of the present invention;
fig. 2 shows that the video sets with three moving targets in the VOT2018 data set are selected for comparison by combining the comparison of the real frame and the SiamMask network tracking effect.
Detailed Description
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. It is to be understood that the described embodiments are merely a few embodiments of the invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. Referring to fig. 1, the example provided by the present invention uses the SiamMask tracking algorithm as a baseline tracker, and the specific implementation manner is as follows.
(1) Reading a video frame sequence, and calculating a regression frame, a segmentation mask (foreground) and a minimum external frame of the frame by using a SimMask method. Firstly, obtaining the vertex coordinates of the minimum external frame M, and recording as:
M=[(xA,yA),(xB,yB),(xC,yC),(xD,yD)] (1)
(2) in the minimum circumscribing frame M, the ratio of the foreground F is:
Figure BDA0003042929110000051
f _ area is the area of foreground F, M _ area is the area of minimum bounding box M, MFIs the proportion of the foreground area to the minimum circumscribing frame.
Set threshold ρ 0.9 to evaluate foreground proportion when MFWhen the ratio of the foreground in the minimum external frame M is higher than the threshold value rho, the original scale of the minimum external frame M is directly output without scale optimization; otherwise, the next dimension optimization is needed.
(3) And when the proportion is smaller than a set threshold value, calculating the reliability of the minimum external frame of the frame. The method comprises the following specific steps:
3.1 the regression frame R is rotated to an angle corresponding to the minimum circumscribing frame M and is denoted as R'. The vertex coordinates of R' are noted as:
R′=[(xa,ya),(xb,yb),(xc,yc),(xd,yd)] (3)
3.2 search area of current frame target { (x)1,y1),(x2,y2),.....,(xn,yn) Denoted sample space S. When a certain sample point appears in the rectangle M, the tracking frame A is formed, and when the certain sample point appears in the rectangle R', the tracking frame A is formedBlock B occurs.
First, under the condition that the tracking frame a occurs, the conditional probability of the tracking frame B occurrence is denoted as P1, specifically:
Figure BDA0003042929110000052
meanwhile, under the condition that the tracking frame B occurs, the conditional probability that the tracking frame a does not occur is recorded as P0, specifically:
Figure BDA0003042929110000053
wherein P (A), P (B), P (A &) B),
Figure BDA0003042929110000061
The probability of the occurrence of the tracking frame A, the probability of the occurrence of the tracking frame B, the probability of the simultaneous occurrence of the tracking frames A and B, and the probability of the occurrence of the tracking frame B but the non-occurrence of the tracking frame A are respectively represented by the following probability calculation formula:
Figure BDA0003042929110000062
the conditional probability P1 reflects the similarity of the two, and P0 reflects the difference of the two. When the P1 value is larger and the P0 value is smaller, the similarity between the M frame and the R' frame is larger, and the reliability of the M frame is higher; conversely, when the smaller the P1 value is or the larger the P0 value is, the greater the difference between the two values is, the lower the reliability of M is, the threshold β of 0.8 and α of 0.2 are set to be compared with the conditional probabilities P1 and P0, respectively, and the reliability of the minimum circumscribed frame M is evaluated.
(4) Selecting different strategies according to the reliability to optimize the minimum bounding box dimension, and totally dividing the minimum bounding box dimension into two strategies:
case 1: when P1> beta and P0< alpha are simultaneously satisfied, the reliability of the minimum bounding box M is higher, and the intersection part of the two is taken as a new tracking box T:
T=M∩R′ (7)
case 2: when P1< β or P0> α, the reliability of the minimum bounding box M is low, redefining the scale as:
Figure BDA0003042929110000063
r ' _ w and R ' _ h are the width and height of R ', respectively, and M _ w and M _ h are the width and height of M, respectively, (in any rectangle, width < height). Setting the minimum value of the widths of M and R' as the width of a new tracking frame T, and recording the minimum value as T _ w; and setting the average value of the height of the two as the height thereof, and marking the height as T _ h, wherein the central coordinate and the angle are unchanged. Finally, outputting a tracking frame T after scale optimization:
T=[(x1,y1),(x2,y2),(x3,y3),(x4,y4)] (9)
further angle optimization is performed on T in (5) - (7).
(5) And performing offset setting aiming at the angle of the tracking frame after the scale optimization. The angle theta of the tracking frame T is respectively shifted by 20 degrees clockwise and anticlockwise, and the generated tracking frame set T1(θ-20°),T2(θ),T3(θ +20 °), represented by Ti(i is 1,2,3) indicates that the next calculation is performed for a rotating frame at an arbitrary angle in the set.
(6) Calculating IoU values for each offset angle rotation frame and the foreground, comprising the steps of:
6.1, adopting a mean shift algorithm to perform noise reduction processing on the frame of picture, and performing smooth filtering on the image color level.
6.2 pairs of rotating frames TiDrawing the edges of the rectangle and filling colors, setting the colors to be (255, 255, 255), and forming a distinguishing degree with other pixels;
6.3, converting the RGB image of the frame into a gray map, and giving different weights to three channels:
Gray=R*0.299+G*0.587+B*0.114 (10)
6.4 binarization processing is carried out on the gray level image, the gray level image is converted into a binary matrix consisting of 0 and 1, and the binary matrix is marked as MTiPI1 is a rotating frame TiInner imagePixel 0 is the other pixel.
6.5 pairs of rotating frames TiAnd performing dot multiplication operation on the binary matrix of the foreground to obtain an intersection matrix:
Figure BDA0003042929110000071
MFPIa binary matrix of the foreground pixel map, which is the dot product of the corresponding element in the matrix, MIPIIs the intersection matrix of the two.
6.6 the matrix area is:
Figure BDA0003042929110000072
h (X, Y) is the value at coordinate (X, Y), and X, Y are the rows and columns of the matrix.
6.7 knowing the intersection, the IoU values for both are:
Figure BDA0003042929110000073
arbitrary TiThe area is equal to T, and is marked as T _ area, and F _ area is the foreground area.
(7) The rotation box with the maximum tracker adaptive output and foreground IoU values is recorded
Figure BDA0003042929110000074
The formula is as follows:
Figure BDA0003042929110000075
Figure BDA0003042929110000076
namely, the final tracking frame obtained after the tracker is jointly optimized in scale and angle.
In order to verify the overall tracking performance of the invention, the invention verifies the VOT data set of a single-target mainstream tracking platform, 60 videos on the VOT2018 data set, and overall evaluation is performed on three indexes of Robustness (Robustness), Accuracy (Accuracy) and Average Overlap Expectation (EAO).
In table 1, the comparison of the present invention with the SiamMask network under three indexes under the VOT2018 data set is shown, as shown in the table, the present invention improves the accuracy by about 3.6% and the EAO by about 1.9% compared to the baseline SiamMask network.
Table 1 comparison of VOT2018 dataset results
Figure BDA0003042929110000077
The above-mentioned embodiments only express several embodiments of the present invention, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the inventive concept, which falls within the scope of the present invention. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (5)

1. A visual tracking method based on foreground condition probability optimization scale and angle is characterized in that an original tracking frame is optimized from two aspects of scale and angle respectively, the foreground proportion in the tracking frame is improved, and steps 2-4 are scale optimization; the steps 5-7 are angle optimization, and the visual tracking method comprises the following specific steps:
step 1: reading a video frame sequence, and calculating a regression frame, a segmentation mask foreground and a minimum external frame of the frame by using a SimMask method;
step 2: calculating the proportion of the foreground area in the minimum external frame;
and step 3: when the proportion is smaller than a set threshold value, calculating the reliability of the minimum external frame of the frame;
and 4, step 4: selecting different strategies according to the reliability to optimize the minimum bounding box dimension;
and 5: performing offset setting aiming at the angle of the tracking frame after the scale optimization;
step 6: IoU values of each offset angle rotation frame and the foreground are calculated;
and 7: the tracker adaptively outputs the rotating frame with the maximum value of the foreground IoU;
the specific steps of step 3 are as follows: the reliability of the minimum external frame M of the frame is calculated, and the method can be divided into the following steps:
s3.1, rotating the regression frame R to form an angle consistent with the minimum external frame M, and recording the vertex coordinates of R' as:
R′=[(xa,ya),(xb,yb),(xc,yc),(xd,yd)] (3)
s3.2, searching the target searching area of the current frame { (x)1,y1),(x2,y2),.....,(xn,yn) Recording the result as a sample space S, and when a certain sample point appears in the minimum external frame M, the result is a tracking frame A, and when the certain sample point appears in a rectangle R', the result is a tracking frame B;
first, under the condition that the tracking frame a occurs, the conditional probability of the tracking frame B occurrence is denoted as P1, specifically:
Figure FDA0003546499920000011
meanwhile, under the condition that the tracking frame B occurs, the conditional probability that the tracking frame a does not occur is recorded as P0, specifically:
Figure FDA0003546499920000012
setting threshold values beta and alpha to be compared with conditional probabilities P1 and P0 respectively, and evaluating the reliability of the minimum external frame M;
the specific steps of step 4 are as follows: the scale of the minimum circumscribed frame M is optimized by two strategies:
case 1: when P1> beta and P0< alpha are simultaneously satisfied, the reliability of the minimum bounding box M is higher, and the intersection part of the two is taken as a new tracking box T:
T=M∩R′ (6)
case 2: when P1< β or P0> α, the reliability of the minimum bounding box M is low, redefining the scale as:
Figure FDA0003546499920000021
r '_ w and R' _ h are the width and height of R 'respectively, M _ w and M _ h are the width and height of M respectively, and the minimum value of the widths of M and R' is set as the width of a new tracking frame T and is recorded as T _ w; setting the height mean value of the two as the height, recording the height as T _ h, keeping the central coordinate and the angle unchanged, and finally outputting a scale-optimized tracking frame T:
T=[(x1,y1),(x2,y2),(x3,y3),(x4,y4)] (8)
further angle optimization is performed for T at steps 5-7.
2. The visual tracking method based on the foreground conditional probability optimization scale and angle as claimed in claim 1, wherein the specific steps of step 2 are as follows: firstly, obtaining the vertex coordinates of the minimum external frame M, and recording as:
M=[(xA,yA),(xB,yB),(xC,yC),(xD,yD)] (1)
in the minimum circumscribing frame M, the ratio of the foreground F is:
Figure FDA0003546499920000022
f _ area is the area of foreground F, M _ area is the area of minimum bounding box M, MFThe ratio of the foreground area to the minimum circumscribing frame;
setting a threshold value rho to evaluate the foreground proportion when MFWhen the ratio of the foreground in the minimum external frame M is higher than the threshold value rho, the original scale of the minimum external frame M is directly output without scale optimization; otherwise, the next dimension optimization is needed.
3. The visual tracking method based on the foreground conditional probability optimization scale and angle as claimed in claim 1, wherein the specific steps of step 5 are as follows: setting n offset thresholds with interval of mu for angle theta of tracking frame T clockwise and anticlockwise respectively, namely generating tracking frame set { T of different angles1(θ-nμ),...,Tn(θ-μ),Tn+1(θ),Tn+2(θ+μ),...,T2n+1(θ + n μ) }, from Ti(i ═ 1,2, …, n, …,2n +1) indicates that the next calculation is performed for a rotating frame at any angle in the set.
4. The visual tracking method based on the foreground conditional probability optimization scale and angle as claimed in claim 1, wherein the calculation of each angle rotation frame T of step 6iThe IoU value with foreground includes the following steps:
s6.1, carrying out noise reduction on the frame of picture by adopting a mean shift algorithm, and carrying out smooth filtering on the image color level;
s6.2 rotating frames TiDrawing the edges of the rectangle and filling colors, setting the colors to be (255, 255, 255), and forming a distinguishing degree with other pixels;
s6.3, converting the RGB image of the frame into a gray-scale image, and giving different weights to three channels:
Gray=R*0.299+G*0.587+B*0.114 (9)
s6.4, carrying out binarization processing on the gray level image, converting the gray level image into a binary matrix consisting of 0 and 1, and recording the binary matrix as
Figure FDA0003546499920000031
1 is a rotating frame TiAll pixels in the pixel array, 0 is other pixels;
s6.5 pairs of rotating frames TiAnd two of the foregroundPerforming dot multiplication operation on the value matrix to obtain an intersection matrix:
Figure FDA0003546499920000032
MFPIa binary matrix of the foreground pixel map, which is the dot product of the corresponding element in the matrix, MIPIIs the intersection matrix of the two;
s6.6 the area of the matrix is:
Figure FDA0003546499920000033
h (X, Y) is the value at coordinate (X, Y), X, Y are the rows and columns of the matrix;
s6.7, after the intersection is known, the IoU values of the two are:
Figure FDA0003546499920000034
arbitrary TiThe area is equal to T, and is marked as T _ area, and F _ area is the foreground area.
5. The visual tracking method based on the foreground conditional probability optimization scale and angle as claimed in claim 3, wherein the specific steps of step 7 are as follows:
the rotating box with the maximum IoU value of the tracker output and the target foreground is recorded as
Figure FDA0003546499920000035
The formula is as follows:
Figure FDA0003546499920000036
Figure FDA0003546499920000037
namely, the final tracking frame obtained after the tracker is jointly optimized in scale and angle.
CN202110462758.7A 2021-04-28 2021-04-28 Vision tracking method based on foreground condition probability optimization scale and angle Active CN112991395B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110462758.7A CN112991395B (en) 2021-04-28 2021-04-28 Vision tracking method based on foreground condition probability optimization scale and angle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110462758.7A CN112991395B (en) 2021-04-28 2021-04-28 Vision tracking method based on foreground condition probability optimization scale and angle

Publications (2)

Publication Number Publication Date
CN112991395A CN112991395A (en) 2021-06-18
CN112991395B true CN112991395B (en) 2022-04-15

Family

ID=76340400

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110462758.7A Active CN112991395B (en) 2021-04-28 2021-04-28 Vision tracking method based on foreground condition probability optimization scale and angle

Country Status (1)

Country Link
CN (1) CN112991395B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993775A (en) * 2019-04-01 2019-07-09 云南大学 Monotrack method based on feature compensation
CN111383244A (en) * 2020-02-28 2020-07-07 浙江大华技术股份有限公司 Target detection tracking method
CN112132855A (en) * 2020-09-22 2020-12-25 山东工商学院 Self-adaptive Gaussian function target tracking method based on foreground segmentation guidance
CN112258558A (en) * 2020-10-23 2021-01-22 复旦大学 Target tracking method based on multi-scale twin network, electronic device and medium
CN112509008A (en) * 2020-12-15 2021-03-16 重庆邮电大学 Target tracking method based on intersection-to-parallel ratio guided twin network

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111445497B (en) * 2020-02-25 2021-03-30 华中科技大学 Target tracking and following method based on scale context regression
CN111968155B (en) * 2020-07-23 2022-05-17 天津大学 Target tracking method based on segmented target mask updating template

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993775A (en) * 2019-04-01 2019-07-09 云南大学 Monotrack method based on feature compensation
CN111383244A (en) * 2020-02-28 2020-07-07 浙江大华技术股份有限公司 Target detection tracking method
CN112132855A (en) * 2020-09-22 2020-12-25 山东工商学院 Self-adaptive Gaussian function target tracking method based on foreground segmentation guidance
CN112258558A (en) * 2020-10-23 2021-01-22 复旦大学 Target tracking method based on multi-scale twin network, electronic device and medium
CN112509008A (en) * 2020-12-15 2021-03-16 重庆邮电大学 Target tracking method based on intersection-to-parallel ratio guided twin network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
A Novel Incremental Multi-Template Update Strategy for Robust Object Tracking;A. Zhiyong 等;《IEEE Access》;20200904;第8卷;全文 *

Also Published As

Publication number Publication date
CN112991395A (en) 2021-06-18

Similar Documents

Publication Publication Date Title
CN109241913B (en) Ship detection method and system combining significance detection and deep learning
CN107194408B (en) Target tracking method of mixed block sparse cooperation model
CN107424171B (en) Block-based anti-occlusion target tracking method
CN108062525B (en) Deep learning hand detection method based on hand region prediction
CN110473231B (en) Target tracking method of twin full convolution network with prejudging type learning updating strategy
CN110766724B (en) Target tracking network training and tracking method and device, electronic equipment and medium
CN110866924A (en) Line structured light center line extraction method and storage medium
CN108320306B (en) Video target tracking method fusing TLD and KCF
CN106991686B (en) A kind of level set contour tracing method based on super-pixel optical flow field
CN109993052B (en) Scale-adaptive target tracking method and system under complex scene
CN111160212B (en) Improved tracking learning detection system and method based on YOLOv3-Tiny
CN111612817A (en) Target tracking method based on depth feature adaptive fusion and context information
CN109035300B (en) Target tracking method based on depth feature and average peak correlation energy
CN112364865B (en) Method for detecting small moving target in complex scene
CN110555868A (en) method for detecting small moving target under complex ground background
CN112270697B (en) Satellite sequence image moving target detection method combined with super-resolution reconstruction
CN111951297A (en) Target tracking method based on structured pixel-by-pixel target attention mechanism
CN109949344B (en) Nuclear correlation filtering tracking method based on color probability target suggestion window
CN109978916B (en) Vibe moving target detection method based on gray level image feature matching
CN108846845B (en) SAR image segmentation method based on thumbnail and hierarchical fuzzy clustering
CN108921872B (en) Robust visual target tracking method suitable for long-range tracking
Sun et al. Adaptive image dehazing and object tracking in UAV videos based on the template updating Siamese network
CN112991395B (en) Vision tracking method based on foreground condition probability optimization scale and angle
JP3716455B2 (en) Region extraction method and region extraction device
CN110322479B (en) Dual-core KCF target tracking method based on space-time significance

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant