CN113298053A - Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium - Google Patents

Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN113298053A
CN113298053A CN202110845085.3A CN202110845085A CN113298053A CN 113298053 A CN113298053 A CN 113298053A CN 202110845085 A CN202110845085 A CN 202110845085A CN 113298053 A CN113298053 A CN 113298053A
Authority
CN
China
Prior art keywords
tracking target
tracking
information
unmanned aerial
aerial vehicle
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110845085.3A
Other languages
Chinese (zh)
Other versions
CN113298053B (en
Inventor
张迪凯
杨鹏
王豪
潘明锋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ji Hua Laboratory
Original Assignee
Ji Hua Laboratory
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ji Hua Laboratory filed Critical Ji Hua Laboratory
Priority to CN202110845085.3A priority Critical patent/CN113298053B/en
Publication of CN113298053A publication Critical patent/CN113298053A/en
Application granted granted Critical
Publication of CN113298053B publication Critical patent/CN113298053B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a multi-target unmanned aerial vehicle tracking identification method, a multi-target unmanned aerial vehicle tracking identification device, electronic equipment and a storage medium, wherein image information acquired by a binocular camera is periodically acquired; inputting current image information into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; calculating the prediction state information of the current moment according to the actual measurement state information acquired at the last moment of each tracking target; matching each tracking target at the current moment with each tracking target at the previous moment according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information and position information of points in a corresponding tracking target frame as distance information and position information of each tracking target; updating the tracking identification information of each tracking target; therefore, the tracking identification of the multiple unmanned aerial vehicles can be accurately carried out.

Description

Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium
Technical Field
The invention relates to the technical field of image processing, in particular to a multi-target unmanned aerial vehicle tracking and identifying method, a multi-target unmanned aerial vehicle tracking and identifying device, electronic equipment and a storage medium.
Background
Because the unmanned aerial vehicle technique constantly promotes and becomes ripe day by day, unmanned aerial vehicle becomes a novel intelligence equipment gradually, because its dexterity, little volume, can realize that the orbit is patrol, and fixed point hovers etc. characteristics are used more and more. In daily life, the phenomenon that the unmanned aerial vehicle flies black sometimes happens, so that the privacy of people is easily invaded, and a plurality of dangerous factors are added for daily travel. Therefore, how to deal with the abnormal flight of the unmanned aerial vehicle becomes a central part of the current society.
In order to take countermeasures for an unmanned aerial vehicle flying abnormally, the unmanned aerial vehicle needs to be tracked and identified firstly, and currently, the unmanned aerial vehicle is generally tracked and identified by adopting a radar perception and video monitoring mode. The radar cannot distinguish targets such as unmanned planes, balloons and birds, and is easily interfered by the environment, so that a large amount of false detection results are easily brought by adopting a radar sensing mode; the video monitoring method requires the staff to undertake all the work tasks, and the method is labor-consuming and unstable.
In addition, when the unmanned aerial vehicle of unusual flight has a plurality ofly, adopt traditional radar perception and video monitoring's mode to be difficult to effectively carry out accurate tracking discernment to a plurality of unmanned aerial vehicles more.
Disclosure of Invention
In view of the defects of the prior art, an object of the embodiment of the present application is to provide a method, an apparatus, an electronic device and a storage medium for tracking and identifying a multi-target unmanned aerial vehicle, which can accurately track and identify a plurality of unmanned aerial vehicles.
In a first aspect, an embodiment of the present application provides a multi-target unmanned aerial vehicle tracking and identifying method, including the steps of:
A1. periodically acquiring image information collected by a binocular camera, the image information including depth information;
A2. inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; the target detection image comprises a plurality of tracking target frames, each tracking target frame corresponds to one tracking target, and the corresponding tracking target is surrounded; the tracking target is an unmanned aerial vehicle;
A3. acquiring actual measurement state information of each tracking target at the current moment according to the target detection image;
A4. based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment;
A5. matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment;
A6. acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target;
A7. and integrating the identity information of each tracking target with the corresponding distance information and position information, and updating the tracking identification information of each tracking target.
The embodiment of the application provides a multi-target unmanned aerial vehicle tracking identification method, a plurality of unmanned aerial vehicles in the image are identified through the deep learning model, compare with radar perception mode, better anti-interference performance has, recognition effect is better, the state at the moment is predicated by unmanned aerial vehicle last quarter's state through the Kalman filtering prediction method, and match with the actual measurement result of moment, can realize the definite to the identity of the unmanned aerial vehicle who discerns, thereby can integrate each unmanned aerial vehicle's that discerns state information under the identity that corresponds, with the tracking identification information of updating each unmanned aerial vehicle, realize a plurality of unmanned aerial vehicle's accurate tracking.
Preferably, the deep learning model is a deep learning model based on a YOLOV3 neural network and a Deepsort algorithm; during operation, the deep learning model firstly performs 2 times of upsampling on a characteristic graph which is output by a YOLOV3 framework and is subjected to 8 times of downsampling, the 2 times of upsampling characteristic graph is spliced with a4 times of downsampling characteristic graph output by a YOLOV3 framework, a fusion information layer which is 4 times of downsampling is output, then the fusion information layer which is 4 times of downsampling is performed again, and the output of the fusion information layer is spliced with 2 times of downsampling of a YOLOV3 framework.
Preferably, the deep learning model is trained by the following training process:
s1, data acquisition: collecting a plurality of unmanned aerial vehicle pictures, and processing the unmanned aerial vehicle pictures into the same size;
s2, data annotation: marking the unmanned aerial vehicle picture;
s3, data enhancement: the method comprises the steps that a randomly selected preset blocking block is adopted to block collected unmanned aerial vehicle pictures to generate new unmanned aerial vehicle pictures, and/or each pixel point of two unmanned aerial vehicle pictures is linearly superposed to obtain new unmanned aerial vehicle pictures, and the collected unmanned aerial vehicle pictures and the new unmanned aerial vehicle pictures are taken as training sets;
s4, training a model: and training the deep learning model by utilizing the training set.
Preferably, in step S3, the step of linearly superimposing the pixel points of the two unmanned aerial vehicle pictures to obtain a new unmanned aerial vehicle picture includes:
by means of arrangement and combination, two different unmanned aerial vehicle pictures are used as a group to obtain a plurality of unmanned aerial vehicle picture combinations;
obtaining a new unmanned aerial vehicle picture aiming at each unmanned aerial vehicle picture combination, wherein the pixel value of each pixel point of the new unmanned aerial vehicle picture is obtained by calculation according to the following formula:
Figure 283484DEST_PATH_IMAGE001
;
wherein the content of the first and second substances,
Figure 525109DEST_PATH_IMAGE002
for the pixel value of the ith pixel point in the new picture of the unmanned aerial vehicle,
Figure 243974DEST_PATH_IMAGE003
the pixel value of the ith pixel point of the first unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination,
Figure 621865DEST_PATH_IMAGE004
the pixel value of the ith pixel point of the second unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination is obtained;
Figure 222611DEST_PATH_IMAGE005
is a scaling factor.
Preferably, step a2 includes:
A201. if the current time is the starting time of the current calculation cycle, then:
amplifying an original image acquired at the current moment according to a preset proportion to obtain amplified image information;
respectively inputting the original image information and the amplified image information acquired at the current moment into a pre-trained deep learning model to obtain two target detection images;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is more than that in the target detection image corresponding to the original image information, determining the amplified image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is not more than that of the tracking targets in the target detection image corresponding to the original image information, determining the original image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
A202. and if the current time is not the starting time of the current calculation period, inputting the corresponding effective input image information into a pre-trained deep learning model to obtain an effective target detection image.
Preferably, step a4 includes:
calculating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 623637DEST_PATH_IMAGE006
;
Figure 968030DEST_PATH_IMAGE007
;
Figure 652958DEST_PATH_IMAGE008
;
Figure 159026DEST_PATH_IMAGE009
;
Figure 312927DEST_PATH_IMAGE010
;
wherein the content of the first and second substances,
Figure 195432DEST_PATH_IMAGE011
to track the predicted state information of the target at the current time,
Figure 547916DEST_PATH_IMAGE012
Figure 411836DEST_PATH_IMAGE013
Figure 849770DEST_PATH_IMAGE014
Figure 473650DEST_PATH_IMAGE015
Figure 680640DEST_PATH_IMAGE016
Figure 981040DEST_PATH_IMAGE017
Figure 906271DEST_PATH_IMAGE018
Figure 130579DEST_PATH_IMAGE019
respectively an abscissa value predicted value of the center of a tracking target frame, an ordinate value predicted value of the center of the tracking target frame, a width predicted value of the tracking target frame, a height predicted value of the tracking target frame, a lateral velocity predicted value of the center of the tracking target frame, a longitudinal velocity predicted value of the center of the tracking target frame, a width change velocity predicted value of the tracking target frame, and a height change velocity predicted value of the tracking target frame at the current time, wherein x is the predicted state information of the tracking target at the previous time,
Figure 395338DEST_PATH_IMAGE020
Figure 679689DEST_PATH_IMAGE021
Figure 16517DEST_PATH_IMAGE022
Figure 44516DEST_PATH_IMAGE023
Figure 226098DEST_PATH_IMAGE024
Figure 619034DEST_PATH_IMAGE025
Figure 518857DEST_PATH_IMAGE026
Figure 271918DEST_PATH_IMAGE027
the predicted value of the horizontal coordinate value of the center of the tracking target frame, the predicted value of the vertical coordinate value of the center of the tracking target frame, the predicted value of the width of the tracking target frame, the predicted value of the height of the tracking target frame, the predicted value of the horizontal velocity of the center of the tracking target frame, the predicted value of the vertical velocity of the center of the tracking target frame, the predicted value of the width change speed of the tracking target frame, and the height change of the tracking target frame at the previous moment are respectivelyThe predicted value of the speed is converted into the speed,
Figure 308007DEST_PATH_IMAGE028
the state covariance of the tracked target at the current moment is P, the state covariance of the tracked target at the last moment is P, Q is a noise matrix coefficient of the system, and F is a conversion matrix.
Preferably, step a5 is followed by:
A8. updating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 871843DEST_PATH_IMAGE029
;
Figure 993383DEST_PATH_IMAGE030
;
Figure 550135DEST_PATH_IMAGE031
;
Figure 706310DEST_PATH_IMAGE032
;
Figure 503365DEST_PATH_IMAGE033
;
Figure 49884DEST_PATH_IMAGE034
;
wherein z is the actual measurement state information of the tracking target at the current moment,
Figure 957797DEST_PATH_IMAGE035
Figure 155429DEST_PATH_IMAGE036
Figure 123385DEST_PATH_IMAGE037
Figure 219517DEST_PATH_IMAGE038
Figure 868804DEST_PATH_IMAGE039
Figure 999571DEST_PATH_IMAGE040
Figure 328309DEST_PATH_IMAGE041
Figure 646158DEST_PATH_IMAGE042
respectively an abscissa value measured value of the center of the tracking target frame, an ordinate value measured value of the center of the tracking target frame, a width measured value of the tracking target frame, a height measured value of the tracking target frame, a lateral velocity measured value of the center of the tracking target frame, a longitudinal velocity measured value of the center of the tracking target frame, a width change velocity measured value of the tracking target frame, a height change velocity measured value of the tracking target frame, y is a residual error, H is an observation matrix, R is a noise matrix, S is a process variable, K is a Kalman gain,
Figure 99136DEST_PATH_IMAGE043
is updated
Figure 84409DEST_PATH_IMAGE011
,
Figure 394168DEST_PATH_IMAGE044
Is updated
Figure 386264DEST_PATH_IMAGE028
,
Figure 439670DEST_PATH_IMAGE045
Is an identity matrix.
In a second aspect, an embodiment of the present application provides a multi-target unmanned aerial vehicle tracks recognition device, includes:
the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for periodically acquiring image information acquired by a binocular camera, and the image information comprises depth information;
the detection module is used for inputting the image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; the target detection image comprises a plurality of tracking target frames, each tracking target frame corresponds to one tracking target, and the corresponding tracking target is surrounded; the tracking target is an unmanned aerial vehicle;
the second acquisition module is used for acquiring the actual measurement state information of each tracking target at the current moment according to the target detection image;
the prediction module is used for calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target obtained at the previous moment based on a Kalman filtering prediction method;
the matching module is used for matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment;
the third acquisition module is used for acquiring the depth information of the midpoint of the tracking target frame of each tracking target as the distance information of each tracking target and acquiring the position information of the midpoint of the tracking target frame of each tracking target as the position information of each tracking target;
and the integration module is used for integrating the identity information of each tracking target and the corresponding distance information and position information and updating the tracking identification information of each tracking target.
In a third aspect, an embodiment of the present application provides an electronic device, including a processor and a memory, where the memory stores a computer program, and the processor is configured to execute the steps of the multi-target drone tracking and identifying method by calling the computer program stored in the memory.
In a fourth aspect, the present application provides a storage medium, on which a computer program is stored, where the computer program runs the steps of the multi-target drone tracking and identifying method as described above when executed by a processor.
Has the advantages that:
according to the multi-target unmanned aerial vehicle tracking and identifying method and device, the electronic equipment and the storage medium, the image information collected by the binocular camera is periodically acquired, and the image information comprises depth information; inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment; matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target; integrating the identity information of each tracking target with corresponding distance information and position information, and updating the tracking identification information of each tracking target; therefore, the tracking identification of the multiple unmanned aerial vehicles can be accurately carried out.
Drawings
Fig. 1 is a flowchart of a multi-target unmanned aerial vehicle tracking and identifying method provided in an embodiment of the present application.
Fig. 2 is a schematic structural diagram of the multi-target unmanned aerial vehicle tracking and identifying device provided in the embodiment of the present application.
Fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Fig. 4 is a network structure diagram of a deep learning model in the embodiment of the present application.
Fig. 5 is a detection effect diagram of the deep learning model in the embodiment of the present application.
Fig. 6 is a diagram of the detection effect of the conventional YOLOV4 network model.
FIG. 7 is a new unmanned aerial vehicle picture obtained by blocking through a blocking block in the embodiment of the present application.
Fig. 8 is a new unmanned aerial vehicle picture obtained by linearly superimposing pixel points in the embodiment of the present application.
Fig. 9 is a target detection image finally displayed in the embodiment of the present application.
Fig. 10 is a structural diagram of a conventional YOLOV3 framework.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the accompanying drawings are illustrative only for the purpose of explaining the present invention, and are not to be construed as limiting the present invention.
In the description of the present invention, it is to be understood that the terms "center", "longitudinal", "lateral", "length", "width", "thickness", "upper", "lower", "front", "rear", "left", "right", "vertical", "horizontal", "top", "bottom", "inner", "outer", "clockwise", "counterclockwise", and the like, indicate orientations and positional relationships based on those shown in the drawings, and are used only for convenience of description and simplicity of description, and do not indicate or imply that the device or element being referred to must have a particular orientation, be constructed and operated in a particular orientation, and thus, should not be considered as limiting the present invention. Furthermore, the terms "first", "second" and "first" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, features defined as "first", "second", may explicitly or implicitly include one or more of the described features. In the description of the present invention, "a plurality" means two or more unless specifically defined otherwise.
The following disclosure provides embodiments or examples for implementing different configurations of the invention. To simplify the disclosure of the present invention, the components and arrangements of specific examples are described below. Of course, they are merely examples and are not intended to limit the present invention. Furthermore, the present invention may repeat reference numerals and/or letters in the various examples, such repetition is for the purpose of simplicity and clarity and does not in itself dictate a relationship between the various embodiments and/or configurations discussed. In addition, the present invention provides examples of various specific processes and materials, but those of ordinary skill in the art will recognize applications of other processes and/or uses of other materials.
Referring to fig. 1, a method for tracking and identifying a multi-target unmanned aerial vehicle provided in an embodiment of the present application includes:
A1. periodically acquiring image information collected by a binocular camera, the image information including depth information;
A2. inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; the target detection image comprises a plurality of tracking target frames, each tracking target frame corresponds to one tracking target, and the corresponding tracking target is surrounded; the tracking target is an unmanned aerial vehicle;
A3. acquiring actual measurement state information of each tracking target at the current moment according to the target detection image;
A4. based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment;
A5. matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment;
A6. acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target;
A7. and integrating the identity information of each tracking target with the corresponding distance information and position information, and updating the tracking identification information of each tracking target.
This multi-target unmanned aerial vehicle tracks recognition method, a plurality of unmanned aerial vehicles in to the image are discerned through the degree of depth learning model, compare with the radar perception mode among the prior art, better anti-interference performance has, the recognition effect is better, can accurately realize the discernment of a plurality of unmanned aerial vehicles in the image, state prediction current moment by unmanned aerial vehicle last moment through the Kalman filtering prediction method, and match with the actual measurement result of current moment with the prediction result, can realize the definite to the identity of the unmanned aerial vehicle who discerns, thereby can integrate the state information of each unmanned aerial vehicle who discerns under the identity that corresponds, with the tracking identification information who updates each unmanned aerial vehicle, realize the accurate tracking of a plurality of unmanned aerial vehicles.
The image information not in A1 is collected by the binocular camera, so that the obtained image information contains depth information, and the distance information between each unmanned aerial vehicle and the binocular camera can be conveniently acquired in the subsequent steps. Wherein, the binocular camera may be a ZED binocular camera, but is not limited thereto.
Preferably, the deep learning model is a deep learning model based on a YOLOV3 neural network and a Deepsort algorithm; during operation, the deep learning model firstly performs 2 times of upsampling on a characteristic graph which is output by a YOLOV3 framework and is subjected to 8 times of downsampling, the 2 times of upsampling characteristic graph is spliced with a4 times of downsampling characteristic graph output by a YOLOV3 framework, a fusion information layer which is 4 times of downsampling is output, then the fusion information layer which is 4 times of downsampling is performed again, and the output of the fusion information layer is spliced with 2 times of downsampling of a YOLOV3 framework.
Specifically, see fig. 4, which is a network structure diagram of the deep learning model, the network structure mainly includes a convolutional layer, a sampling layer, and a residual structure, and finally 5 detection layers are formed to detect an object from different dimensions.
The network structure of the conventional YOLOV3 frame is as shown in fig. 10, the detection of a tiny target is performed on an 8-time down-sampled output characteristic diagram, and when a target pixel is smaller than 8X8 pixel, the detection task cannot ensure accuracy, so the detection effect on the small target is not good. The conventional YOLOV4 network structure has a good detection effect on different types of targets, but the application mainly aims to identify a single type of unmanned aerial vehicle, so that a plurality of modules in the YOLOV4 network structure are redundant, only the training time and the reasoning time are increased, and the identification effect on the single type of targets is poorer than that of the deep learning model of the application; if fig. 5 is an effect graph of adopting the deep learning model of this application to identify, fig. 6 is an effect graph of adopting conventional YOLOV4 network model to identify the same image, and each tracked target frame in the graph represents an unmanned aerial vehicle that is identified, and it can be seen that the deep learning model of this application can identify more unmanned aerial vehicles, and its recognition effect to single kind of target is better.
In some embodiments, the deep learning model is trained by the following training process:
s1, data acquisition: collecting a plurality of unmanned aerial vehicle pictures, and processing the unmanned aerial vehicle pictures into the same size;
s2, data annotation: marking the unmanned aerial vehicle picture;
s3, data enhancement: the method comprises the steps that a randomly selected preset blocking block is adopted to block collected unmanned aerial vehicle pictures to generate new unmanned aerial vehicle pictures, and/or each pixel point of two unmanned aerial vehicle pictures is linearly superposed to obtain new unmanned aerial vehicle pictures, and the collected unmanned aerial vehicle pictures and the new unmanned aerial vehicle pictures are taken as training sets;
s4, training a model: and training the deep learning model by utilizing the training set.
In step S1, the drone image may be processed to a size of 608X608 (pixels), and the processing may specifically be performed by stretching, zooming, and the like.
In step S2, labeling may be performed by a labelImg-master, but is not limited thereto.
Preferably, in step S3, the step of linearly superimposing the pixel points of the two unmanned aerial vehicle pictures to obtain a new unmanned aerial vehicle picture includes:
by means of arrangement and combination, two different unmanned aerial vehicle pictures are used as a group to obtain a plurality of unmanned aerial vehicle picture combinations;
obtaining a new unmanned aerial vehicle picture aiming at each unmanned aerial vehicle picture combination, wherein the pixel value of each pixel point of the new unmanned aerial vehicle picture is obtained by calculation according to the following formula:
Figure 217133DEST_PATH_IMAGE001
;
wherein the content of the first and second substances,
Figure 697793DEST_PATH_IMAGE002
for the pixel value of the ith pixel point in the new picture of the unmanned aerial vehicle,
Figure 255814DEST_PATH_IMAGE003
the pixel value of the ith pixel point of the first unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination,
Figure 34283DEST_PATH_IMAGE004
the pixel value of the ith pixel point of the second unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination is obtained;
Figure 994148DEST_PATH_IMAGE005
is a scaling factor.
For example, in fig. 8, a new picture of the unmanned aerial vehicle is obtained by linearly superimposing two pictures, where each of the two original pictures of the unmanned aerial vehicle has one unmanned aerial vehicle.
Further, in step S3, after obtaining the new picture of the unmanned aerial vehicle, it may be determined whether the new picture of the unmanned aerial vehicle is qualified, and if not, the new picture of the unmanned aerial vehicle is deleted. Wherein, judge whether qualified judgement condition of new unmanned aerial vehicle picture is: and if the overlapping part of the unmanned aerial vehicle is less than 30%, judging that the picture is qualified, otherwise, judging that the picture is unqualified.
Preferably, in step S3, the step of blocking the acquired picture of the unmanned aerial vehicle by using the randomly selected preset blocking block to generate a new picture of the unmanned aerial vehicle includes:
randomly selecting a numerical value in a preset numerical value range (the range is [1, N ], and N is a positive integer larger than 1) as the number of the shielding objects;
and randomly selecting a corresponding number of shelters from a preset shelter database according to the number of the shelters, and adding the shelters into the unmanned aerial vehicle picture for sheltering.
The preset shelter database is stored with a plurality of shelter block images with different shapes and/or different sizes in advance. When adding the blocking object into the unmanned aerial vehicle picture, the center of the blocking object is arranged in the tracking target frame area (a pixel point position can be randomly selected in the tracking target frame area, and the center of the blocking object is arranged at the pixel point position). For example, in fig. 7, for a new picture of the unmanned aerial vehicle obtained after the blocking, a rectangular blocking block and an oval blocking block are adopted for blocking.
By the data enhancement processing in step S3, the sample size of the training set can be increased significantly, thereby avoiding the situation of insufficient sample size and insufficient data complexity of the training set. Compared with the traditional cutting and splicing method, the mode of linearly superposing the pixel points of the two unmanned aerial vehicle pictures to obtain the new unmanned aerial vehicle picture avoids the problems of unnatural splicing effect and excessively harsh edge details.
In some embodiments, in step S4, the learning rate is set to 0.0013 and the number of training times is 10000, but the method is not limited thereto.
Preferably, the deep learning model further comprises the steps of, through the following training process:
s5, adjusting parameters: obtaining the prior frame size of the training set by using K-means + + clustering, and modifying the prior frame size;
s6, repeatedly executing the steps S4 and S5 (the repeated times are preset according to actual needs).
In some preferred embodiments, step a2 includes:
A201. if the current time is the starting time of the current calculation cycle, then:
amplifying an original image acquired at the current moment according to a preset proportion to obtain amplified image information;
respectively inputting the original image information and the amplified image information acquired at the current moment into a pre-trained deep learning model to obtain two target detection images;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is more than that in the target detection image corresponding to the original image information, determining the amplified image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is not more than that of the tracking targets in the target detection image corresponding to the original image information, determining the original image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
A202. and if the current time is not the starting time of the current calculation period, inputting the corresponding effective input image information into a pre-trained deep learning model to obtain an effective target detection image.
The calculation period is a preset time period, which is longer than a sampling period for acquiring the image information acquired by the binocular camera, and for example, the calculation period is 10s, but is not limited thereto. At the initial moment of each calculation period, inputting original image information and amplified image information into a depth learning model respectively to obtain two target detection images, wherein if the number of unmanned aerial vehicles identified in an identification result corresponding to the original image information is more, the identification by using the original image information is more accurate, so that the original image information is adopted for identification in the period; if the number of the unmanned aerial vehicles identified in the identification result corresponding to the amplified image information is larger, the identification by the amplified image information is more accurate, and therefore the amplified image information is adopted for identification in the period. Therefore, the identification efficiency and accuracy are greatly improved, and meanwhile, the real-time calculation time is saved, and the stability of the frame number is ensured as far as possible.
The preset magnification ratio can be set according to actual needs, for example, the size of the original picture is 608X608 (pixels), and the magnified size is 1024X 1024 (pixels).
In this embodiment, the state information of the tracking target includes an abscissa value of the center of the tracking target frame, an ordinate value of the center of the tracking target frame, a width of the tracking target frame (length in the abscissa axis direction), a height of the tracking target frame (length in the ordinate axis direction), a lateral velocity of the center of the tracking target frame, a longitudinal velocity of the center of the tracking target frame, a width change velocity of the tracking target frame, and a height change velocity of the tracking target frame. The horizontal speed can be obtained by calculating the horizontal coordinate value at the current moment, the horizontal coordinate value at the last moment and the sampling period, the longitudinal speed can be obtained by calculating the vertical coordinate value at the current moment, the vertical coordinate value at the last moment and the sampling period, the width change speed can be obtained by calculating the width at the current moment, the width at the last moment and the sampling period, and the height change speed can be obtained by calculating the height at the current moment, the height at the last moment and the sampling period.
Accordingly, the actual measurement state information in step a3 includes an abscissa value actual measurement value of the tracking target frame center, an ordinate value actual measurement value of the tracking target frame center, a tracking target frame width actual measurement value, a tracking target frame height actual measurement value, a lateral velocity actual measurement value of the tracking target frame center, a longitudinal velocity actual measurement value of the tracking target frame center, a width change velocity actual measurement value of the tracking target frame, and a height change velocity actual measurement value of the tracking target frame.
Preferably, step a4 includes:
calculating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 583393DEST_PATH_IMAGE006
;
Figure 363130DEST_PATH_IMAGE007
;
Figure 945290DEST_PATH_IMAGE008
;
Figure 759662DEST_PATH_IMAGE009
;
Figure 582125DEST_PATH_IMAGE010
;
wherein the content of the first and second substances,
Figure 786841DEST_PATH_IMAGE011
to track the predicted state information of the target at the current time,
Figure 985741DEST_PATH_IMAGE012
Figure 844500DEST_PATH_IMAGE013
Figure 837864DEST_PATH_IMAGE014
Figure 592193DEST_PATH_IMAGE015
Figure 532468DEST_PATH_IMAGE016
Figure 55853DEST_PATH_IMAGE017
Figure 407069DEST_PATH_IMAGE018
Figure 648694DEST_PATH_IMAGE019
the predicted value of the horizontal coordinate value of the center of the tracking target frame, the predicted value of the vertical coordinate value of the center of the tracking target frame, the predicted value of the width of the tracking target frame and the height of the tracking target frame at the current moment are respectivelyDegree predicted value, transverse velocity predicted value of tracking target frame center, longitudinal velocity predicted value of tracking target frame center, width change velocity predicted value of tracking target frame, height change velocity predicted value of tracking target frame, x is predicted state information of tracking target at last moment,
Figure 189397DEST_PATH_IMAGE020
Figure 504972DEST_PATH_IMAGE021
Figure 105717DEST_PATH_IMAGE022
Figure 756010DEST_PATH_IMAGE023
Figure 100404DEST_PATH_IMAGE024
Figure 270485DEST_PATH_IMAGE025
Figure 42132DEST_PATH_IMAGE026
Figure 179721DEST_PATH_IMAGE027
respectively, an abscissa value predicted value of the center of the tracking target frame, an ordinate value predicted value of the center of the tracking target frame, a width predicted value of the tracking target frame, a height predicted value of the tracking target frame, a lateral velocity predicted value of the center of the tracking target frame, a longitudinal velocity predicted value of the center of the tracking target frame, a width change velocity predicted value of the tracking target frame, and a height change velocity predicted value of the tracking target frame at the previous time,
Figure 62227DEST_PATH_IMAGE028
the state covariance of the tracked target at the current moment is P, the state covariance of the tracked target at the last moment is P, Q is a noise matrix coefficient of the system, and F is a conversion matrix.
The identity information in step a5 may be name information and/or number information. Through the step a5, matching between each tracking target recognized at the current time and each tracking target recognized at the previous time can be realized, so as to determine the identity of each tracking target recognized at the current time, and allocate correct identity information to each tracking target recognized at the current time, thereby ensuring that the finally integrated tracking and recognizing information of each tracking target is information really belonging to the corresponding tracking target. Preferably, the matching threshold used when matching using the hungarian matching algorithm is 0.6.
In some preferred embodiments, step a5 is further followed by:
A8. updating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 414711DEST_PATH_IMAGE029
;
Figure 294942DEST_PATH_IMAGE030
;
Figure 732877DEST_PATH_IMAGE031
;
Figure 620672DEST_PATH_IMAGE032
;
Figure 827662DEST_PATH_IMAGE033
;
Figure 941112DEST_PATH_IMAGE034
;
wherein z is the actual measurement state information of the tracking target at the current moment,
Figure 804026DEST_PATH_IMAGE035
Figure 28334DEST_PATH_IMAGE036
Figure 542360DEST_PATH_IMAGE037
Figure 561132DEST_PATH_IMAGE038
Figure 973659DEST_PATH_IMAGE039
Figure 939341DEST_PATH_IMAGE040
Figure 120923DEST_PATH_IMAGE041
Figure 763126DEST_PATH_IMAGE042
respectively an abscissa value measured value of the center of the tracking target frame, an ordinate value measured value of the center of the tracking target frame, a width measured value of the tracking target frame, a height measured value of the tracking target frame, a lateral velocity measured value of the center of the tracking target frame, a longitudinal velocity measured value of the center of the tracking target frame, a width change velocity measured value of the tracking target frame, a height change velocity measured value of the tracking target frame, y is a residual error, H is an observation matrix, R is a noise matrix, S is a process variable, K is a Kalman gain,
Figure 397370DEST_PATH_IMAGE043
is updated
Figure 901164DEST_PATH_IMAGE011
,
Figure 202832DEST_PATH_IMAGE044
Is updated
Figure 828985DEST_PATH_IMAGE028
,
Figure 137476DEST_PATH_IMAGE045
Is an identity matrix.
Therefore, when the predicted state information is calculated at the next time, the updated state covariance and the updated predicted state information are used for calculation. By updating the predicted state information and the state covariance in the above manner, the predicted state information obtained by calculation at the next moment is closer to the corresponding actually measured state information, so that the matching accuracy can be improved, and the tracking effect is further improved.
The probability of the midpoint of the tracking target frame falling on the final target body is higher, and the probability of the midpoint falling on the background part is lower, so that the midpoint position of the tracking target frame can represent the position of the corresponding tracking target, and the depth information of the midpoint of the tracking target frame can represent the depth information of the corresponding tracking target; therefore, in step a6, the probability that the acquired distance information and position information of the tracking target are the true distance information and position information of the corresponding tracking target is high. The distance information of the tracking target refers to the distance information of the tracking target from the center point of the binocular camera.
Specifically, step a7 includes: and adding the distance information and the position information of each tracking target acquired at the current moment into the corresponding tracking identification information data set according to the identity information. Therefore, the identity information, the distance information data sequence and the position information data sequence of the corresponding tracking target are recorded in the tracking identification information data set of each tracking target.
In some preferred embodiments, after step a7, the method further comprises the steps of:
A9. marking corresponding identity information on one side of each tracking target frame in the target detection image, and adding a motion trajectory line of each tracking target to the target detection image;
A10. the target detection image is displayed.
The motion trajectory line of the tracked target is a line formed by sequentially connecting all position points of a position information data sequence in the tracking identification information data set of the tracked target by straight lines. Preferably, the connecting line between each position point is a straight line with an arrow, so as to mark the moving direction of the tracking target; such as the target detection image shown in fig. 9. Since the image information is periodically acquired and processed, the displayed target detection image is also periodically updated.
According to the multi-target unmanned aerial vehicle tracking and identifying method, the image information collected by the binocular camera is periodically acquired, and the image information comprises depth information; inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment; matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target; integrating the identity information of each tracking target with corresponding distance information and position information, and updating the tracking identification information of each tracking target; therefore, the unmanned aerial vehicles can be accurately tracked and identified; the multi-target unmanned aerial vehicle tracking and identifying method has the following advantages:
1. the method utilizes the multilayer neural network to extract and learn the characteristics of the unmanned aerial vehicle so as to achieve the recognition effect, so that the method avoids the problem of various interferences of the traditional detection method in multi-target detection in principle; meanwhile, the neural network model carries out parameterized learning on the characteristics of the unmanned aerial vehicle, so that the neural network model has good compatibility with future unmanned aerial vehicles;
2. the method has the advantages that detection is carried out on 5 dimensions in the deep learning model, so that the method has good detection precision, the detection effect on small targets is effectively improved, and the problem of poor detection precision of the small target unmanned aerial vehicle in the prior art is solved;
3. the method can realize the accurate tracking of the multiple unmanned aerial vehicles and accurately obtain the motion information of the multiple unmanned aerial vehicles;
4. the deep learning model of the method adopts a unique data enhancement method during training, so that the quantity and the quality of data are effectively improved, the difficulty that the existing training data is insufficient is solved, and the time and the economic cost for artificially constructing the data are saved.
Referring to fig. 2, the embodiment of the application provides a multi-target unmanned aerial vehicle tracks recognition device, includes:
a first acquisition module 1, configured to periodically acquire image information acquired by a binocular camera, where the image information includes depth information;
the detection module 2 is used for inputting the image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; the target detection image comprises a plurality of tracking target frames, each tracking target frame corresponds to one tracking target, and the corresponding tracking target is surrounded; the tracking target is an unmanned aerial vehicle;
the second obtaining module 3 is configured to obtain actual measurement state information of each tracking target at the current moment according to the target detection image;
the prediction module 4 is used for calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target obtained at the previous moment based on a Kalman filtering prediction method;
the matching module 5 is used for matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment;
a third obtaining module 6, configured to obtain depth information of a midpoint in a tracking target frame of each tracking target as distance information of each tracking target, and obtain position information of the midpoint in the tracking target frame of each tracking target as position information of each tracking target;
and the integration module 7 is used for integrating the identity information of each tracking target with the corresponding distance information and position information and updating the tracking identification information of each tracking target.
Preferably, the deep learning model is a deep learning model based on a YOLOV3 neural network and a Deepsort algorithm; during operation, the deep learning model firstly performs 2 times of upsampling on a characteristic graph which is output by a YOLOV3 framework and is subjected to 8 times of downsampling, the 2 times of upsampling characteristic graph is spliced with a4 times of downsampling characteristic graph output by a YOLOV3 framework, a fusion information layer which is 4 times of downsampling is output, then the fusion information layer which is 4 times of downsampling is performed again, and the output of the fusion information layer is spliced with 2 times of downsampling of a YOLOV3 framework.
Specifically, see fig. 4, which is a network structure diagram of the deep learning model, the network structure mainly includes a convolutional layer, a sampling layer, and a residual structure, and finally 5 detection layers are formed to detect an object from different dimensions.
In some embodiments, the deep learning model is trained by the following training process:
s1, data acquisition: collecting a plurality of unmanned aerial vehicle pictures, and processing the unmanned aerial vehicle pictures into the same size;
s2, data annotation: marking the unmanned aerial vehicle picture;
s3, data enhancement: the method comprises the steps that a randomly selected preset blocking block is adopted to block collected unmanned aerial vehicle pictures to generate new unmanned aerial vehicle pictures, and/or each pixel point of two unmanned aerial vehicle pictures is linearly superposed to obtain new unmanned aerial vehicle pictures, and the collected unmanned aerial vehicle pictures and the new unmanned aerial vehicle pictures are taken as training sets;
s4, training a model: and training the deep learning model by utilizing the training set.
In step S1, the drone image may be processed to a size of 608X608 (pixels), and the processing may specifically be performed by stretching, zooming, and the like. The number of the collected pictures can be set according to actual measurement requirements, such as 1800 and 2000 pictures.
In step S2, labeling may be performed by a labelImg-master, but is not limited thereto.
Preferably, in step S3, the step of linearly superimposing the pixel points of the two unmanned aerial vehicle pictures to obtain a new unmanned aerial vehicle picture includes:
by means of arrangement and combination, two different unmanned aerial vehicle pictures are used as a group to obtain a plurality of unmanned aerial vehicle picture combinations;
obtaining a new unmanned aerial vehicle picture aiming at each unmanned aerial vehicle picture combination, wherein the pixel value of each pixel point of the new unmanned aerial vehicle picture is obtained by calculation according to the following formula:
Figure 507277DEST_PATH_IMAGE001
;
wherein the content of the first and second substances,
Figure 601135DEST_PATH_IMAGE002
for the pixel value of the ith pixel point in the new picture of the unmanned aerial vehicle,
Figure 398190DEST_PATH_IMAGE003
the pixel value of the ith pixel point of the first unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination,
Figure 7026DEST_PATH_IMAGE004
the pixel value of the ith pixel point of the second unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination is obtained;
Figure 104819DEST_PATH_IMAGE005
is a scaling factor.
For example, in fig. 8, a new picture of the unmanned aerial vehicle is obtained by linearly superimposing two pictures, where each of the two original pictures of the unmanned aerial vehicle has one unmanned aerial vehicle.
Further, in step S3, after obtaining the new picture of the unmanned aerial vehicle, it may be determined whether the new picture of the unmanned aerial vehicle is qualified, and if not, the new picture of the unmanned aerial vehicle is deleted. Wherein, judge whether qualified judgement condition of new unmanned aerial vehicle picture is: and if the overlapping part of the unmanned aerial vehicle is less than 30%, judging that the picture is qualified, otherwise, judging that the picture is unqualified.
Preferably, in step S3, the step of blocking the acquired picture of the unmanned aerial vehicle by using the randomly selected preset blocking block to generate a new picture of the unmanned aerial vehicle includes:
randomly selecting a numerical value in a preset numerical value range (the range is [1, N ], and N is a positive integer larger than 1) as the number of the shielding objects;
and randomly selecting a corresponding number of shelters from a preset shelter database according to the number of the shelters, and adding the shelters into the unmanned aerial vehicle picture for sheltering.
The preset shelter database is stored with a plurality of shelter block images with different shapes and/or different sizes in advance. When adding the blocking object into the unmanned aerial vehicle picture, the center of the blocking object is arranged in the tracking target frame area (a pixel point position can be randomly selected in the tracking target frame area, and the center of the blocking object is arranged at the pixel point position). For example, in fig. 7, for a new picture of the unmanned aerial vehicle obtained after the blocking, a rectangular blocking block and an oval blocking block are adopted for blocking.
By the data enhancement processing in step S3, the sample size of the training set can be increased significantly, thereby avoiding the situation of insufficient sample size and insufficient data complexity of the training set. Compared with the traditional cutting and splicing method, the mode of linearly superposing the pixel points of the two unmanned aerial vehicle pictures to obtain the new unmanned aerial vehicle picture avoids the problems of unnatural splicing effect and excessively harsh edge details.
In some embodiments, in step S4, the learning rate is set to 0.0013 and the number of training times is 10000, but the method is not limited thereto.
Preferably, the deep learning model further comprises the steps of, through the following training process:
s5, adjusting parameters: obtaining the prior frame size of the training set by using K-means + + clustering, and modifying the prior frame size;
s6, repeatedly executing the steps S4 and S5 (the repeated times are preset according to actual measurement requirements).
In some preferred embodiments, when inputting the image information acquired at the current time into the pre-trained deep learning model to obtain the target detection image, the detection module 2:
if the current time is the starting time of the current calculation cycle, then:
amplifying an original image acquired at the current moment according to a preset proportion to obtain amplified image information;
respectively inputting the original image information and the amplified image information acquired at the current moment into a pre-trained deep learning model to obtain two target detection images;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is more than that in the target detection image corresponding to the original image information, determining the amplified image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is not more than that of the tracking targets in the target detection image corresponding to the original image information, determining the original image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
and if the current time is not the starting time of the current calculation period, inputting the corresponding effective input image information into a pre-trained deep learning model to obtain an effective target detection image.
The calculation period is a preset time period, which is longer than a sampling period for acquiring the image information acquired by the binocular camera, and for example, the calculation period is 10s, but is not limited thereto. At the initial moment of each calculation period, inputting original image information and amplified image information into a depth learning model respectively to obtain two target detection images, wherein if the number of unmanned aerial vehicles identified in an identification result corresponding to the original image information is more, the identification by using the original image information is more accurate, so that the original image information is adopted for identification in the period; if the number of the unmanned aerial vehicles identified in the identification result corresponding to the amplified image information is larger, the identification by the amplified image information is more accurate, and therefore the amplified image information is adopted for identification in the period. Therefore, the identification efficiency and accuracy are greatly improved, and meanwhile, the real-time calculation time is saved, and the stability of the frame number is ensured as far as possible.
The preset scale of the amplification can be set according to actual measurement requirements, for example, the size of the original picture is 608X608 (pixels), and the size after the amplification is 1024X 1024 (pixels).
In this embodiment, the state information of the tracking target includes an abscissa value of the center of the tracking target frame, an ordinate value of the center of the tracking target frame, a width of the tracking target frame (length in the abscissa axis direction), a height of the tracking target frame (length in the ordinate axis direction), a lateral velocity of the center of the tracking target frame, a longitudinal velocity of the center of the tracking target frame, a width change velocity of the tracking target frame, and a height change velocity of the tracking target frame. The horizontal speed can be obtained by calculating the horizontal coordinate value at the current moment, the horizontal coordinate value at the last moment and the sampling period, the longitudinal speed can be obtained by calculating the vertical coordinate value at the current moment, the vertical coordinate value at the last moment and the sampling period, the width change speed can be obtained by calculating the width at the current moment, the width at the last moment and the sampling period, and the height change speed can be obtained by calculating the height at the current moment, the height at the last moment and the sampling period.
Accordingly, the actual measurement state information acquired by the second acquisition module 3 includes an abscissa value actual measurement value of the center of the tracking target frame, an ordinate value actual measurement value of the center of the tracking target frame, a tracking target frame width actual measurement value, a tracking target frame height actual measurement value, a tracking target frame center lateral velocity actual measurement value, a tracking target frame center longitudinal velocity actual measurement value, a tracking target frame width change velocity actual measurement value, and a tracking target frame height change velocity actual measurement value of each tracking target.
Preferably, when calculating the prediction state information of each tracking target at the current time, the prediction module 4:
calculating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 115501DEST_PATH_IMAGE006
;
Figure 21140DEST_PATH_IMAGE007
;
Figure 117272DEST_PATH_IMAGE008
;
Figure 828876DEST_PATH_IMAGE009
;
Figure 881014DEST_PATH_IMAGE010
;
wherein the content of the first and second substances,
Figure 19872DEST_PATH_IMAGE011
to track the predicted state information of the target at the current time,
Figure 540983DEST_PATH_IMAGE012
Figure 56278DEST_PATH_IMAGE013
Figure 775972DEST_PATH_IMAGE014
Figure 272681DEST_PATH_IMAGE015
Figure 343406DEST_PATH_IMAGE016
Figure 334495DEST_PATH_IMAGE017
Figure 174275DEST_PATH_IMAGE018
Figure 841886DEST_PATH_IMAGE019
respectively an abscissa value predicted value of the center of a tracking target frame, an ordinate value predicted value of the center of the tracking target frame, a width predicted value of the tracking target frame, a height predicted value of the tracking target frame, a lateral velocity predicted value of the center of the tracking target frame, a longitudinal velocity predicted value of the center of the tracking target frame, a width change velocity predicted value of the tracking target frame, and a height change velocity predicted value of the tracking target frame at the current time, wherein x is the predicted state information of the tracking target at the previous time,
Figure 399906DEST_PATH_IMAGE020
Figure 991425DEST_PATH_IMAGE021
Figure 623394DEST_PATH_IMAGE022
Figure 540535DEST_PATH_IMAGE023
Figure 510152DEST_PATH_IMAGE024
Figure 905361DEST_PATH_IMAGE025
Figure 719734DEST_PATH_IMAGE026
Figure 479879DEST_PATH_IMAGE027
respectively, an abscissa value predicted value of the center of the tracking target frame, an ordinate value predicted value of the center of the tracking target frame, a width predicted value of the tracking target frame, a height predicted value of the tracking target frame, a lateral velocity predicted value of the center of the tracking target frame, a longitudinal velocity predicted value of the center of the tracking target frame, a width change velocity predicted value of the tracking target frame, and a height change velocity predicted value of the tracking target frame at the previous time,
Figure 746913DEST_PATH_IMAGE028
the state covariance of the tracked target at the current moment is P, the state covariance of the tracked target at the last moment is P, Q is a noise matrix coefficient of the system, and F is a conversion matrix.
Wherein, the identity information can be name information and/or number information. Preferably, the matching threshold used when matching using the hungarian matching algorithm is 0.6.
In some preferred embodiments, the apparatus for identifying and tracking multiple target drones further includes:
the updating module is used for updating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 132763DEST_PATH_IMAGE029
;
Figure 801642DEST_PATH_IMAGE030
;
Figure 795006DEST_PATH_IMAGE031
;
Figure 487018DEST_PATH_IMAGE032
;
Figure 410981DEST_PATH_IMAGE033
;
Figure 934366DEST_PATH_IMAGE034
;
wherein z is the actual measurement state information of the tracking target at the current moment,
Figure 301894DEST_PATH_IMAGE035
Figure 543519DEST_PATH_IMAGE036
Figure 818643DEST_PATH_IMAGE037
Figure 649064DEST_PATH_IMAGE038
Figure 984231DEST_PATH_IMAGE039
Figure 650835DEST_PATH_IMAGE040
Figure 729650DEST_PATH_IMAGE041
Figure 227627DEST_PATH_IMAGE042
respectively an abscissa value measured value of the center of the tracking target frame, an ordinate value measured value of the center of the tracking target frame, a width measured value of the tracking target frame, a height measured value of the tracking target frame, a lateral velocity measured value of the center of the tracking target frame, a longitudinal velocity measured value of the center of the tracking target frame, a width change velocity measured value of the tracking target frame, a height change velocity measured value of the tracking target frame, y is a residual error, H is an observation matrix, R is a noise matrix, S is a process variable, K is a Kalman gain,
Figure 189155DEST_PATH_IMAGE043
is updated
Figure 139793DEST_PATH_IMAGE011
,
Figure 959981DEST_PATH_IMAGE044
Is updated
Figure 312465DEST_PATH_IMAGE028
,
Figure 255014DEST_PATH_IMAGE045
Is an identity matrix.
Specifically, when the integration module 7 integrates the identity information of each tracking target with the corresponding distance information and position information to update the tracking identification information of each tracking target: and adding the distance information and the position information of each tracking target acquired at the current moment into the corresponding tracking identification information data set according to the identity information. Therefore, the identity information, the distance information data sequence and the position information data sequence of the corresponding tracking target are recorded in the tracking identification information data set of each tracking target.
In some preferred embodiments, after step a7, the apparatus for tracking and identifying multi-target drones further includes:
the image processing module is used for marking corresponding identity information on one side of each tracking target frame in the target detection image and adding a motion trajectory line of each tracking target to the target detection image;
and the display module is used for displaying the target detection image.
The motion trajectory line of the tracked target is a line formed by sequentially connecting all position points of a position information data sequence in the tracking identification information data set of the tracked target by straight lines. Preferably, the connecting line between each position point is a straight line with an arrow, so as to mark the moving direction of the tracking target; such as the target detection image shown in fig. 9. Since the image information is periodically acquired and processed, the displayed target detection image is also periodically updated.
According to the multi-target unmanned aerial vehicle tracking and identifying device, the image information collected by the binocular camera is periodically acquired, and the image information comprises depth information; inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment; matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target; integrating the identity information of each tracking target with corresponding distance information and position information, and updating the tracking identification information of each tracking target; therefore, the unmanned aerial vehicles can be accurately tracked and identified; this multi-target unmanned aerial vehicle tracks up recognition device has following advantage:
1. the device utilizes the multilayer neural network to extract and learn the characteristics of the unmanned aerial vehicle so as to achieve the recognition effect, so that the method avoids the problem of various interferences of the traditional detection method in multi-target detection in principle; meanwhile, the neural network model carries out parameterized learning on the characteristics of the unmanned aerial vehicle, so that the neural network model has good compatibility with future unmanned aerial vehicles;
2. the device has the advantages that detection is performed on 5 dimensions in the deep learning model, so that the device has good detection precision, the detection effect on small targets is effectively improved, and the problem of poor detection precision of the small target unmanned aerial vehicle in the prior art is solved;
3. the method can realize the accurate tracking of the multiple unmanned aerial vehicles and accurately obtain the motion information of the multiple unmanned aerial vehicles;
4. the deep learning model of the device adopts a unique data enhancement method during training, so that the quantity and the quality of data are effectively improved, the difficulty that the existing training data is insufficient is solved, and the time and the economic cost for artificially constructing the data are saved. .
Referring to fig. 3, an electronic device 100 according to an embodiment of the present application further includes a processor 101 and a memory 102, where the memory 102 stores a computer program, and the processor 101 is configured to execute the steps of the multi-target drone tracking and identifying method as described above by calling the computer program stored in the memory 102.
The processor 101 is electrically connected to the memory 102. The processor 101 is a control center of the electronic device 100, connects various parts of the entire electronic device using various interfaces and lines, and performs various functions of the electronic device and processes data by running or calling a computer program stored in the memory 102 and calling data stored in the memory 102, thereby performing overall monitoring of the electronic device.
The memory 102 may be used to store computer programs and data. The memory 102 stores computer programs containing instructions executable in the processor. The computer program may constitute various functional modules. The processor 101 executes various functional applications and data processing by calling a computer program stored in the memory 102.
In this embodiment, the processor 101 in the electronic device 100 loads instructions corresponding to one or more processes of the computer program into the memory 102, and the processor 101 runs the computer program stored in the memory 102 according to the following steps, so as to implement various functions: periodically acquiring image information collected by a binocular camera, the image information including depth information; inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment; matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target; and integrating the identity information of each tracking target with the corresponding distance information and position information, and updating the tracking identification information of each tracking target.
As can be seen from the above, the electronic device periodically acquires image information collected by the binocular camera, the image information including depth information; inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment; matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target; integrating the identity information of each tracking target with corresponding distance information and position information, and updating the tracking identification information of each tracking target; therefore, the tracking identification of the multiple unmanned aerial vehicles can be accurately carried out.
The embodiment of the application further provides a storage medium, on which a computer program is stored, and when the computer program is executed by a processor, the method for tracking and identifying a multi-target unmanned aerial vehicle runs the steps of the method for tracking and identifying a multi-target unmanned aerial vehicle, so as to implement the following functions: periodically acquiring image information collected by a binocular camera, the image information including depth information; inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; acquiring actual measurement state information of each tracking target at the current moment according to the target detection image; based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment; matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment; acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target; and integrating the identity information of each tracking target with the corresponding distance information and position information, and updating the tracking identification information of each tracking target.
The storage medium may be implemented by any type of volatile or nonvolatile storage device or combination thereof, such as a Static Random Access Memory (SRAM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), an Erasable Programmable Read-Only Memory (EPROM), a Programmable Read-Only Memory (PROM), a Read-Only Memory (ROM), a magnetic Memory, a flash Memory, a magnetic disk, or an optical disk.
In summary, although the present invention has been described with reference to the preferred embodiments, the above-described preferred embodiments are not intended to limit the present invention, and those skilled in the art can make various changes and modifications without departing from the spirit and scope of the present invention, which are substantially the same as the present invention.

Claims (10)

1. The multi-target unmanned aerial vehicle tracking and identifying method is characterized by comprising the following steps:
A1. periodically acquiring image information collected by a binocular camera, the image information including depth information;
A2. inputting image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; the target detection image comprises a plurality of tracking target frames, each tracking target frame corresponds to one tracking target, and the corresponding tracking target is surrounded; the tracking target is an unmanned aerial vehicle;
A3. acquiring actual measurement state information of each tracking target at the current moment according to the target detection image;
A4. based on a Kalman filtering prediction method, calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target acquired at the previous moment;
A5. matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment;
A6. acquiring depth information of the midpoint of a tracking target frame of each tracking target as distance information of each tracking target, and acquiring position information of the midpoint of the tracking target frame of each tracking target as position information of each tracking target;
A7. and integrating the identity information of each tracking target with the corresponding distance information and position information, and updating the tracking identification information of each tracking target.
2. The multi-target unmanned aerial vehicle tracking identification method according to claim 1, wherein the deep learning model is a deep learning model based on a YOLOV3 neural network and a Deepsort algorithm; during operation, the deep learning model firstly performs 2 times of upsampling on a characteristic graph which is output by a YOLOV3 framework and is subjected to 8 times of downsampling, the 2 times of upsampling characteristic graph is spliced with a4 times of downsampling characteristic graph output by a YOLOV3 framework, a fusion information layer which is 4 times of downsampling is output, then the fusion information layer which is 4 times of downsampling is performed again, and the output of the fusion information layer is spliced with 2 times of downsampling of a YOLOV3 framework.
3. The multi-target unmanned aerial vehicle tracking and identifying method according to claim 2, wherein the deep learning model is obtained by training through the following training process:
s1, data acquisition: collecting a plurality of unmanned aerial vehicle pictures, and processing the unmanned aerial vehicle pictures into the same size;
s2, data annotation: marking the unmanned aerial vehicle picture;
s3, data enhancement: the method comprises the steps that a randomly selected preset blocking block is adopted to block collected unmanned aerial vehicle pictures to generate new unmanned aerial vehicle pictures, and/or each pixel point of two unmanned aerial vehicle pictures is linearly superposed to obtain new unmanned aerial vehicle pictures, and the collected unmanned aerial vehicle pictures and the new unmanned aerial vehicle pictures are taken as training sets;
s4, training a model: and training the deep learning model by utilizing the training set.
4. The multi-target unmanned aerial vehicle tracking and identifying method according to claim 3, wherein in the step S3, the step of linearly superimposing each pixel point of the two unmanned aerial vehicle pictures to obtain a new unmanned aerial vehicle picture comprises:
by means of arrangement and combination, two different unmanned aerial vehicle pictures are used as a group to obtain a plurality of unmanned aerial vehicle picture combinations;
obtaining a new unmanned aerial vehicle picture aiming at each unmanned aerial vehicle picture combination, wherein the pixel value of each pixel point of the new unmanned aerial vehicle picture is obtained by calculation according to the following formula:
Figure 328121DEST_PATH_IMAGE001
;
wherein the content of the first and second substances,
Figure 510841DEST_PATH_IMAGE002
for the pixel value of the ith pixel point in the new picture of the unmanned aerial vehicle,
Figure 290578DEST_PATH_IMAGE003
the pixel value of the ith pixel point of the first unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination,
Figure 544842DEST_PATH_IMAGE004
the pixel value of the ith pixel point of the second unmanned aerial vehicle picture in the unmanned aerial vehicle picture combination is obtained;
Figure 359214DEST_PATH_IMAGE005
is a scaling factor.
5. The multi-target unmanned aerial vehicle tracking and identifying method according to claim 1, wherein the step A2 comprises:
A201. if the current time is the starting time of the current calculation cycle, then:
amplifying an original image acquired at the current moment according to a preset proportion to obtain amplified image information;
respectively inputting the original image information and the amplified image information acquired at the current moment into a pre-trained deep learning model to obtain two target detection images;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is more than that in the target detection image corresponding to the original image information, determining the amplified image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
if the number of the tracking targets in the target detection image corresponding to the amplified image information is not more than that of the tracking targets in the target detection image corresponding to the original image information, determining the original image information in the current calculation period as effective input image information, and taking the target detection image corresponding to the effective input image information as an effective target detection image;
A202. and if the current time is not the starting time of the current calculation period, inputting the corresponding effective input image information into a pre-trained deep learning model to obtain an effective target detection image.
6. The multi-target unmanned aerial vehicle tracking and identifying method according to claim 1, wherein the step A4 comprises:
calculating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 181677DEST_PATH_IMAGE006
;
Figure 12492DEST_PATH_IMAGE007
;
Figure 211392DEST_PATH_IMAGE008
;
Figure 942588DEST_PATH_IMAGE009
;
Figure 935951DEST_PATH_IMAGE010
;
wherein the content of the first and second substances,
Figure 690281DEST_PATH_IMAGE011
to track the predicted state information of the target at the current time,
Figure 224030DEST_PATH_IMAGE012
Figure 747415DEST_PATH_IMAGE013
Figure 177260DEST_PATH_IMAGE014
Figure 481202DEST_PATH_IMAGE015
Figure 756326DEST_PATH_IMAGE016
Figure 399797DEST_PATH_IMAGE017
Figure 86693DEST_PATH_IMAGE018
Figure 815615DEST_PATH_IMAGE019
respectively an abscissa value predicted value of the center of a tracking target frame, an ordinate value predicted value of the center of the tracking target frame, a width predicted value of the tracking target frame, a height predicted value of the tracking target frame, a lateral velocity predicted value of the center of the tracking target frame, a longitudinal velocity predicted value of the center of the tracking target frame, a width change velocity predicted value of the tracking target frame, and a height change velocity predicted value of the tracking target frame at the current time, wherein x is the predicted state information of the tracking target at the previous time,
Figure 894429DEST_PATH_IMAGE020
Figure 454724DEST_PATH_IMAGE021
Figure 226371DEST_PATH_IMAGE022
Figure 239326DEST_PATH_IMAGE023
Figure 121831DEST_PATH_IMAGE024
Figure 474315DEST_PATH_IMAGE025
Figure 479180DEST_PATH_IMAGE026
Figure 917115DEST_PATH_IMAGE027
the predicted value of the horizontal coordinate value of the center of the tracking target frame, the predicted value of the vertical coordinate value of the center of the tracking target frame, the predicted value of the width of the tracking target frame, the predicted value of the height of the tracking target frame, the predicted value of the horizontal velocity of the center of the tracking target frame, the predicted value of the vertical velocity of the center of the tracking target frame, the predicted value of the tracking target frameThe predicted value of the width change speed of the target frame, the predicted value of the height change speed of the tracking target frame,
Figure 337732DEST_PATH_IMAGE028
the state covariance of the tracked target at the current moment is P, the state covariance of the tracked target at the last moment is P, Q is a noise matrix coefficient of the system, and F is a conversion matrix.
7. The multi-target drone tracking and identifying method according to claim 6, further comprising after step a 5:
A8. updating the prediction state information of the tracking target at the current moment according to the following formula:
Figure 108504DEST_PATH_IMAGE029
;
Figure 221954DEST_PATH_IMAGE030
;
Figure 147184DEST_PATH_IMAGE031
;
Figure 433809DEST_PATH_IMAGE032
;
Figure 495306DEST_PATH_IMAGE033
;
Figure 841974DEST_PATH_IMAGE034
;
wherein z is the actual measurement state information of the tracking target at the current moment,
Figure 254501DEST_PATH_IMAGE035
Figure 282500DEST_PATH_IMAGE036
Figure 821672DEST_PATH_IMAGE037
Figure 11345DEST_PATH_IMAGE038
Figure 911168DEST_PATH_IMAGE039
Figure 539595DEST_PATH_IMAGE040
Figure 841263DEST_PATH_IMAGE041
Figure 201838DEST_PATH_IMAGE042
respectively an abscissa value measured value of the center of the tracking target frame, an ordinate value measured value of the center of the tracking target frame, a width measured value of the tracking target frame, a height measured value of the tracking target frame, a lateral velocity measured value of the center of the tracking target frame, a longitudinal velocity measured value of the center of the tracking target frame, a width change velocity measured value of the tracking target frame, a height change velocity measured value of the tracking target frame, y is a residual error, H is an observation matrix, R is a noise matrix, S is a process variable, K is a Kalman gain,
Figure 385694DEST_PATH_IMAGE043
is updated
Figure 755496DEST_PATH_IMAGE011
,
Figure 646091DEST_PATH_IMAGE044
Is updated
Figure 505463DEST_PATH_IMAGE028
,
Figure 114299DEST_PATH_IMAGE045
Is an identity matrix.
8. The utility model provides a multi-target unmanned aerial vehicle tracks recognition device which characterized in that includes:
the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for periodically acquiring image information acquired by a binocular camera, and the image information comprises depth information;
the detection module is used for inputting the image information acquired at the current moment into a pre-trained deep learning model to obtain a target detection image; the target detection image comprises a plurality of tracking target frames, each tracking target frame corresponds to one tracking target, and the corresponding tracking target is surrounded; the tracking target is an unmanned aerial vehicle;
the second acquisition module is used for acquiring the actual measurement state information of each tracking target at the current moment according to the target detection image;
the prediction module is used for calculating the prediction state information of each tracking target at the current moment according to the actual measurement state information of each tracking target obtained at the previous moment based on a Kalman filtering prediction method;
the matching module is used for matching each tracking target at the current moment with each tracking target at the previous moment by adopting a Hungarian matching algorithm according to the actual measurement state information and the prediction state information of each tracking target at the current moment so as to determine the identity information of each tracking target at the current moment;
the third acquisition module is used for acquiring the depth information of the midpoint of the tracking target frame of each tracking target as the distance information of each tracking target and acquiring the position information of the midpoint of the tracking target frame of each tracking target as the position information of each tracking target;
and the integration module is used for integrating the identity information of each tracking target and the corresponding distance information and position information and updating the tracking identification information of each tracking target.
9. An electronic device, characterized by comprising a processor and a memory, wherein the memory stores a computer program, and the processor is used for executing the steps of the multi-target drone tracking and identifying method according to any one of claims 1-7 by calling the computer program stored in the memory.
10. A storage medium having a computer program stored thereon, wherein the computer program, when executed by a processor, performs the steps of the multi-target drone tracking identification method according to any one of claims 1-7.
CN202110845085.3A 2021-07-26 2021-07-26 Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium Active CN113298053B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110845085.3A CN113298053B (en) 2021-07-26 2021-07-26 Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110845085.3A CN113298053B (en) 2021-07-26 2021-07-26 Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN113298053A true CN113298053A (en) 2021-08-24
CN113298053B CN113298053B (en) 2021-11-30

Family

ID=77330880

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110845085.3A Active CN113298053B (en) 2021-07-26 2021-07-26 Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113298053B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114255407A (en) * 2021-12-13 2022-03-29 中国电子科技集团公司第三十八研究所 High-resolution-based anti-unmanned aerial vehicle multi-target identification and tracking video detection method
CN114281096A (en) * 2021-11-09 2022-04-05 中时讯通信建设有限公司 Unmanned aerial vehicle tracking control method, device and medium based on target detection algorithm
CN115035425A (en) * 2022-06-07 2022-09-09 北京庚图科技有限公司 Target identification method and system based on deep learning, electronic equipment and storage medium
CN116528062A (en) * 2023-07-05 2023-08-01 合肥中科类脑智能技术有限公司 Multi-target tracking method
CN117079196A (en) * 2023-10-16 2023-11-17 长沙北斗产业安全技术研究院股份有限公司 Unmanned aerial vehicle identification method based on deep learning and target motion trail

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109961061A (en) * 2019-04-15 2019-07-02 艾物智联(北京)科技有限公司 A kind of edge calculations video data structure method and system
CN111462184A (en) * 2020-04-02 2020-07-28 桂林电子科技大学 Online sparse prototype tracking method based on twin neural network linear representation model
CN111476817A (en) * 2020-02-27 2020-07-31 浙江工业大学 Multi-target pedestrian detection tracking method based on yolov3
CN111583307A (en) * 2020-05-11 2020-08-25 山东万腾电子科技有限公司 Real-time detection and tracking system and method for moving target
CN111832608A (en) * 2020-05-29 2020-10-27 上海海事大学 Multi-abrasive-particle identification method for ferrographic image based on single-stage detection model yolov3

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109961061A (en) * 2019-04-15 2019-07-02 艾物智联(北京)科技有限公司 A kind of edge calculations video data structure method and system
CN111476817A (en) * 2020-02-27 2020-07-31 浙江工业大学 Multi-target pedestrian detection tracking method based on yolov3
CN111462184A (en) * 2020-04-02 2020-07-28 桂林电子科技大学 Online sparse prototype tracking method based on twin neural network linear representation model
CN111583307A (en) * 2020-05-11 2020-08-25 山东万腾电子科技有限公司 Real-time detection and tracking system and method for moving target
CN111832608A (en) * 2020-05-29 2020-10-27 上海海事大学 Multi-abrasive-particle identification method for ferrographic image based on single-stage detection model yolov3

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
LIANG WANG ET AL.: "Design of Airport Obstacle-Free Zone Monitoring UAV System Based on Computer Vision", 《SENSORS》 *
刘鑫明: "基于双目视觉的无人船水上目标测距和识别", 《中国优秀博硕士学位论文全文数据库(硕士)工程科技II辑》 *
杜文凯: "基于图像处理的室内尾随检测算法设计", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *
杨光: "无人机集群目标几何特征识别", 《中国优秀博硕士学位论文全文数据库(硕士)工程科技II辑》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114281096A (en) * 2021-11-09 2022-04-05 中时讯通信建设有限公司 Unmanned aerial vehicle tracking control method, device and medium based on target detection algorithm
CN114255407A (en) * 2021-12-13 2022-03-29 中国电子科技集团公司第三十八研究所 High-resolution-based anti-unmanned aerial vehicle multi-target identification and tracking video detection method
CN115035425A (en) * 2022-06-07 2022-09-09 北京庚图科技有限公司 Target identification method and system based on deep learning, electronic equipment and storage medium
CN115035425B (en) * 2022-06-07 2024-02-09 北京庚图科技有限公司 Target recognition method, system, electronic equipment and storage medium based on deep learning
CN116528062A (en) * 2023-07-05 2023-08-01 合肥中科类脑智能技术有限公司 Multi-target tracking method
CN116528062B (en) * 2023-07-05 2023-09-15 合肥中科类脑智能技术有限公司 Multi-target tracking method
CN117079196A (en) * 2023-10-16 2023-11-17 长沙北斗产业安全技术研究院股份有限公司 Unmanned aerial vehicle identification method based on deep learning and target motion trail
CN117079196B (en) * 2023-10-16 2023-12-29 长沙北斗产业安全技术研究院股份有限公司 Unmanned aerial vehicle identification method based on deep learning and target motion trail

Also Published As

Publication number Publication date
CN113298053B (en) 2021-11-30

Similar Documents

Publication Publication Date Title
CN113298053B (en) Multi-target unmanned aerial vehicle tracking identification method and device, electronic equipment and storage medium
US10706285B2 (en) Automatic ship tracking method and system based on deep learning network and mean shift
CN110136199B (en) Camera-based vehicle positioning and mapping method and device
US20210326624A1 (en) Method, system and device for difference automatic calibration in cross modal target detection
Levinson et al. Traffic light mapping, localization, and state detection for autonomous vehicles
CN112417926B (en) Parking space identification method and device, computer equipment and readable storage medium
CN111582234B (en) Large-scale oil tea tree forest fruit intelligent detection and counting method based on UAV and deep learning
CN111860352B (en) Multi-lens vehicle track full tracking system and method
CN110765906A (en) Pedestrian detection algorithm based on key points
CN105955308A (en) Aircraft control method and device
CN113112526B (en) Target tracking method, device, equipment and medium
DE102019102866A1 (en) Identification of a manufacturing part using computer vision and machine learning
CN111931582A (en) Image processing-based highway traffic incident detection method
CN105374049B (en) Multi-corner point tracking method and device based on sparse optical flow method
Wu et al. Multivehicle object tracking in satellite video enhanced by slow features and motion features
CN111428539A (en) Target tracking method and device
CN111881322B (en) Target searching method and device, electronic equipment and storage medium
CN114202643A (en) Apple leaf disease identification terminal and method based on multi-sensor fusion
CN111881984A (en) Target detection method and device based on deep learning
CN116823884A (en) Multi-target tracking method, system, computer equipment and storage medium
CN112215714A (en) Rice ear detection method and device based on unmanned aerial vehicle
CN112639400A (en) Distance measurement method, device, equipment and system based on semantic segmentation
CN115115935A (en) Automatic weed identification method, system, equipment and storage medium
CN112818837B (en) Aerial photography vehicle weight recognition method based on attitude correction and difficult sample perception
Zhang et al. Vehicle detection and tracking in remote sensing satellite vidio based on dynamic association

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant