CN114154018A - Cloud-edge collaborative video stream processing method and system for unmanned system - Google Patents

Cloud-edge collaborative video stream processing method and system for unmanned system Download PDF

Info

Publication number
CN114154018A
CN114154018A CN202210117381.6A CN202210117381A CN114154018A CN 114154018 A CN114154018 A CN 114154018A CN 202210117381 A CN202210117381 A CN 202210117381A CN 114154018 A CN114154018 A CN 114154018A
Authority
CN
China
Prior art keywords
cloud
edge
picture
identification
local
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210117381.6A
Other languages
Chinese (zh)
Other versions
CN114154018B (en
Inventor
郭锐
陈忠
施晓东
徐俊瑜
马驰
李磊
姚传明
王超
刘佳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CETC 28 Research Institute
Original Assignee
CETC 28 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CETC 28 Research Institute filed Critical CETC 28 Research Institute
Priority to CN202210117381.6A priority Critical patent/CN114154018B/en
Publication of CN114154018A publication Critical patent/CN114154018A/en
Application granted granted Critical
Publication of CN114154018B publication Critical patent/CN114154018B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7844Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using original textual content or text extracted from visual content or transcript of audio data

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Library & Information Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a cloud-edge collaborative video stream processing method and system for an unmanned system. The method comprises the following steps: the terminal unattended platform cuts the collected video stream, and sends each cut frame of picture to a designated edge terminal device for processing according to a preset distribution strategy; the edge terminal equipment responds to the picture identification request, adds the picture into a cloud sending queue to wait for sending to the cloud, and adds the picture into a local identification queue to wait for identification; the edge terminal equipment takes out the picture to be identified from the local identification queue for identification processing; according to the confidence coefficient obtained by local picture recognition, the recognition result with the confidence coefficient larger than or equal to the specified threshold value is taken as a correct recognition result to be stored; and the edge terminal equipment gives a final picture identification result according to the local identification result and the cloud identification result. According to the invention, the video information processing efficiency of the unattended system in a complex environment can be improved through cloud-edge cooperative processing.

Description

Cloud-edge collaborative video stream processing method and system for unmanned system
Technical Field
The invention relates to unmanned system video stream processing, in particular to an unmanned system-oriented cloud-side collaborative video stream processing method and system.
Background
The unattended system collects video streams through video monitoring and recording equipment to perform real-time analysis and target detection. Most of existing unmanned systems send video streams collected from a terminal unmanned platform to a cloud for processing. However, if all the image data are executed in the cloud, huge overhead is caused to the transmission network; if all detection tasks are executed at the edge end, the edge end equipment is required to have better computing and storing hardware configuration, and the method depends on continuous and stable guarantee, and is difficult to be suitable for the complex and changeable environment of the edge end. Therefore, a single cloud processing or a single edge processing is unrealistic, which needs to combine the cloud computing and the edge computing technologies to construct a cloud-edge collaborative video stream processing system, thereby improving the video information processing efficiency of the unattended system in a complex environment.
Disclosure of Invention
The purpose of the invention is as follows: in order to solve the technical problems, the invention provides a cloud-edge collaborative video stream processing method and system for an unmanned system, which optimize the video stream processing capability of the unmanned system in a cloud-edge collaborative mode by utilizing the low time-delay property of an edge end and the resource sufficiency of a cloud end.
The technical scheme is as follows: a cloud edge collaborative video stream processing method facing an unmanned system comprises the following steps:
the terminal unattended platform cuts the collected video stream, and sends each cut frame of picture to the appointed edge end equipment for processing according to a preset distribution strategy;
the edge terminal equipment responds to the picture identification request, adds the picture into a cloud sending queue to wait for sending to the cloud, and adds the picture into a local identification queue to wait for identification;
the edge terminal equipment identifies the picture data in the local identification queue;
the edge terminal equipment identifies the obtained confidence coefficient f according to the local picture, and the confidence coefficient f is larger than or equal to the specified threshold value
Figure 100002_DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
the edge end equipment judges whether a cloud end identification result is received or not, if the edge end equipment receives the cloud end identification result, comprehensive judgment and scoring are carried out according to the time and the accuracy of the cloud end and the local identification of the edge end, when the score is larger than or equal to a threshold value s, the cloud end identification result is judged to be superior to the local identification result, and the edge end equipment returns the cloud end identification result to the human-computer interaction interface; otherwise, the edge terminal equipment returns a local identification result to the man-machine interaction interface;
and if the edge terminal equipment does not receive the identification result of the cloud, directly returning the local identification result to the human-computer interaction interface.
Wherein the preset allocation policy includes: selecting the edge terminal equipment with the minimum expected waiting time to process the picture according to the identification queue length and the identification time of the adjacent edge terminal equipment stored in the local database:
Figure 100002_DEST_PATH_IMAGE004
where d represents the selected edge device, N represents the total number of edge devices,
Figure 100002_DEST_PATH_IMAGE006
indicating the number of pictures waiting to be identified in the ith edge device,
Figure 100002_DEST_PATH_IMAGE008
is shown as
Figure 100002_DEST_PATH_IMAGE010
The individual edge devices identify the average time of a picture.
The specified threshold value
Figure 852749DEST_PATH_IMAGE002
The identification result is obtained according to the test set, and the values are as follows:
Figure 100002_DEST_PATH_IMAGE012
where M is the number of pictures of the object to be detected in the test set,
Figure 100002_DEST_PATH_IMAGE014
the number of correctly identified pictures in the test set.
Further, comprehensive judgment and scoring are carried out according to the duration of cloud identification and the accuracy of promotion, and the formula is as follows:
Figure 100002_DEST_PATH_IMAGE016
Figure 100002_DEST_PATH_IMAGE018
the accuracy of the promotion is represented, the accuracy of the cloud model and the accuracy of the edge end are obtained through the test set, and the difference between the accuracy of the cloud model and the accuracy of the edge end is the difference
Figure 94856DEST_PATH_IMAGE018
A value of (d);
Figure 100002_DEST_PATH_IMAGE020
indicating slave transmissionThe time that the picture has passed to the cloud to return the results,
Figure 100002_DEST_PATH_IMAGE022
and the average time delay from the edge end to the cloud end is represented.
An unmanned system-oriented cloud-edge collaborative video stream processing system comprises:
the terminal unattended platform is used for cutting the collected video stream and sending each frame of cut pictures to appointed edge-end equipment for processing according to a preset distribution strategy;
the edge terminal equipment is used for responding to the picture identification request, adding the picture to be identified into a cloud sending queue to wait for sending to the cloud, and adding the picture to be identified into a local identification queue to wait for identification; determining a final recognition result of the picture based on the local recognition result and the cloud recognition result;
the cloud end is used for identifying the pictures sent by the edge end equipment and returning an identification result to the edge end equipment;
the edge terminal device determines a final recognition result of the picture based on the local recognition result and the cloud recognition result, and the determination comprises the following steps:
the edge terminal equipment identifies the picture data in the local identification queue, and according to the confidence coefficient f obtained by local picture identification, the confidence coefficient f is greater than or equal to the specified threshold value
Figure 733648DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
the edge end equipment judges whether the cloud end identification result exists or not, if the edge end equipment receives the cloud end identification result, comprehensive judgment and scoring are carried out according to the time and the accuracy of cloud end and edge end local identification, when the score is larger than or equal to a threshold value s, the cloud end identification result is judged to be superior to the local identification result, and the edge end equipment returns the cloud end identification result to the human-computer interaction interface; otherwise, the edge terminal equipment returns a local identification result;
and if the edge terminal equipment does not receive the identification result of the cloud, directly returning the local identification result to the human-computer interaction interface.
An unmanned system oriented video stream processing device comprising:
the data acquisition module is used for acquiring a picture identification request of each frame of picture after video stream cutting from the tail end unattended platform, wherein each frame of picture after video stream cutting is distributed according to a preset distribution strategy;
the data queue module is used for responding to the picture identification request, adding the picture to be identified into a cloud sending queue to wait for being sent to the cloud, and adding the picture to be identified into a local identification queue to wait for being identified;
the local picture identification module is used for identifying the picture data in the local identification queue and making the confidence coefficient f be more than or equal to a specified threshold value according to the confidence coefficient f obtained by local picture identification
Figure 706021DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
the cloud side comprehensive analysis decision module is used for judging whether a cloud side identification result is received or not, if the cloud side identification result is received, comprehensive judgment and scoring are carried out according to the time and the accuracy of the cloud side and the local identification respectively, when the score is larger than or equal to a threshold value s, the cloud side identification result is judged to be superior to the local identification result, and the cloud side identification result is returned to the human-computer interaction interface; otherwise, returning a local identification result; and if the identification result of the cloud is not received, directly returning the local identification result to the human-computer interaction interface.
A computer device, comprising:
one or more processors;
a memory; and
one or more programs, wherein the one or more programs are stored in the memory and configured for execution by the one or more processors, the programs when executed by the processors implement the steps of:
acquiring a picture identification request of each frame of picture after video stream cutting from a terminal unattended platform, wherein each frame of picture after video stream cutting is distributed according to a preset distribution strategy;
responding to the picture identification request, adding the picture to be identified into a cloud end sending queue to wait for sending to the cloud end, and adding the picture to be identified into a local identification queue to wait for identification;
identifying the picture data in the local identification queue, and according to the confidence f obtained by local picture identification, making the confidence f be greater than or equal to a specified threshold value
Figure 51551DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
judging whether a cloud identification result is received or not, if the cloud identification result is received, comprehensively judging and scoring according to the time and accuracy of cloud and local identification respectively, judging that the cloud identification result is superior to the local identification result when the score is greater than or equal to a threshold value s, and returning the cloud identification result to the human-computer interaction interface; otherwise, returning a local identification result; and if the identification result of the cloud is not received, directly returning the local identification result to the human-computer interaction interface.
Has the advantages that: according to the method, each frame of picture cut from the video stream by the edge end is sent to the edge end device with the minimum estimated waiting time for processing the picture, the edge end device also sends a copy of picture to the cloud end device after receiving the picture, the edge end and the cloud end are processed simultaneously, if the cloud end is not processed, the identification result of the edge end is obtained, if the cloud end has a return result, the return result is scored according to time delay and improvement precision to comprehensively judge, and the low ductility of the edge end and the resource sufficiency of the cloud end can be combined through cloud-edge cooperative processing, so that the efficiency of unmanned system video stream processing is improved.
Drawings
Fig. 1 is a flowchart of a method for processing a cloud-edge collaborative video stream for an unmanned system according to an embodiment of the present invention;
fig. 2 is a schematic diagram of an object detection process according to an embodiment of the invention.
Detailed Description
The technical scheme of the invention is further explained by combining the attached drawings.
The invention provides a collaborative video stream processing method and system for an unmanned system. Cloud refers to a cloud server or any platform or infrastructure that provides cloud computing services. The video stream processing method based on the system comprises the following steps: firstly, the unattended platform collects video streams and then sends pictures in the videos to the edge end device, then the edge end device and the cloud end device conduct cooperative processing, if a picture processing result is obtained through cloud end return, scoring is comprehensively judged according to time delay and precision improvement rate, whether the recognition result of the cloud end needs to be used is determined, and if the cloud end does not return the result, the picture processing result of the edge end is obtained. Referring to fig. 1, the method specifically includes the following steps:
(1) data distribution:
and the tail end unattended platform cuts the collected video stream and sends each cut frame of picture to the edge end equipment. Suppose there is
Figure DEST_PATH_IMAGE024
An edge terminal device, each edge terminal device
Figure 421484DEST_PATH_IMAGE010
Maintain a queue with
Figure 485255DEST_PATH_IMAGE006
And waiting for detection and identification of the picture.
Figure 998669DEST_PATH_IMAGE008
Is shown as
Figure 198706DEST_PATH_IMAGE010
The average time of identifying one picture by each edge terminal device is obtained by averaging the identification time of the pictures with the same type and a large number of samples in the model training process. The specific sending edge end equipment is determined by real-time task allocation. The edge end determines which close device to process the current picture through the real-time task distribution system. Real-time task scoringThe matching system selects an edge device with the smallest expected waiting time for processing the picture. Namely:
Figure 100002_DEST_PATH_IMAGE004A
each edge device acquires the neighboring edge devices through timing data synchronization
Figure DEST_PATH_IMAGE026
And
Figure DEST_PATH_IMAGE028
values and stored in a local mini-database SQLite, able to be based on
Figure 441338DEST_PATH_IMAGE026
And
Figure 461246DEST_PATH_IMAGE028
the dynamic change of the value triggers a change of the distribution strategy.
(2) And (3) cloud edge collaborative image processing:
(2-1) for edge terminal device
Figure 994996DEST_PATH_IMAGE010
When receiving a picture identification request, the terminal device is added into the cloud sending queue
Figure DEST_PATH_IMAGE030
Waiting to be sent to the cloud end, and adding the picture to be identified into a local identification queue
Figure DEST_PATH_IMAGE032
Awaiting recognition.
(2-2) whenever edge end device
Figure 128168DEST_PATH_IMAGE010
From the local queue when a picture has been identified
Figure 89171DEST_PATH_IMAGE032
And taking out the next picture for local identification processing. The identification model of the edge-end device is as follows: yolov3 was used first for detecting common objects in the video stream and then ResNet-152 or MobileNet was used for more accurate and specific classification of the detected objects and labeling them. Fig. 2 is a process diagram of target detection.
(2-3) edge terminal device
Figure 842713DEST_PATH_IMAGE010
Obtaining a confidence level after identifying a picture
Figure DEST_PATH_IMAGE034
When is coming into contact with
Figure DEST_PATH_IMAGE036
When the picture does not identify the target, the picture returns a result without identifying the target. I.e. only confidence levels greater than or equal to a specified threshold
Figure 993203DEST_PATH_IMAGE002
Is saved to be less than
Figure 902253DEST_PATH_IMAGE002
The recognition results of (2) are all discarded.
Figure 34157DEST_PATH_IMAGE002
The value of (a) is selected to maximize the accuracy of the whole edge terminal. Is provided with a test center
Figure DEST_PATH_IMAGE038
A picture of the object to be detected in
Figure DEST_PATH_IMAGE040
In the result of (A), there are
Figure 136980DEST_PATH_IMAGE014
The sheet is correct and the sheet is,
Figure 746953DEST_PATH_IMAGE002
the values of (A) are as follows:
Figure 100002_DEST_PATH_IMAGE012A
(2-4) when the picture is identified by the cloud equipment, the cloud identification model is as follows: yolov3 was used first for detecting common objects in the video stream and then ResNet-152 or MobileNet was used for more accurate and specific classification of the detected objects and labeling them. Fig. 2 is a process diagram of target detection. Cloud equipment returns detection result to edge equipment
Figure 169232DEST_PATH_IMAGE010
(2-5) if edge terminal device
Figure 206458DEST_PATH_IMAGE010
When local identification is completed, the identification result of the cloud is also received, comprehensive judgment and scoring are carried out according to the time and the accuracy of the respective identification of the cloud and the edge, and the scoring details are as follows:
Figure 100002_DEST_PATH_IMAGE016A
wherein
Figure 94780DEST_PATH_IMAGE018
Expressing the accuracy of the promotion, obtaining the accuracy of the cloud model and the accuracy of the edge end through the test set, and then performing difference to obtain the accuracy of the cloud model and the accuracy of the edge end
Figure 242864DEST_PATH_IMAGE018
The value of (c).
Figure 641353DEST_PATH_IMAGE020
Representing the time elapsed from sending the picture to the cloud returning the results,
Figure 115060DEST_PATH_IMAGE022
and the average time delay from the edge end to the cloud end is represented. Here, the
Figure DEST_PATH_IMAGE042
Represents the proportion of cloud processing time delay to end-to-end time delay, because
Figure 897202DEST_PATH_IMAGE022
Representing the one-way average delay from the edge end to the cloud end,
Figure 848978DEST_PATH_IMAGE020
indicating the time elapsed from sending the picture to returning the result to the cloud, is bi-directional, and therefore
Figure 118285DEST_PATH_IMAGE022
Multiply by 2 times and
Figure 274810DEST_PATH_IMAGE020
a comparison is made.
(2-6) when score is larger than or equal to a threshold value s, judging that the cloud identification result is superior to the edge end identification result, and returning the cloud identification result to the human-computer interaction interface by the edge end equipment; otherwise, judging that the edge end identification result is superior to the cloud end identification result, and returning the local identification result to the human-computer interaction interface by the edge end equipment. In an embodiment of the present invention, the threshold s is 0.6.
And if the edge terminal equipment does not receive the identification result of the cloud, directly returning the local identification result to the human-computer interaction interface.
(3) And outputting a result:
and after the identification is finished, returning the identified result to the user, and storing the result in the cloud database.
The present invention also provides an unmanned system-oriented video stream processing device, which corresponds to an edge device in the video stream processing system, and the device includes:
the data acquisition module is used for acquiring a picture identification request of each frame of picture after video stream cutting from the tail end unattended platform, wherein each frame of picture after video stream cutting is distributed according to a preset distribution strategy;
the data queue module is used for responding to the picture identification request, adding the picture to be identified into a cloud sending queue to wait for being sent to the cloud, and adding the picture to be identified into a local identification queue to wait for being identified;
the local picture identification module is used for identifying the picture data in the local identification queue and making the confidence coefficient f be more than or equal to a specified threshold value according to the confidence coefficient f obtained by local picture identification
Figure 731199DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
the cloud side comprehensive analysis decision module is used for judging whether a cloud side identification result is received or not, if the cloud side identification result is received, comprehensive judgment and scoring are carried out according to the time and the accuracy of the cloud side and the local identification respectively, when the score is larger than or equal to a threshold value s, the cloud side identification result is judged to be superior to the local identification result, and the cloud side identification result is returned to the human-computer interaction interface; otherwise, returning a local identification result; and if the identification result of the cloud is not received, directly returning the local identification result to the human-computer interaction interface.
The data acquisition module obtains each frame of picture to be processed according to a preset distribution strategy, wherein the distribution strategy is as follows: the video stream processing equipment synchronizes the length of a data queue of the video stream processing equipment and the average processing time of a picture at regular time, and each video stream processing equipment selects equipment with the minimum expected waiting time to process the picture according to the identification queue length and the identification time of the adjacent edge end equipment stored in the local database:
Figure DEST_PATH_IMAGE004AA
where d denotes a selected video stream processing apparatus, N denotes a total number of video stream processing apparatuses,
Figure 142458DEST_PATH_IMAGE006
at the i-th video streamThe number of pictures to be detected and identified in the management device,
Figure 735113DEST_PATH_IMAGE008
is shown as
Figure 550623DEST_PATH_IMAGE010
A video stream processing apparatus identifies an average time of a picture.
The identification model of the local picture identification module is as follows: yolov3 was used first for detecting common objects in the video stream and then ResNet-152 or MobileNet was used for more accurate and specific classification of the detected objects and labeling them.
Specified threshold for use by local picture recognition module
Figure 494308DEST_PATH_IMAGE002
The identification result is obtained according to the test set, and the values are as follows:
Figure DEST_PATH_IMAGE012AA
where M is the number of pictures of the object to be detected in the test set,
Figure 587553DEST_PATH_IMAGE014
the number of correctly identified pictures in the test set.
The cloud side comprehensive analysis decision module carries out comprehensive judgment and scoring according to the time and the accuracy rate respectively identified by the cloud side and the local, and the formula is as follows:
Figure DEST_PATH_IMAGE016AA
wherein
Figure 205354DEST_PATH_IMAGE018
The accuracy of the promotion is represented and obtained by the difference between the accuracy of the cloud and the accuracy of the local area;
Figure 676918DEST_PATH_IMAGE020
representing the time elapsed from sending the picture to the cloud returning the results,
Figure 170216DEST_PATH_IMAGE022
representing the average latency from the video stream processing device to the cloud.
The present invention also provides a computer apparatus comprising:
one or more processors;
a memory; and
one or more programs, wherein the one or more programs are stored in the memory and configured for execution by the one or more processors, the programs when executed by the processors implement the steps of:
acquiring a picture identification request of each frame of picture after video stream cutting from a terminal unattended platform, wherein each frame of picture after video stream cutting is distributed according to a preset distribution strategy;
responding to the picture identification request, adding the picture to be identified into a cloud end sending queue to wait for sending to the cloud end, and adding the picture to be identified into a local identification queue to wait for identification;
identifying the picture data in the local identification queue, and according to the confidence f obtained by local picture identification, making the confidence f be greater than or equal to a specified threshold value
Figure 68419DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
judging whether a cloud identification result is received or not, if the cloud identification result is received, comprehensively judging and scoring according to the time and accuracy of cloud and local identification respectively, judging that the cloud identification result is superior to the local identification result when the score is greater than or equal to a threshold value s, and returning the cloud identification result to the human-computer interaction interface; otherwise, returning a local identification result; and if the identification result of the cloud is not received, directly returning the local identification result to the human-computer interaction interface.
The steps executed by the processor correspond to the functions implemented by the functional modules of the video stream processing device facing the unmanned system, and the specific calculation method involved in each step may refer to the corresponding description of the video stream processing device, which is not described herein again.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting the same, and although the present invention is described in detail with reference to the above embodiments, those of ordinary skill in the art should understand that: modifications and equivalents may be made to the embodiments of the invention without departing from the spirit and scope of the invention, which is to be covered by the claims.

Claims (11)

1. A cloud edge collaborative video stream processing method for an unmanned system is characterized by comprising the following steps:
the terminal unattended platform cuts the collected video stream, and sends each cut frame of picture to a designated edge terminal device for processing according to a preset distribution strategy;
the edge terminal equipment responds to the picture identification request, adds the picture to be identified into a cloud sending queue to wait for sending to the cloud, and adds the picture into a local identification queue to wait for local identification;
the edge terminal equipment identifies the picture data in the local identification queue;
the edge terminal equipment identifies the obtained confidence coefficient f according to the local picture, and the confidence coefficient f is larger than or equal to the specified threshold value
Figure DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
the edge end equipment judges whether a cloud end identification result is received or not, if the edge end equipment receives the cloud end identification result, comprehensive judgment and scoring are carried out according to the time and the accuracy of the cloud end and the local identification of the edge end, when the score is larger than or equal to a threshold value s, the cloud end identification result is judged to be superior to the local identification result, and the edge end equipment returns the cloud end identification result to the human-computer interaction interface; otherwise, the edge terminal equipment returns a local identification result;
and if the edge terminal equipment does not receive the identification result of the cloud, directly returning the local identification result to the human-computer interaction interface.
2. The unmanned-system-oriented cloud-edge collaborative video stream processing method according to claim 1, wherein the preset allocation policy includes: selecting the edge equipment with the minimum expected waiting time to process the picture according to the identification queue length and the identification time of the adjacent edge equipment stored in the local database of each edge equipment:
Figure DEST_PATH_IMAGE004
where d represents the selected edge device, N represents the total number of edge devices,
Figure DEST_PATH_IMAGE006
indicating the number of pictures waiting to be identified in the ith edge device,
Figure DEST_PATH_IMAGE008
is shown as
Figure DEST_PATH_IMAGE010
The individual edge devices identify the average time of a picture.
3. The unmanned-system-oriented cloud-edge collaborative video streaming method according to claim 1, wherein the specified threshold is set
Figure 217118DEST_PATH_IMAGE002
The identification result is obtained according to the test set, and the values are as follows:
Figure DEST_PATH_IMAGE012
where M is the number of pictures of the object to be detected in the test set,
Figure DEST_PATH_IMAGE014
the number of correctly identified pictures in the test set.
4. The unmanned-system-oriented cloud-edge collaborative video stream processing method according to claim 1, wherein comprehensive judgment and scoring are performed according to time and accuracy rates respectively identified by a cloud end and an edge end, and a formula is as follows:
Figure DEST_PATH_IMAGE016
wherein
Figure DEST_PATH_IMAGE018
The accuracy of the promotion is represented, and the accuracy is obtained by making a difference between the accuracy of the cloud end and the accuracy of the edge end;
Figure DEST_PATH_IMAGE020
representing the time elapsed from sending the picture to the cloud returning the results,
Figure DEST_PATH_IMAGE022
and the average time delay from the edge end to the cloud end is represented.
5. The unmanned-system-oriented cloud-edge collaborative video stream processing method according to claim 1, wherein a local identification model of an edge device performs primary detection by using Yolov3, and performs secondary detection and marking on an object which is detected for the first time by using ResNet-152 or MobileNet.
6. An unmanned system-oriented cloud-edge collaborative video stream processing system is characterized by comprising:
the terminal unattended platform is used for cutting the collected video stream and sending each frame of cut pictures to appointed edge-end equipment for processing according to a preset distribution strategy;
the edge terminal equipment is used for responding to the picture identification request, adding the picture to be identified into a cloud sending queue to wait for sending to the cloud, and adding the picture to be identified into a local identification queue to wait for identification; determining a final recognition result of the picture based on the local recognition result and the cloud recognition result;
the cloud end is used for identifying the pictures sent by the edge end equipment and returning an identification result to the edge end equipment;
the edge terminal device determines a final recognition result of the picture based on the local recognition result and the cloud recognition result, and the determination comprises the following steps:
the edge terminal equipment identifies the picture data in the local identification queue, and according to the confidence coefficient f obtained by local picture identification, the confidence coefficient f is greater than or equal to the specified threshold value
Figure 604456DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
the edge end equipment judges whether a cloud end identification result is received or not, if the edge end equipment receives the cloud end identification result, comprehensive judgment and scoring are carried out according to the time and the accuracy of the cloud end and the local identification of the edge end, when the score is larger than or equal to a threshold value s, the cloud end identification result is judged to be superior to the local identification result, and the edge end equipment returns the cloud end identification result to the human-computer interaction interface; otherwise, the edge terminal equipment returns a local identification result;
and if the edge terminal equipment does not receive the identification result of the cloud, directly returning the local identification result to the human-computer interaction interface.
7. The unmanned-system-oriented cloud-edge collaborative video stream processing system according to claim 6, wherein the edge device performs comprehensive judgment and scoring according to the time and accuracy respectively identified by the cloud and the edge, and the formula is as follows:
Figure DEST_PATH_IMAGE016A
wherein
Figure 601362DEST_PATH_IMAGE018
The accuracy of the promotion is represented, and the accuracy is obtained by making a difference between the accuracy of the cloud end and the accuracy of the edge end;
Figure 751590DEST_PATH_IMAGE020
representing the time elapsed from sending the picture to the cloud returning the results,
Figure 438923DEST_PATH_IMAGE022
and the average time delay from the edge end to the cloud end is represented.
8. The unmanned-system-oriented cloud-edge collaborative video streaming system according to claim 6, wherein the preset allocation policy comprises: selecting the edge equipment with the minimum expected waiting time to process the picture according to the identification queue length and the identification time of the adjacent edge equipment stored in the local database of each edge equipment:
Figure DEST_PATH_IMAGE004A
where d represents the selected edge device, N represents the total number of edge devices,
Figure 376923DEST_PATH_IMAGE006
indicating the number of pictures waiting to be identified in the ith edge device,
Figure 251338DEST_PATH_IMAGE008
is shown as
Figure 473765DEST_PATH_IMAGE010
The individual edge devices identify the average time of a picture.
9. The unmanned-system-oriented cloud-edge-collaborative video streaming system of claim 6, wherein the specified threshold is
Figure 281184DEST_PATH_IMAGE002
The identification result is obtained according to the test set, and the values are as follows:
Figure DEST_PATH_IMAGE012A
where M is the number of pictures of the object to be detected in the test set,
Figure 232829DEST_PATH_IMAGE014
the number of correctly identified pictures in the test set.
10. An unmanned system-oriented video stream processing apparatus, comprising:
the data acquisition module is used for acquiring a picture identification request of each frame of picture after video stream cutting from the tail end unattended platform, wherein each frame of picture after video stream cutting is distributed by the tail end unattended platform according to a preset distribution strategy;
the data queue module is used for responding to the picture identification request, adding the picture to be identified into a cloud sending queue to wait for being sent to the cloud, and adding the picture to be identified into a local identification queue to wait for being identified;
the local picture identification module is used for identifying the picture data in the local identification queue and making the confidence coefficient f be more than or equal to a specified threshold value according to the confidence coefficient f obtained by local picture identification
Figure 860119DEST_PATH_IMAGE002
Is recognized as a resultSaved as a correct recognition result;
the cloud side comprehensive analysis decision module is used for judging whether a cloud side identification result is received or not, if the cloud side identification result is received, comprehensive judgment and scoring are carried out according to the time and the accuracy of the cloud side and the local identification respectively, when the score is larger than or equal to a threshold value s, the cloud side identification result is judged to be superior to the local identification result, and the cloud side identification result is returned to the human-computer interaction interface; otherwise, returning a local identification result; and if the identification result of the cloud is not received, directly returning the local identification result to the human-computer interaction interface.
11. A computer device, comprising:
one or more processors;
a memory; and
one or more programs, wherein the one or more programs are stored in the memory and configured for execution by the one or more processors, the programs when executed by the processors implement the steps of:
acquiring a picture identification request of each frame of picture after video stream cutting from the tail-end unattended platform, wherein each frame of picture after video stream cutting is distributed by the tail-end unattended platform according to a preset distribution strategy;
responding to the picture identification request, adding the picture to be identified into a cloud end sending queue to wait for sending to the cloud end, and adding the picture to be identified into a local identification queue to wait for identification;
identifying the picture data in the local identification queue, and according to the confidence f obtained by local picture identification, making the confidence f be greater than or equal to a specified threshold value
Figure 102882DEST_PATH_IMAGE002
The recognition result of (2) is saved as a correct recognition result;
judging whether a cloud identification result is received or not, if the cloud identification result is received, comprehensively judging and scoring according to the time and accuracy of cloud and local identification respectively, judging that the cloud identification result is superior to the local identification result when the score is greater than or equal to a threshold value s, and returning the cloud identification result to the human-computer interaction interface; otherwise, returning a local identification result; and if the identification result of the cloud is not received, directly returning the local identification result to the human-computer interaction interface.
CN202210117381.6A 2022-02-08 2022-02-08 Cloud-edge collaborative video stream processing method and system for unmanned system Active CN114154018B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210117381.6A CN114154018B (en) 2022-02-08 2022-02-08 Cloud-edge collaborative video stream processing method and system for unmanned system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210117381.6A CN114154018B (en) 2022-02-08 2022-02-08 Cloud-edge collaborative video stream processing method and system for unmanned system

Publications (2)

Publication Number Publication Date
CN114154018A true CN114154018A (en) 2022-03-08
CN114154018B CN114154018B (en) 2022-05-10

Family

ID=80450257

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210117381.6A Active CN114154018B (en) 2022-02-08 2022-02-08 Cloud-edge collaborative video stream processing method and system for unmanned system

Country Status (1)

Country Link
CN (1) CN114154018B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116366649A (en) * 2023-06-01 2023-06-30 中电云脑(天津)科技有限公司 Side cloud cooperative electroencephalogram data task scheduling method and system

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160057188A1 (en) * 2011-11-09 2016-02-25 Microsoft Technology Licensing, Llc Generating and updating event-based playback experiences
CN107145816A (en) * 2017-02-24 2017-09-08 北京悉见科技有限公司 Object identifying tracking and device
CN109872406A (en) * 2019-01-23 2019-06-11 北京影谱科技股份有限公司 A kind of gradual judgement update method, device and face punch card system
CN111368824A (en) * 2020-02-24 2020-07-03 河海大学常州校区 Instrument identification method, mobile device and storage medium
CN111400047A (en) * 2020-03-18 2020-07-10 青岛牛利智能科技有限公司 Method for detecting and identifying human face from monitoring video stream through cloud edge cooperation
CN111698470A (en) * 2020-06-03 2020-09-22 河南省民盛安防服务有限公司 Security video monitoring system based on cloud edge cooperative computing and implementation method thereof
US20200327739A1 (en) * 2012-12-10 2020-10-15 Nant Holdings Ip, Llc Interaction analysis systems and methods
CN111787280A (en) * 2020-06-30 2020-10-16 清华大学 Video real-time target tracking method and device based on edge calculation
CN111885375A (en) * 2020-07-15 2020-11-03 中国工商银行股份有限公司 Method, device, server and system for testing double-recorded video
CN112241719A (en) * 2020-10-23 2021-01-19 杭州卷积云科技有限公司 Monitoring video target real-time query method based on edge cloud convolution neural network cascade

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160057188A1 (en) * 2011-11-09 2016-02-25 Microsoft Technology Licensing, Llc Generating and updating event-based playback experiences
US20200327739A1 (en) * 2012-12-10 2020-10-15 Nant Holdings Ip, Llc Interaction analysis systems and methods
CN107145816A (en) * 2017-02-24 2017-09-08 北京悉见科技有限公司 Object identifying tracking and device
CN109872406A (en) * 2019-01-23 2019-06-11 北京影谱科技股份有限公司 A kind of gradual judgement update method, device and face punch card system
CN111368824A (en) * 2020-02-24 2020-07-03 河海大学常州校区 Instrument identification method, mobile device and storage medium
CN111400047A (en) * 2020-03-18 2020-07-10 青岛牛利智能科技有限公司 Method for detecting and identifying human face from monitoring video stream through cloud edge cooperation
CN111698470A (en) * 2020-06-03 2020-09-22 河南省民盛安防服务有限公司 Security video monitoring system based on cloud edge cooperative computing and implementation method thereof
CN111787280A (en) * 2020-06-30 2020-10-16 清华大学 Video real-time target tracking method and device based on edge calculation
CN111885375A (en) * 2020-07-15 2020-11-03 中国工商银行股份有限公司 Method, device, server and system for testing double-recorded video
CN112241719A (en) * 2020-10-23 2021-01-19 杭州卷积云科技有限公司 Monitoring video target real-time query method based on edge cloud convolution neural network cascade

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
MUHAMMAD USMAN YASEEN 等: "High Performance Video Processing in Cloud Data Centres", 《2016 IEEE SYMPOSIUM ON SERVICE-ORIENTED SYSTEM ENGINEERING (SOSE)》 *
王梓凯: "基于边缘计算和微服务的视频监控系统", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
韩义波 等: "基于云边协同的物联网任务卸载策略建模与分析", 《南阳理工学院学报》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116366649A (en) * 2023-06-01 2023-06-30 中电云脑(天津)科技有限公司 Side cloud cooperative electroencephalogram data task scheduling method and system
CN116366649B (en) * 2023-06-01 2023-09-05 中电云脑(天津)科技有限公司 Side cloud cooperative electroencephalogram data task scheduling method and system

Also Published As

Publication number Publication date
CN114154018B (en) 2022-05-10

Similar Documents

Publication Publication Date Title
US11017220B2 (en) Classification model training method, server, and storage medium
CN109582903B (en) Information display method, device, equipment and storage medium
WO2017127976A1 (en) Method for training and scheduling incremental learning cloud system and related device
CN106815254B (en) Data processing method and device
WO2022213565A1 (en) Review method and apparatus for prediction result of artificial intelligence model
WO2019019652A1 (en) Push-information processing method, apparatus, computer device, and storage medium
CN114154018B (en) Cloud-edge collaborative video stream processing method and system for unmanned system
CN114048387B (en) Content recommendation method based on big data and AI prediction and artificial intelligence cloud system
WO2017143773A1 (en) Crowdsourcing learning method and device
CN112636942B (en) Method and device for monitoring service host node
CN111753875A (en) Power information system operation trend analysis method and device and storage medium
CN112966608A (en) Target detection method, system and storage medium based on edge-side cooperation
CN107277557B (en) A kind of methods of video segmentation and system
CN111027397B (en) Comprehensive feature target detection method, system, medium and equipment suitable for intelligent monitoring network
CN106611021B (en) Data processing method and equipment
CN113660687A (en) Network difference cell processing method, device, equipment and storage medium
CN112925964A (en) Big data acquisition method based on cloud computing service and big data acquisition service system
CN112001622A (en) Health degree evaluation method, system, equipment and storage medium of cloud virtual gateway
CN111061911A (en) Target detection and tracking method, device and equipment for multi-video monitoring data
CN114580664A (en) Training analysis method and device, storage medium and electronic equipment
CN110544182B (en) Power distribution communication network fusion control method and system based on machine learning technology
CN115100124A (en) Image defect detection system based on cloud service
CN111737371B (en) Data flow detection classification method and device capable of dynamically predicting
CN112989869B (en) Optimization method, device, equipment and storage medium of face quality detection model
CN114692706A (en) Model training method, system, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant