CN114661466A - Task unloading method for intelligent workflow application in edge computing environment - Google Patents

Task unloading method for intelligent workflow application in edge computing environment Download PDF

Info

Publication number
CN114661466A
CN114661466A CN202210275145.7A CN202210275145A CN114661466A CN 114661466 A CN114661466 A CN 114661466A CN 202210275145 A CN202210275145 A CN 202210275145A CN 114661466 A CN114661466 A CN 114661466A
Authority
CN
China
Prior art keywords
task
edge
terminal
model
dag
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210275145.7A
Other languages
Chinese (zh)
Inventor
东方
阮文龙
沈典
张竞慧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202210275145.7A priority Critical patent/CN114661466A/en
Publication of CN114661466A publication Critical patent/CN114661466A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Signal Processing (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a task unloading method for intelligent workflow application in an edge computing environment, which comprises the steps of firstly constructing an edge environment model aiming at an edge system, and depicting a workflow to be executed by each terminal into a Directed Acyclic Graph (DAG); then constructing a task unloading model for the intelligent workflow task to be executed by the terminal, and serializing DAG according to an average execution time calculation method based on topological sorting; and then defining a maximum user experience (QoE) problem according to the optimization target of the terminal, wherein the user experience is determined by task execution delay and model precision used for executing the intelligent task. And finally, solving the optimization problem of DAG task unloading based on multi-agent reinforcement learning, and taking the output result of the network as an unloading decision of the DAG sub-task. The method and the system consider terminal unloading rational decision and resource competition among multiple terminals, and greatly improve the user experience of the terminal for executing the intelligent workflow task.

Description

Task unloading method for intelligent workflow application in edge computing environment
Technical Field
The invention belongs to the field of edge intelligence and the field of deep learning in computer technology, and particularly relates to a task unloading method for realizing intelligent workflow application in a scene of executing the intelligent workflow application in an edge computing environment.
Background
In recent years, the rapid development of new computing and communication technologies has driven the constant emergence of innovative mobile applications and services, such as augmented reality, virtual reality, face recognition, and autonomous driving. These applications are usually implemented by deep learning, first training a deep neural network offline through a large amount of data, and then using the neural network as the computation of the intelligent application. However, the deep learning method provides effective computation for the intelligent application, and simultaneously gives the intelligent application extremely high computation demand and storage demand, so that it is increasingly difficult for the terminal to meet the resource demand of the intelligent application. In addition, mobile applications are rapidly growing, and while providing better services to people, these applications are becoming more and more complex. Accordingly, workflow techniques are widely used in mobile applications in order to better execute complex smart applications. Workflow (workflow) technology organizes a set of subtasks to complete some complex task flow. With this technique, complex intelligent applications can often be decomposed into several related subtasks, and furthermore, the dependency relationships between the subtasks define their execution order and the requirements of data transfer. More and more intelligent applications improve the execution efficiency in this way, for example, autopilot applications, which provide the possibility of flexible execution of applications by breaking a complex autopilot application into finer-grained modules. Similar examples are face recognition applications, object and recognition applications, gesture recognition applications, etc. developed by OpenCV.
In summary, the workflow of intelligent application has the characteristics of dynamics and complexity, and the intelligent subtasks therein exhibit the characteristics of intensive computation, intensive data and sensitive delay, so that the tasks executed at the terminal with weak computing power cannot be realized. The cloud computing mode selects to transmit the applications to the cloud end for running so as to save the electric quantity of the mobile terminal and meet higher computing requirements. However, with the rapid spread of smart terminals, it is estimated that 850ZB data will be generated on the terminals in 2021, and transmitting these data to the cloud for processing causes huge network pressure and communication delay, which is unacceptable for delay-sensitive applications such as autopilot. Moving Edge Computing (MEC) is gaining increasing attention as a new computational paradigm. The MEC relieves the pressure of the network and introduces lower communication delay to the execution of the application by deploying the server to the edge end closer to the terminal. However, considering economic factors, people usually cannot provide computing power and storage resources of the cloud end at the edge end, so that it is difficult for the edge end to deploy all required services of the user, and the transmission of all applications to the edge end results in too high load at the edge end to provide satisfactory services for the user, and how to perform tasks with high computing resource requirements and low response time requirements is a main problem to be solved by the edge computing.
To solve these problems, task offloading has become a research hotspot with the aim of optimizing the execution of applications to meet the execution of high-demand tasks under limited resource constraints. The advent of edge computing has enabled people to place these applications on edge servers for execution, namely task offloading (task offloading). In a task unloading mode in the edge computing, a user transmits the whole task or a part of the task to an edge server for execution, and after the execution of the edge end is finished, the operation result of the application is sent to the terminal, so that the user obtains lower time delay and saves the energy consumption of the terminal. However, considering the limited factors of the edge end providing services, how to select the position where the task is performed to obtain the highest user experience is also a problem to be solved.
This problem has been extensively studied as edge calculations have evolved due to the importance of task offloading, however, conventional methods still have many deficiencies. From the viewpoint of unloading of the workflow, the traditional method lacks the problem of multi-model selection of the subtasks in the workflow, and the requirements and experience of the user are related to the delay of task execution and the accuracy of the model, so that the traditional method causes the unloaded service to be difficult to match the requirements of the user; secondly, the traditional method ignores the situation that a plurality of users compete for the same resource, and often causes long queuing time caused by service preemption when different users have the same service requirements; finally, the traditional method is difficult to solve the workflow unloading problem of the complex edge environment with multiple edge ends and multiple terminals in the edge computing environment, which often results in high algorithm complexity and difficulty in obtaining the optimal solution.
In summary, the existing task offloading method still has great limitation when applied to the scene of the intelligent workflow in the edge computing environment, and cannot meet the running requirements of low time delay and high precision of the edge intelligent application, and the invention is generated thereby.
Disclosure of Invention
The invention aims to provide a task unloading method for intelligent workflow application in an edge computing environment, which is based on multi-agent reinforcement learning and a workflow task segmentation method and can solve the problems of low user experience, high algorithm complexity and the like of the existing task unloading method applied to the intelligent workflow task in the edge computing environment in the prior art. The method carries out priority sequencing on the workflow so as to reduce the complexity of the workflow, simultaneously describes a multi-model processing method of intelligent subtasks in the workflow so as to meet the user requirements, and finally constructs a workflow task unloading model based on multi-agent reinforcement learning to solve the task unloading problem, thereby improving the user experience brought by application execution.
In order to achieve the purpose, the invention provides the following technical scheme:
the task unloading method for the intelligent workflow application in the edge computing environment comprises the following steps:
step 1, modeling an environment model and a task of an edge system, wherein the system comprises a plurality of edge terminals and a plurality of terminals, a plurality of services are provided for the terminals to select unloading execution on each edge terminal, and each terminal has an intelligent workflow of the terminal and needs to carry out unloading execution; constructing intelligent workflow tasks needing to be unloaded on a terminal into a Directed Acyclic Graph (DAG) model, wherein in the DAG model, points in the graph represent a set of intelligent subtasks forming the workflow, and edges in the graph represent a dependency relationship among the subtasks;
step 2, based on the DAG model provided in the step 1, adopting an algorithm based on topological sorting to obtain the average execution time of each subtask, and serializing the DAG according to the sequence, thereby determining the execution sequence of the subtasks in the workflow;
step 3, modeling a task unloading model of the workflow in the edge environment based on the environment model and the task model constructed in the steps 1 and 2, namely, determining the execution position of each subtask in the DAG, so that an unloading decision is constructed into an integer planning problem; according to an optimization target of a terminal, describing a problem into an optimization problem which maximizes user experience QoE, wherein the QoE comprises model precision used when an intelligent subtask is executed and total DAG execution time;
step 4, designing a limited breadth first search algorithm to solve DAG execution time delay based on the optimization target constructed in the step 3, and finally obtaining the total QoE obtained by the unloading scheme;
and 5, designing a task unloading algorithm based on multi-agent reinforcement learning to solve the optimization problem by combining with the DAG unloading delay solving method in the step 4, obtaining a stable inference model through training a deep neural network, taking the output of the inference model as an unloading strategy executed by a workflow, and applying the unloading strategy to improve the QoE.
Further, the step 1 comprises the following sub-steps:
step 101, constructing a service model included in the edge system, and defining S as a service set provided by the system for w services provided by the edge system, wherein S iswRepresenting the w-th service, the edge system is executed by using u machine learning models with different precision and time for each intelligent service and using MDw,uIndicates execution of SwThe u-th model available;
102, constructing an edge end model contained in an edge system, and aiming at the edge systemE edge terminals included, using ESeRepresents the e-th edge, each edge providing a certain number of services, using SeRepresenting a set of services provided by the edge terminal, SwFor the w-th service, the number of MD is definede wDenotes SwAt ESeThe model of the corresponding deployment; finally, use
Figure BDA0003555572180000031
Representing marginal end
Figure BDA0003555572180000032
And
Figure BDA0003555572180000033
the data transmission rate therebetween;
103, modeling a workflow model to be executed by the terminal in the edge system, and using N for i terminals included in the edge systemiRepresenting the ith terminal, each terminal comprises a workflow task needing to be unloaded, describing the workflow task into a DAG model, and using Gi={Vi,EGiDenotes the DAG model on the ith terminal; wherein, V is usedi={Ti,1,Ti,2....Ti,jDenotes the set of nodes in the DAG, i.e. the set of subtasks in the workflow, Ti,jDenotes the jth subtask, idx, of the ith terminali,jRepresents Ti,jThe corresponding service sequence number.
Further, the step 2 comprises the following sub-steps:
step 201: for each subtask in the DAG, the computation load requirement C executed by the subtask is obtainedi,jTerminal calculation force fvmEdge-end average calculated force fueCalculating the local execution delay of the subtask by using the following formula
Figure BDA0003555572180000034
And execution latency at the edge end
Figure BDA0003555572180000035
Figure BDA0003555572180000036
Figure BDA0003555572180000037
Step 202: obtaining, for each subtask in the DAG, an upload data volume requirement to be executed by each subtask
Figure BDA0003555572180000038
Download data volume requirement
Figure BDA0003555572180000041
Average data transmission rate R of terminal and edge terminali,jCalculating the upload delay of the subtask using the following formula
Figure BDA0003555572180000042
And download latency
Figure BDA0003555572180000043
Figure BDA0003555572180000044
Figure BDA0003555572180000045
Step 203: based on the time summation obtained in step 201 and step 202
Figure BDA0003555572180000046
And (3) representing the average execution time delay of the task, and applying the following rules to the subtasks in the DAG according to the time to obtain the priority:
Figure BDA0003555572180000047
wherein T isi,k∈child(Ti,j) Represents Ti,kIs Ti,jThe predecessor task of (1) needs to precede Ti,jAnd executing, so that the priority of each subtask is obtained by applying topological sorting, and a sorted task sequence is obtained.
Further, the step 3 comprises the following sub-steps:
step 301, the unloading decision model for the workflow is described as follows: for Ti,jI sub-task of the represented i terminal, offe i,jIndicating that the task is unloaded to the e-th edge end;
step 302, describing the user experience QoE as follows: the QoE comprehensively considers the task execution delay and the adopted model precision, and the execution delay part comprises the following parts: for task Ti,jUsing tqueue i,jThe queuing time before the task is executed is represented, and the queuing time is difficult to directly calculate due to the consideration of the resource preemption problem on the edge end; using texe i,jRepresenting the execution delay of the task, determined by the selected task execution model, for two connected nodes in the DAG,
Figure BDA0003555572180000048
representing data transmission delay generated when two subtasks with dependency relationship are unloaded to different edge terminals; in the model accuracy part adopted, q is usede i,jRepresenting model accuracy weighted by user demand;
step 303, depicting the optimization problem of maximizing QoE, and comprehensively considering the effects of delay and model accuracy, where the optimization objective is expressed as:
Figure BDA0003555572180000049
s.t.zi,j,k∈Se
wherein the optimization objective first item represents for workflow GiThe sum of the execution, queuing and transmission time of the generated tasks, considering that the limited number of services on the edge terminal are occupied by multiple users and multiple loads, the queuing time is difficult to directly calculate, and T is usedi finishRepresenting an execution delay; the second item of the optimization target is the sum of precision weighting of the task execution usage model, the sum of the two items is the final QoE, and the target is the maximum QoE; the constraint representing ES only at the edgeeThereon is deployed with Ti,jThe corresponding model can select Ti,jOff-loading to ESe
Further, the step 4 comprises the following sub-steps:
step 401, initialize a data structure including a priority queue PQ1Taking the starting time of the task as the priority for the executable task queue; priority queue PQ2Taking the task ending time as a task queue for the executing task queue; table map1Representing a set of tasks that are not executable due to resource constraints, list L1Representing an edge terminal service resource quantity list; initializing DAG, edge end service resource quantity list and unloading scheme as input, and enqueuing PQ to DAG start node1
Step 402, determine executable task queue PQ1If the value is null, entering step 403 if the value is not null, or entering step 404 if the value is not null;
step 403, dequeuing the tasks in sequence, determining whether there is a resource at the edge, enqueuing if there is a resource, and executing queue PQ2And the number of corresponding edge resources is reduced, and go to step 403; otherwise, adding the non-executable task list map1In the process, waiting for the edge terminal to release the resource, and proceeding to step 404;
step 404, enqueue PQ being executed2Dequeue task, on behalf of execution of task, enqueue PQ the dequeue task's successor in DAG1Meanwhile, releasing the resources corresponding to the edge terminal, and proceeding to step 405;
step 405, retrieving map when releasing edge resources1Whether or not to wrap inIncluding the task, if the task is included, the task is located in map1In-team on-the-fly queue PQ2Step 406 is entered;
step 406, determining whether all the tasks in the DAG have been executed, and if so, then PQ is executed2The completion time of the last task dequeued is taken as the total execution delay of the offload scenario and exits, otherwise step 402 is entered.
Further, the step 5 comprises the following sub-steps:
step 501, constructing a state element of reinforcement learning based on the environment information constructed in step 1, namely, the edge server information, the network environment information, the service deployment information in the environment, and the terminal intelligent workflow task information, wherein S ═ S { (S) }1,s2…snRepresenting the state information received by the ith terminal;
step 502, based on the task unloading decision in step 3, using the unloading decision of the terminal task to form a reinforcement learning action element, wherein a is a set of all terminal unloading decisions,
Figure BDA0003555572180000051
represents the set of offload decisions for the serialized DAG on the ith terminal, S '═ S'1,s′2…s′nRepresents new environment information that will be formed by executing a under the condition that the environment is S;
step 503, based on step 3, using the optimized QoE to form the reward elements of reinforcement learning, wherein RD ═ RD1,rd2,...,rdnDenotes the reward obtained after performing the action in step 502;
step 504, constructing a multi-agent reinforcement learning model, wherein the multi-agent reinforcement learning model comprises two parts, namely an operator network and a critic network; in the operator network part, for each terminal, initializing its deep learning model piiWherein each DNN is represented by siAs an input, the initial weight is wiThe DNN is composed of LSTM to embody the task priority characteristic of the serialized DAG; at the critical network part, for each terminal, DNN is initialized to a value of S,a } as an input;
step 505, initializing an experience pool to record a quadruple { state, action, reward, next state }, updating by using a round updating mode, entering step 506 if the number of rounds is less than a preset maximum training round number, and entering step 510 if the number of rounds is not less than the preset maximum training round number;
step 506, for the operator network on each terminal, using the local information siAs input, selecting action by using epsilon-greedy exploration strategy in action selection process to finally obtain action
Figure BDA0003555572180000052
Performing an action
Figure BDA0003555572180000053
To obtain new state s'iAnd earn the reward rdiRecord the quadruple as
Figure BDA0003555572180000054
Putting the obtained product into an experience pool, and entering a step 507;
step 507, for the critical network on each terminal, sampling a random minipatch sample from the experience pool, taking global information { S, A } as input, and recording the output of the target critical network as QiIndicating the output of the ith terminal critical network, and proceeding to step 508;
step 508, updating the main operator network and iteratively updating the target operator network of each terminal by using the policy gradient method, and proceeding to step 509:
Figure BDA0003555572180000061
step 509, update the critic network and iteratively update the target critic network according to a manner that minimizes the following loss, and proceed to step 505:
Figure BDA0003555572180000062
step 510, save the actor model, and then use the output of the actor as an unload decision.
Compared with the prior art, the invention has the following advantages and beneficial effects:
(1) according to the method, the intelligent workflow application is deployed in the edge computing environment, so that the division of the workflow application is realized, and a foundation is provided for a fine-grained task unloading method; the problem is reduced by adopting the mode of carrying out priority serialization on the workflow and multi-agent reinforcement learning, the complexity of the algorithm is effectively reduced, and the practicability is high, so that the method can be suitable for large-scale task environments.
(2) The invention fully considers the accuracy and time requirements of the user on the execution of the subtasks in the workflow, thereby being better suitable for the actual requirements of different intelligent workflow tasks.
(3) Because the invention is based on multi-agent reinforcement learning, the invention can effectively adapt to the dynamic and changeable conditions of the environment in the edge computing environment, effectively utilize the resources of the edge terminal and the terminal, reduce the task execution time and improve the model precision used by the intelligent subtask execution.
(4) The method of the invention considers the rational decision of terminal unloading and the resource competition among multiple terminals, and can greatly improve the user experience of the terminal for executing the intelligent workflow task.
Drawings
FIG. 1 is a schematic diagram of an edge computing execution framework of the intelligent workflow-oriented application of the present invention;
FIG. 2 is a flow diagram illustrating the execution of an intelligent workflow application in an edge computing environment according to the present invention;
FIG. 3 is a diagram of an agent model architecture for multi-agent reinforcement learning in an edge computing environment of the present invention;
FIG. 4 is a flow chart of reinforcement learning training of the present invention.
Detailed Description
The technical solutions provided by the present invention will be described in detail below with reference to specific examples, and it should be understood that the following specific embodiments are only illustrative of the present invention and are not intended to limit the scope of the present invention.
Fig. 1 shows a schematic diagram of an edge computation execution framework for an intelligent workflow application, which includes an edge layer and a terminal layer. In the terminal layer, a plurality of intelligent terminals are included, each terminal can generate an intelligent workflow task according to the requirement, however, the terminal cannot meet the delay requirement of application execution because the terminal is not powerful and the intelligent workflow task often reflects the characteristics of sensitive delay and intensive calculation. Therefore, a plurality of edge servers with certain calculation power are deployed at the edge layer, and machine learning models required by intelligent application execution are deployed in the edge servers for the terminal to select. By adopting the end-edge cooperative architecture, the terminal can unload a part of tasks to the edge layer for execution in a network transmission mode, and the edge end returns the result to the terminal after the execution of the tasks, so that the experience of application execution is improved.
The method comprises the steps of firstly, constructing an edge environment model aiming at an edge system, wherein the edge environment model comprises an intelligent service model provided by the system, an environment resource model composed of an edge terminal and a terminal, and a network environment model, and describing a workflow required to be executed by each terminal into a Directed Acyclic Graph (DAG). And then constructing a task unloading model for the intelligent workflow tasks to be executed by the terminal, serializing the DAG according to an average execution time calculation method based on topological sorting, wherein the unloading problem is that the task execution position is determined for each task in the sequence. Then, a maximum user experience (QoE) problem is defined according to the optimization objective of the terminal, wherein the user experience is determined by task execution delay and model precision used for executing the intelligent task. And finally, solving the optimization problem of DAG task unloading based on multi-agent reinforcement learning, obtaining a stable inference network through training of the deep neural network, and taking the output result of the network as an unloading decision of the DAG sub-task.
As shown in fig. 2, the invention applies a fine-grained workflow execution mode to split a complex intelligent task into a plurality of intelligent subtasks, and combines the task unloading idea to unload a part of subtasks to the edge end for execution, thereby effectively improving the parallelism of the subtasks execution while utilizing the strong computing resources of the edge end and the high-precision machine learning model, thereby reducing the task execution time, improving the model precision adopted by the intelligent task execution, and finally improving the user experience.
The specific steps executed by the intelligent workflow comprise:
step 1: modeling terminal-generated workflows into DAGs
The method comprises the steps of modeling an environment model and workflow tasks of an edge system, wherein the system environment comprises a plurality of edge terminals and a plurality of terminals, each edge terminal provides a plurality of services for the terminals to select to unload and execute, and each terminal has an intelligent workflow of the terminal and needs to unload and execute. And constructing the intelligent workflow tasks needing to be unloaded on the terminal into a Directed Acyclic Graph (DAG) model, wherein in the DAG model, points in the graph represent a set of intelligent subtasks forming the workflow, and edges in the graph represent the dependency relationship among the subtasks.
The specific content of the step is as follows:
step 101, constructing a service model included in the edge system, and defining S as a service set provided by the system for w services provided by the edge system, wherein S isWRepresenting the w-th service, the edge system can be executed using u precision, time-diverse machine learning models, using MD, for each intelligent servicew,uIndicates execution of SwThe u-th model available.
102, constructing edge end models contained in the edge system, and using ES for e edge ends contained in the edge systemeIndicating the e-th edge, each edge providing a certain number of services, using SeRepresenting a set of services provided by the edge terminal, SwFor the w-th service provision quantity, for this intelligent service, an MD is definede wDenotes SwAt ESeThe model of the corresponding deployment. Finally, use
Figure BDA0003555572180000081
Representing marginal end
Figure BDA0003555572180000082
And with
Figure BDA0003555572180000083
The data transmission rate therebetween.
103, modeling a workflow model to be executed by the terminal in the edge system, and using N for i terminals included in the edge systemiRepresenting the ith terminal, each terminal comprises a workflow task needing to be unloaded, describing the workflow task into a DAG model, and using Gi={Vi,EGiDenotes a DAG model at the ith terminal. Wherein V is usedi={Ti,1,Ti,2....Ti,jDenotes the set of nodes in the DAG, i.e. the set of subtasks in the workflow, Ti,jDenotes the jth sub-task, idx, of the ith terminali,jRepresents Ti,jA corresponding service sequence number; using EGi={egi,1,egi,2....egi,lDenotes the set of edges in the DAG, i.e. the dependencies of the sub-tasks in the workflow, where
Figure BDA0003555572180000084
Representing two subtasks
Figure BDA0003555572180000085
There is a dependency relationship between them.
And 2, step: and calculating task execution priority by using a topological sorting method based on average execution time, and sorting the subtasks to obtain a serialized task queue. And (2) on the basis of the DAG model proposed in the step 1, adopting the topological ordering-based algorithm to obtain the average execution time of each subtask for the complex DAG model, and serializing the DAG according to the sequence, thereby determining the execution sequence of the subtasks in the workflow.
The specific content of the step is as follows:
step 201: for each subtask in the DAG, the computation load requirement C executed by the subtask is obtainedi,jThe terminal has large calculation powerSmall fvmEdge-end average calculated force fueUsing the following formula, thereby calculating the execution time delay of the subtask locally
Figure BDA0003555572180000086
And execution latency at the edge end
Figure BDA0003555572180000087
Figure BDA0003555572180000088
Figure BDA0003555572180000089
Step 202: obtaining, for each subtask in the DAG, an upload data volume requirement to be executed by each subtask
Figure BDA00035555721800000810
Download data volume requirement
Figure BDA00035555721800000811
Average data transmission rate R of terminal and edge terminali,jThe following formula is used, whereby the upload delay of the subtask is calculated
Figure BDA00035555721800000812
And download latency
Figure BDA00035555721800000813
Figure BDA00035555721800000814
Figure BDA00035555721800000815
Step 203:based on the time summation obtained in step 201 and step 202
Figure BDA00035555721800000816
And (3) representing the average execution time delay of the tasks, and applying the following rules to the subtasks in the DAG according to the time to obtain the priority:
Figure BDA00035555721800000817
wherein T isi,k∈child(Ti,j) Represents Ti,kIs Ti,jThe predecessor task of (1) needs to precede Ti,jAnd executing, so that the topological sorting is applied to obtain the priority of each subtask, and a sorted task sequence is obtained.
And 3, step 3: based on the environment model and the task model constructed in the step 1 and the step 2, a task unloading model of the workflow in the edge environment is modeled, namely the execution position of each subtask in the DAG needs to be decided, so that an unloading decision is constructed into an integer programming problem. According to the optimization objective of the terminal, the problem is characterized as an optimization problem for maximizing the user experience (QoE), wherein the QoE comprises the model precision used in the execution of the intelligent subtask and the total execution time of DAG.
The specific content of the step is as follows:
step 301, the unloading decision model for the workflow is described as follows: for Ti,jJ sub-tasks of the i terminal indicated, offe i,jIndicating that the task is offloaded to the e-th edge.
Step 302, describing the user experience QoE as follows: the QoE comprehensively considers the task execution delay and the adopted model precision, and the execution delay part comprises the following parts: for task Ti,jUsing tqueue i,jThe queuing time before the task is executed is represented, and the queuing time is difficult to directly calculate due to the consideration of the resource preemption problem on the edge end; using texe i,jRepresenting the execution delay of the task, determined by the selected task execution model, for two connected DAGsThe node(s) of (a) is (are),
Figure BDA0003555572180000091
representing the data transmission delay generated when two subtasks with dependency relationship are unloaded to different edge terminals. In the model accuracy part adopted, q is usede i,jRepresenting the accuracy of the model weighted on the user demand.
Step 303, the optimization problem of maximizing QoE is characterized, and in order to comprehensively consider the effects of delay and model accuracy, the optimization objective is expressed as:
Figure BDA0003555572180000092
s.t.zi,j,k∈Se
wherein the optimization objective first item represents for workflow GiThe sum of the execution, queuing and transmission time of the generated tasks, considering that the limited number of services on the edge terminal are occupied by multiple users and multiple loads, the queuing time is difficult to directly calculate, and T is usedi finishRepresenting an execution delay; the second item of the optimization target is the sum of precision weighting of the task execution usage model, and the sum of the two items is the final QoE, wherein the target is to maximize the QoE. The constraint representing ES only at the edgeeThereon is deployed with Ti,jThe corresponding model can select Ti,jOff-loading to ESe
And inputting the task queue into a decision maker to obtain a task unloading decision of each subtask, wherein the task unloading decision can be executed at an edge terminal or a terminal.
And 4, step 4: based on the optimization target constructed in the step 3, DAG execution delay is difficult to directly solve due to the resource preemption problem of multiple terminals of the edge system, a limited breadth first search algorithm is designed to solve the DAG execution delay, and finally the total QoE obtained by the unloading scheme is obtained; the resource-limited breadth-first search DAG task execution delay algorithm comprises the following steps:
step 401, initialize a data structure, whichIncluding a priority queue PQ1Taking the starting time of the task as the priority for the executable task queue; priority queue PQ2Taking the task ending time as a task queue for the executing task queue; table map1Representing a set of tasks that are not executable due to resource constraints, list L1And representing the list of the number of the edge terminal service resources. Initializing DAG, edge end service resource quantity list and unloading scheme as input, and enqueuing PQ to DAG start nodei
Step 402, determine executable task queue PQ1If it is empty, go to step 403 if it is not empty, otherwise go to step 404.
Step 403, dequeuing the tasks in sequence, determining whether there is a resource at the edge, and enqueuing if there is a resource and executing a queue PQ2And the number of corresponding edge resources is reduced, and go to step 403; otherwise, adding the non-executable task table map1Wait for the edge to release the resource, and proceed to step 404.
Step 404, enqueue PQ being executed2Dequeue task, on behalf of execution of task, enqueue PQ the dequeue task's successor in DAG1Meanwhile, the corresponding resources of the edge terminal are released, and step 405 is proceeded.
Step 405, retrieving map when releasing edge resources1Whether the task is contained in the table, if so, the task is in the map1In-team on-the-fly queue PQ2Step 406 is entered.
Step 406, determining whether all the tasks in the DAG have been executed, and if so, then PQ is executed2And the completion time of the last task of the intermediate dequeue is used as the total execution delay of the unloading scheme, and the process is exited. Otherwise step 402 is entered.
And 5, designing a task unloading algorithm based on multi-agent reinforcement learning to solve the optimization problem by combining with the DAG unloading delay solving method in the step 4, obtaining a stable inference model through training a deep neural network, taking the output of the inference model as an unloading strategy executed by a workflow, and applying the unloading strategy to improve the QoE.
In the step, machine learning models of the terminal and the edge terminal are applied to execute the intelligent subtasks, intermediate and final results are transmitted by using a network, and finally the terminal receives a workflow execution result.
As shown in FIG. 3, the invention combines the idea of multi-agent reinforcement learning to provide a task unloading method for intelligent workflow application in edge computing environment. Corresponding to the edge computing environment, the multi-agent reinforcement learning architecture comprises a plurality of agents formed by terminals, each agent is composed of two parts of networks, namely an actor network used for selecting output actions, namely making unloading decisions, and the other part of the agent network is a critic network used for evaluating the actions output by the actor network.
Specifically, in the multi-agent reinforcement learning architecture, a centralized training and distributed inference mode is adopted. Wherein, the operator network basic structure of each agent is composed of LSTM to embody the sequential characteristic of the serialized DAG. The actor network uses the edge environment information and local workflow information which needs to be executed locally as input, and finally outputs the unloading decision of the subtask sequence through forward propagation. The criticc network of each agent is composed of 3 layers of full connection layers, global edge server information, workflow task information of all terminals and task unloading decision information of all terminals are used as input, and finally Q values are output to evaluate unloading decisions.
And the centralized global information and the Q value calculated by the critic network are used for training the actor network and the critic network of the intelligent agent by utilizing the reverse transmission of the neural network. The operator network uses a policy gradient training mode, the critic network uses a Q-learning training mode, and finally the two networks converge. After network convergence, each intelligent terminal can use an actor network and local information to carry out distributed inference so as to obtain an unloading decision maximizing user experience, centralized global information is not needed, and the effect of privacy protection is achieved.
As shown in fig. 4, in the multi-agent reinforcement learning method for solving the task offloading problem in the edge computing environment according to the present invention, in the model, the agents are embodied as a plurality of intelligent terminals in the edge computing environment, each terminal needs to select an edge server for task offloading, and a deep neural network obtained through reinforcement learning training is used to obtain an optimal offloading decision. The multi-agent reinforcement learning method specifically comprises the following steps:
step 501, constructing a state element of reinforcement learning based on the environment information constructed in step 1, namely, the edge server information, the network environment information, the service deployment information in the environment, and the terminal intelligent workflow task information, wherein S ═ S { (S) }1,s2…snAnd represents the status information received by the ith terminal.
Step 502, based on the task offloading decision of step 3, using the terminal task offloading decision to form a reinforcement learning action element, wherein a is a set of all terminal offloading decisions,
Figure BDA0003555572180000111
represents the set of offload decisions for the serialized DAG on the ith terminal, S '═ S'1,s′2…s′nDenotes execution under the condition that the environment is S
Figure BDA0003555572180000112
New context information will be formed.
Step 503, based on step 3, using the optimized QoE to form the reward elements of reinforcement learning, wherein RD ═ RD1,rd2,...,rdnDenotes the prize earned after performing the action in step 502.
Step 504, a multi-agent reinforcement learning model is constructed, which comprises two parts, namely an actor network and a critic network. In the operator network part, for each terminal, initializing its deep learning model piiWherein each DNN is represented by siAs input, 256 neurons are included with an initial weight of wiThe DNN is comprised of LSTMs to embody the task priority characteristics of the serialized DAG. In the critic network part, for each terminal, DNN is initialized with S, A as input and 256 neuronsAnd (4) forming.
Step 505, initializing an experience pool to record quadruple { state, action, reward, next state }, updating by using a round updating mode, setting the maximum number of training rounds to be 1000, if the number of rounds is less than 1000, entering step 506, otherwise, entering step 510.
Step 506, for the operator network on each terminal, using the local information siAs input, selecting an action by using an Ee-greedy exploration strategy in the action selection process to finally obtain the action
Figure BDA0003555572180000113
Performing an action
Figure BDA0003555572180000114
To obtain a new state s'iAnd earn a reward rdiRecord the quadruple as
Figure BDA0003555572180000115
Put into the experience pool and go to step 507.
Step 507, for the critical network on each terminal, sampling a random minipatch sample from the experience pool, taking global information { S, A } as input, and recording the output of the target critical network as QiAnd indicates the output of the ith terminal critical network, the process proceeds to step 508.
Step 508, updating the main operator network and iteratively updating the target operator network of each terminal by using the policy gradient method, and proceeding to step 509:
Figure BDA0003555572180000116
step 509, update the critic network and iteratively update the target critic network according to a manner that minimizes the following loss, and proceed to step 505:
Figure BDA0003555572180000121
step 510, save the actor model, and then use the output of the actor as an unload decision.
And finally, the terminal equipment takes the generated intelligent workflow task as the input of the operator network, obtains the output of the network as an unloading decision, and performs task scheduling and execution on the subtask according to the unloading decision.
The technical means disclosed in the invention scheme are not limited to the technical means disclosed in the above embodiments, but also include the technical scheme formed by any combination of the above technical features. It should be noted that those skilled in the art can make various improvements and modifications without departing from the principle of the present invention, and such improvements and modifications are also considered to be within the scope of the present invention.

Claims (6)

1. The task unloading method for the intelligent workflow application in the edge computing environment is characterized by comprising the following steps of:
step 1, modeling an environment model and a task of an edge system, wherein the system comprises a plurality of edge terminals and a plurality of terminals, a plurality of services are provided for the terminals to select unloading execution on each edge terminal, and each terminal has an intelligent workflow of the terminal and needs to carry out unloading execution; constructing intelligent workflow tasks needing to be unloaded on a terminal into a Directed Acyclic Graph (DAG) model, wherein in the DAG model, points in the graph represent a set of intelligent subtasks forming the workflow, and edges in the graph represent a dependency relationship among the subtasks;
step 2, based on the DAG model provided in the step 1, adopting an algorithm based on topological sorting to obtain the average execution time of each subtask, and serializing the DAG according to the sequence, thereby determining the execution sequence of the subtasks in the workflow;
step 3, modeling a task unloading model of the workflow in the edge environment based on the environment model and the task model constructed in the steps 1 and 2, namely, determining the execution position of each subtask in the DAG, so that an unloading decision is constructed into an integer planning problem; according to an optimization target of a terminal, describing a problem into an optimization problem which maximizes user experience QoE, wherein the QoE comprises model precision used when an intelligent subtask is executed and total DAG execution time;
step 4, designing a limited breadth first search algorithm to solve DAG execution time delay based on the optimization target constructed in the step 3, and finally obtaining the total QoE obtained by the unloading scheme;
and 5, designing a task unloading algorithm based on multi-agent reinforcement learning to solve the optimization problem by combining with the DAG unloading delay solving method in the step 4, obtaining a stable inference model through training a deep neural network, taking the output of the inference model as an unloading strategy executed by a workflow, and applying the unloading strategy to improve the QoE.
2. The method for task offload for intelligent workflow-oriented applications in edge computing environment according to claim 1, wherein the step 1 comprises the following sub-steps:
step 101, constructing a service model included in the edge system, and defining S as a service set provided by the system for w services provided by the edge system, wherein S iswRepresenting the w-th service, the edge system is executed by using u machine learning models with different precision and time for each intelligent service and MDw,uIndicates performing SwThe u-th model available;
102, constructing edge end models contained in the edge system, and using ES for e edge ends contained in the edge systemeIndicating the e-th edge, each edge providing a certain number of services, using SeRepresenting a set of services provided by the edge terminal, SwFor the w-th service, the number of MD is definede wDenotes SwAt ESeThe model of the corresponding deployment; finally, use
Figure FDA0003555572170000011
Indicating marginal end
Figure FDA0003555572170000012
And
Figure FDA0003555572170000013
the data transmission rate therebetween;
103, modeling a workflow model to be executed by the terminal in the edge system, and using N for i terminals included in the edge systemiRepresenting the ith terminal, each terminal comprises a workflow task needing to be unloaded, describing the workflow task into a DAG model, and using Gi={Vi,EGiDenotes a DAG model at the ith terminal; wherein V is usedi={Ti,1,Ti,2…Ti,jDenotes the set of nodes in the DAG, i.e. the set of subtasks in the workflow, Ti,jDenotes the jth sub-task, idx, of the ith terminali,jRepresents Ti,jThe corresponding service sequence number.
3. The method for task offloading of an intelligent workflow-oriented application in an edge computing environment according to claim 1, wherein the step 2 comprises the sub-steps of:
step 201: for each subtask in the DAG, the computation load requirement C executed by the subtask is obtainedi,jCalculating the strength of the terminal fvmEdge average calculated force fueCalculating the local execution delay of the subtask using the following formula
Figure FDA0003555572170000021
And execution latency at the edge end
Figure FDA0003555572170000022
Figure FDA0003555572170000023
Figure FDA0003555572170000024
Step 202: obtaining, for each subtask in the DAG, an upload data volume requirement to be executed by each subtask
Figure FDA0003555572170000025
Download data volume requirement
Figure FDA0003555572170000026
Average data transmission rate R of terminal and edge terminali,jCalculating the upload delay of the subtask using the following formula
Figure FDA0003555572170000027
And download latency
Figure FDA0003555572170000028
Figure FDA0003555572170000029
Figure FDA00035555721700000210
Step 203: based on the time summation obtained in step 201 and step 202
Figure FDA00035555721700000211
And (3) representing the average execution time delay of the tasks, and applying the following rules to the subtasks in the DAG according to the time to obtain the priority:
Figure FDA00035555721700000212
wherein T isi,k∈child(Ti,j) Represents Ti,kIs Ti,jThe predecessor task of (1) needs to precede Ti,jAnd executing, so that the priority of each subtask is obtained by applying topological sorting, and a sorted task sequence is obtained.
4. The method for task offload for intelligent workflow-oriented applications in edge computing environment according to claim 1, wherein the step 3 comprises the sub-steps of:
step 301, the unloading decision model for the workflow is described as follows: for Ti,jJ sub-task of the i terminal indicated, offe i,jIndicating that the task is unloaded to the e-th edge end;
step 302, describing the user experience QoE as follows: the QoE comprehensively considers the task execution delay and the adopted model precision, and the execution delay part comprises the following parts: for task Ti,jUsing tqueue i,jThe queuing time before the task is executed is represented, and the queuing time is difficult to directly calculate due to the consideration of the resource preemption problem on the edge end; using texe i,jRepresenting the execution delay of the task, determined by the selected task execution model, for two connected nodes in the DAG,
Figure FDA00035555721700000213
representing data transmission delay generated when two subtasks with dependency relationship are unloaded to different edge terminals; in the model accuracy part adopted, q is usede i,jRepresenting model accuracy weighted against user demand;
step 303, depicting the optimization problem of maximizing QoE, and comprehensively considering the effects of delay and model accuracy, where the optimization objective is expressed as:
max:
Figure FDA0003555572170000031
s.t.zi,j,k∈Se
wherein the optimization objective first item represents for workflow GiExecution of, result inThe total time of executing, queuing and transmitting the tasks is taken into consideration that the queuing time is difficult to directly obtain by considering the limited number of services on the edge terminals occupied by multiple users and multiple loads, and T is usedi finishRepresenting an execution delay; the second item of the optimization target is the sum of precision weighting of the task execution usage model, the sum of the two items is the final QoE, and the target is the maximum QoE; the constraint representing ES only at the edgeeThereon is deployed with Ti,jThe corresponding model can select Ti,jUnloading to ESe
5. The method for task offload for intelligent workflow-oriented applications in edge computing environment according to claim 1, wherein the step 4 comprises the following sub-steps:
step 401, initialize a data structure including a priority queue PQ1Taking the starting time of the task as the priority for the executable task queue; priority queue PQ2Taking the task ending time as a task queue for the executing task queue; table map1Representing a set of tasks that are not executable due to resource constraints, list L1Representing an edge terminal service resource quantity list; initializing DAG, edge end service resource quantity list and unloading scheme as input, and enqueuing PQ to DAG start node1
Step 402, determine executable task queue PQ1If the value is null, entering step 403 if the value is not null, or entering step 404 if the value is not null;
step 403, dequeuing the tasks in sequence, determining whether there is a resource at the edge, and enqueuing if there is a resource and executing a queue PQ2And the number of corresponding edge resources is reduced, and go to step 403; otherwise, adding the non-executable task list map1In the process, waiting for the edge terminal to release the resource, and proceeding to step 404;
step 404, enqueue PQ being executed2Dequeue task, on behalf of execution of task, enqueue PQ the dequeue task's successor in DAG1Meanwhile, releasing the resources corresponding to the edge terminal, and proceeding to step 405;
step 405, retrieving map when releasing edge resources1Whether the task is contained in the list, if so, the task is stored in the map1Merge queue PQ2Step 406 is entered;
step 406, determining whether all tasks in the DAG have been executed, and if so, making the PQ2The completion time of the last task dequeued is taken as the total execution delay of the offload scenario and exits, otherwise step 402 is entered.
6. The method for task offload for intelligent workflow-oriented applications in edge computing environment according to claim 1, wherein the step 5 comprises the sub-steps of:
step 501, constructing a state element of reinforcement learning based on the environment information constructed in step 1, namely, the edge server information, the network environment information, the service deployment information in the environment, and the terminal intelligent workflow task information, wherein S ═ S { (S) }1,s2…snRepresents the status information received by the ith terminal;
step 502, based on the task unloading decision in step 3, using the unloading decision of the terminal task to form a reinforcement learning action element, wherein a is a set of all terminal unloading decisions,
Figure FDA0003555572170000041
represents the set of offload decisions for the serialized DAG on the ith terminal, S '═ S'1,s′2…s′nRepresents new environment information that will be formed by executing a under the condition that the environment is S;
step 503, based on step 3, using the optimization objective QoE to construct the reward element of reinforcement learning, where RD ═ is1,rd2,…,rdnDenotes the reward obtained after performing the action in step 502;
step 504, constructing a multi-agent reinforcement learning model, wherein the multi-agent reinforcement learning model comprises two parts, namely an actor network and a critic network; in the actor network part, for each terminal, initializing the deep learning model pi thereofiWherein each DNN is represented by siAs an input, the initial weight is wiThe DNN is composed of LSTM to embody the task priority characteristic of the serialized DAG; at the critical network part, for each terminal, DNN is initialized with S, a as input;
step 505, initializing an experience pool to record a quadruple { state, action, reward, next state }, updating by using a round updating mode, entering step 506 if the number of rounds is less than a preset maximum training round number, and entering step 510 if the number of rounds is not less than the preset maximum training round number;
step 506, for the operator network on each terminal, using the local information siAs input, selecting an action by using an Ee-greedy exploration strategy in the action selection process to finally obtain the action
Figure FDA0003555572170000042
Performing an action
Figure FDA0003555572170000043
To obtain a new state s'iAnd earn a reward rdiRecord the quadruple as
Figure FDA0003555572170000044
Putting the obtained product into an experience pool, and entering a step 507;
step 507, for the critical network on each terminal, sampling a random minipatch sample from the experience pool, taking global information { S, A } as input, and recording the output of the target critical network as QiIndicating the output of the ith terminal critical network, and proceeding to step 508;
step 508, updating main operator network of each terminal and iteratively updating target operator network by using policy gradient method, and proceeding to step 509:
Figure FDA0003555572170000045
step 509, update the critic network and iteratively update the target critic network according to a manner that minimizes the following loss, and proceed to step 505:
Figure FDA0003555572170000046
step 510, save the actor model, and then use the output of the actor as an unload decision.
CN202210275145.7A 2022-03-21 2022-03-21 Task unloading method for intelligent workflow application in edge computing environment Pending CN114661466A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210275145.7A CN114661466A (en) 2022-03-21 2022-03-21 Task unloading method for intelligent workflow application in edge computing environment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210275145.7A CN114661466A (en) 2022-03-21 2022-03-21 Task unloading method for intelligent workflow application in edge computing environment

Publications (1)

Publication Number Publication Date
CN114661466A true CN114661466A (en) 2022-06-24

Family

ID=82030478

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210275145.7A Pending CN114661466A (en) 2022-03-21 2022-03-21 Task unloading method for intelligent workflow application in edge computing environment

Country Status (1)

Country Link
CN (1) CN114661466A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115941791A (en) * 2022-11-25 2023-04-07 重庆邮电大学 Hot spot service caching method and system based on server-free edge calculation
CN116562172A (en) * 2023-07-07 2023-08-08 中国人民解放军国防科技大学 Geographical scene time deduction method, device and equipment for space-time narrative
CN116662283A (en) * 2023-07-28 2023-08-29 北京孔皆数智科技有限公司 Data sharing and calculating method with end-edge cooperation
CN117632298A (en) * 2023-12-06 2024-03-01 江西理工大学 Task unloading and resource allocation method based on priority list indexing mechanism
CN117632298B (en) * 2023-12-06 2024-05-31 江西理工大学 Task unloading and resource allocation method based on priority list indexing mechanism

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115941791A (en) * 2022-11-25 2023-04-07 重庆邮电大学 Hot spot service caching method and system based on server-free edge calculation
CN116562172A (en) * 2023-07-07 2023-08-08 中国人民解放军国防科技大学 Geographical scene time deduction method, device and equipment for space-time narrative
CN116562172B (en) * 2023-07-07 2023-09-15 中国人民解放军国防科技大学 Geographical scene time deduction method, device and equipment for space-time narrative
CN116662283A (en) * 2023-07-28 2023-08-29 北京孔皆数智科技有限公司 Data sharing and calculating method with end-edge cooperation
CN116662283B (en) * 2023-07-28 2023-12-12 北京孔皆数智科技有限公司 Data sharing and calculating method with end-edge cooperation
CN117632298A (en) * 2023-12-06 2024-03-01 江西理工大学 Task unloading and resource allocation method based on priority list indexing mechanism
CN117632298B (en) * 2023-12-06 2024-05-31 江西理工大学 Task unloading and resource allocation method based on priority list indexing mechanism

Similar Documents

Publication Publication Date Title
CN108880663B (en) Space-ground integrated network resource allocation method based on improved genetic algorithm
CN111756812B (en) Energy consumption perception edge cloud cooperation dynamic unloading scheduling method
CN111858009B (en) Task scheduling method of mobile edge computing system based on migration and reinforcement learning
CN114661466A (en) Task unloading method for intelligent workflow application in edge computing environment
CN113220356B (en) User computing task unloading method in mobile edge computing
CN113064671A (en) Multi-agent-based edge cloud extensible task unloading method
CN112214301B (en) Smart city-oriented dynamic calculation migration method and device based on user preference
CN113626104B (en) Multi-objective optimization unloading strategy based on deep reinforcement learning under edge cloud architecture
CN112988285B (en) Task unloading method and device, electronic equipment and storage medium
CN113867843B (en) Mobile edge computing task unloading method based on deep reinforcement learning
CN116893861A (en) Multi-agent cooperative dependency task unloading method based on space-ground cooperative edge calculation
CN115033359A (en) Internet of things agent multi-task scheduling method and system based on time delay control
CN116886703A (en) Cloud edge end cooperative computing unloading method based on priority and reinforcement learning
CN113741999B (en) Dependency-oriented task unloading method and device based on mobile edge calculation
Paul et al. Efficient planning of multi-robot collective transport using graph reinforcement learning with higher order topological abstraction
CN113946423A (en) Multi-task edge computing scheduling optimization method based on graph attention network
Chen et al. An intelligent task offloading algorithm (iTOA) for UAV network
CN116828541A (en) Edge computing dependent task dynamic unloading method and system based on multi-agent reinforcement learning
CN117436485A (en) Multi-exit point end-edge-cloud cooperative system and method based on trade-off time delay and precision
CN116489708B (en) Meta universe oriented cloud edge end collaborative mobile edge computing task unloading method
CN116009990B (en) Cloud edge collaborative element reinforcement learning computing unloading method based on wide attention mechanism
CN116954866A (en) Edge cloud task scheduling method and system based on deep reinforcement learning
CN116204319A (en) Yun Bianduan collaborative unloading method and system based on SAC algorithm and task dependency relationship
CN114217881B (en) Task unloading method and related device
CN115220818A (en) Real-time dependency task unloading method based on deep reinforcement learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination