CN113127169B - Efficient link scheduling method for dynamic workflow in data center network - Google Patents

Efficient link scheduling method for dynamic workflow in data center network Download PDF

Info

Publication number
CN113127169B
CN113127169B CN202110373804.6A CN202110373804A CN113127169B CN 113127169 B CN113127169 B CN 113127169B CN 202110373804 A CN202110373804 A CN 202110373804A CN 113127169 B CN113127169 B CN 113127169B
Authority
CN
China
Prior art keywords
coflow
node
scheduling
tasks
nodes
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110373804.6A
Other languages
Chinese (zh)
Other versions
CN113127169A (en
Inventor
沈鸿
王鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Yat Sen University
Original Assignee
Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Yat Sen University filed Critical Sun Yat Sen University
Priority to CN202110373804.6A priority Critical patent/CN113127169B/en
Publication of CN113127169A publication Critical patent/CN113127169A/en
Application granted granted Critical
Publication of CN113127169B publication Critical patent/CN113127169B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • G06F9/4843Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
    • G06F9/4881Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/48Indexing scheme relating to G06F9/48
    • G06F2209/484Precedence
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention provides a high-efficiency link scheduling method of dynamic workflow in a data center network, which comprises the following steps: s1: using a directed acyclic graph neural network to simultaneously process the n arriving coflows; s2: forming J incomplete operation DAGs, inputting the J incomplete operation DAGs as a large non-connected directed acyclic graph into a directed acyclic graph neural network, and outputting to obtain an emmbedding vector of each node; s3: s2, obtaining n subedding vectors, using the n subedding vectors as the input of a strategy network in deep reinforcement learning, obtaining score of each node, and calculating to obtain a weighted score value of each node; s4: according to the partial DAG graphs of the current different jobs, finding out all nodes with the current ingress degree of 0, calculating the probability of each node based on the weighted score value through a softmax operation, and obtaining a coflow scheduling priority list according to the probability arrangement of the nodes; s5: and carrying out priority scheduling tasks based on the coflow scheduling priority list.

Description

Efficient link scheduling method for dynamic workflow in data center network
Technical Field
The present invention relates to the field of high performance computing, and more particularly, to a method for efficient link scheduling of dynamic workflows in a data center network.
Background
Modern parallel computing platforms (e.g., hadoop, spark, dryad) have supported processing large data sets at data centers. The processing is typically composed of multiple computing and communication stages. The computation phase involves the local operation of the servers, while the communication phase involves the transmission of data between servers in the data center network to initiate the next computation phase. Such an intermediate communication phase has a great influence on the application delay. Coflow is an abstraction proposed to model such communication modes, which represents a set of intermediate parallel data streams and is transmitted between servers to initiate the next phase.
For jobs with a single communication phase, minimizing the average completion time of the coflow may improve the latency of the job. However, for multi-stage jobs, minimizing the average Coflow completion time may not be the correct metric, and may even lead to worse performance, as it ignores the dependencies between Coflow in the job: starts-After and finish-Before. For multi-stage jobs, each job consists of multiple Cofiows, typically represented by a DAG (directed acyclic graph) that captures the dependencies between the Cofiows (Starts-After).
Although the Coflow scheduling for single-stage jobs is studied very much, the Coflow scheduling for multi-stage jobs and the dependency between Coflows are largely ignored. The problem of the Coflow scheduling of multi-stage jobs has proven to be an NP-hard problem. The problem of the Coflow scheduling of multi-stage jobs is that there are many processing difficulties and inherent factors including how to process different job DAGs, how to effectively extract characteristic information (including node information, side information, dependency relationships, etc.) of job DAGs, and factors such as different numbers of Coflow in different jobs, different numbers of parallel flows in a single Coflow. The prior art mainly has the following limitations:
existing jobs, including heuristic and approximation algorithms, focus on the Coflow scheduling of single-stage jobs. For the Coflow scheduling problem of multi-stage jobs, in Aalo, the authors simply discuss a straightforward heuristic to reduce the completion time of the multi-stage job. The heuristic solution of manual debugging simplifies the problem by giving some relaxation conditions, ensuring only an optimal solution of the roughly approximated NP-hard problem.
Therefore, we think that whether an adaptive scheduling model without manual guidance can be constructed, and the dependent Coflow from different jobs is dynamically scheduled by directly interacting with the environment, so that the sum of the completion time of the weighted jobs is optimized, and the working efficiency is improved. Weights may capture different priorities for different jobs, with the problem equating to minimizing average job completion time in the special case of all weights being equal.
Disclosure of Invention
The present invention provides a method for efficient link scheduling of dynamic workflows in a data center network, which can dynamically schedule dependent Coflow from different jobs without relying on a lot of manual debugging, so that the sum of weighted job completion times is minimized.
In order to solve the technical problems, the technical scheme of the invention is as follows: a method for efficient link scheduling of dynamic workflow in a data center network comprises the following steps:
s1: using a directed acyclic graph neural network to process n arriving coflows simultaneously, and forming a job by a plurality of dependent tasks, wherein the job is expressed by DAG;
s2: forming J incomplete operation DAGs, inputting the J incomplete operation DAGs as a large non-connected directed acyclic graph into a directed acyclic graph neural network, and outputting to obtain an emmbedding vector of each node;
s3: s2, obtaining n subedding vectors, using the n subedding vectors as the input of a strategy network in deep reinforcement learning, obtaining score of each node, and calculating to obtain a weighted score value of each node;
s4: according to the partial DAG graphs of the current different jobs, finding out all nodes with the current ingress degree of 0, calculating the probability of each node based on the weighted score value through a softmax operation, and obtaining a coflow scheduling priority list according to the probability arrangement of the nodes; nodes with current ingress degree not being 0 are temporarily stored in a reflow waiting list;
s5: and carrying out priority scheduling tasks based on the coflow scheduling priority list, updating the coflow scheduling priority list and the coflow waiting list every time one coflow is scheduled until n coflow scheduling is completed, and feeding back the quality of the report evaluating action by the environment.
Further, the updating formula of the directed acyclic graph neural network for calculating the global information and node characteristics of the job DAG is as follows:
Figure BDA0003010370770000031
Figure BDA0003010370770000032
wherein ,
Figure BDA0003010370770000033
representation of all nodes v representing layer i, h g A representation representing the entire DAG graph; p (v) represents a set of direct predecessor nodes of node v; t (T)Representing a set of nodes that are not directly successor; g l ,F l And R are both parameterized neural networks.
Still further, in step S2, J incomplete job DAGs are constructed, specifically as follows:
constructing a waiting queue W with the capacity of m and a waiting queue D with the capacity of n, wherein m > n; sequentially arranging tasks in a waiting queue W until the number of the tasks is greater than or equal to n, taking out the first n tasks and putting the first n tasks into a waiting queue D; because the n tasks come from different jobs and the sequence coming successively meets the respective dependency relationship, J incomplete operation DAGs are formed.
Further, the weighted score value of each node is obtained by multiplying the task weight corresponding to the node by the score of each node; each job corresponds to a task weight, and the task weights in the same job are the same.
Still further, step S5, specifically:
s1: performing priority scheduling tasks based on the coflow scheduling priority list; after each time a coflow is scheduled, removing corresponding nodes and edges in the DAG graph, updating a node set with the ingress degree of 0, and further updating a coflow scheduling priority list and a coflow waiting list;
s2: and continuously repeating the operation until the n coflow scheduling is completed, and feeding back the environmental feedback to evaluate whether the action is good or bad.
And further, in step S5, after the execution of the coflow scheduling priority list is finished, the queue D to be processed is updated, namely, the first n tasks are selected again from the queue W to be placed in the queue D to be processed, and the step 2 is returned to continue the execution.
Still further, step S5 evaluates the quality of the action by calculating the sum of the weighted job completion times as a reward by which the agent is optimized so that the sum of the weighted job completion times is minimized.
Compared with the prior art, the technical scheme of the invention has the beneficial effects that:
1. on the premise that the GNN can extract the characteristic of the operation DAG, the invention uses the directed acyclic graph neural network instead, makes full use of the special structure of the DAG, and can generate a more favorable vector representation of the graph. And the directed acyclic graph neural network can directly process nodes in sequence and output the ebedding vector obeying the dependency sequence.
2. In order to avoid the problem of processing a complete set of job DAGs at a time, resulting in too large an action space, requiring a long time to complete a training, and the input size of the policy network is not fixed. The invention assumes that the coflow arrival sequences in different jobs are based on their respective dependency relationships, and only n tasks in the queue D to be processed need to be processed each time, so that the training speed is faster, the action space is small, and the input size of the strategy network is also fixed.
3. Considering the importance of different jobs, the optimization goal is to weight the sum of job completion times. For optimization purposes, the invention considers the influence of the weight on the basis of the policy network output score, and the input of softmax is also a weighted score value, so that the reinforcement learning agent can learn the optimal action more easily.
Drawings
Fig. 1 is a schematic diagram of a dynamic Coflow scheduling method according to this embodiment.
Fig. 2 is a multi-stage job DAG provided in this embodiment, in which nodes represent computation stages and edges represent communication stages between nodes.
Fig. 3 is a schematic diagram of the directed acyclic graph neural network according to the embodiment for processing different job DAGs.
Detailed Description
The technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only some embodiments of the present invention, which are only for illustration and not to be construed as limitations of the present patent. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The technical scheme of the invention is further described below with reference to the accompanying drawings and examples.
Example 1
As shown in fig. 1, a method for efficient link scheduling of dynamic workflows in a data center network includes the following steps:
s1: using a directed acyclic graph neural network to process n arriving coflows simultaneously, and forming a job by a plurality of dependent tasks, wherein the job is expressed by DAG;
s2: forming J incomplete operation DAGs, inputting the J incomplete operation DAGs as a large non-connected directed acyclic graph into a directed acyclic graph neural network, and outputting to obtain an emmbedding vector of each node;
s3: s2, obtaining n subedding vectors, using the n subedding vectors as the input of a strategy network in deep reinforcement learning, obtaining score of each node, and calculating to obtain a weighted score value of each node;
s4: according to the partial DAG graphs of the current different jobs, finding out all nodes with the current ingress degree of 0, calculating the probability of each node based on the weighted score value through a softmax operation, and obtaining a coflow scheduling priority list according to the probability arrangement of the nodes; nodes with current ingress degree not being 0 are temporarily stored in a reflow waiting list;
s5: and carrying out priority scheduling tasks based on the coflow scheduling priority list, updating the coflow scheduling priority list and the coflow waiting list every time one coflow is scheduled until n coflow scheduling is completed, and feeding back the quality of the report evaluating action by the environment.
The present embodiment does not employ a Graph Neural Network (GNN), and in general, the most common GNN architecture aggregates information from neighbors based on messaging. This common GNN architecture can handle undirected graphs efficiently, but for a job DAG that contains dependencies, it may not be possible to extract its intrinsic features efficiently, which are also needed for our neural network (dependency or partial ordering). Therefore, to obtain higher predictive power, acyclic graph neural networks (DAGNNs) are used to integrate this information into the architectural representation.
In a specific embodiment, the update formula of the directed acyclic graph neural network for calculating the global information and node characteristics of the job DAG is as follows:
Figure BDA0003010370770000051
Figure BDA0003010370770000052
wherein ,
Figure BDA0003010370770000053
representation of all nodes v representing layer i, h g A representation representing the entire DAG graph; p (v) represents a set of direct predecessor nodes of node v; t represents a set of nodes that are not directly successor; g l ,F l And R are both parameterized neural networks.
It can be seen that the acyclic graph neural network is all using the information of the current layer, i.e. the acyclic graph neural network always uses the latest information to update the node representation. Because it uses only the predecessor node of the current node for aggregation and no successor node is needed, the same layer of information is used when updating the current node
Figure BDA0003010370770000054
These differences are all unique to DAG. A vector representation of a more advantageous graph is generated by using this structure appropriately. The main idea of the acyclic graph neural network is to process nodes according to the partial order defined by the DAG.
In contrast to previous GNN-processed undirected graphs: (1) The acyclic graph neural network updates the representation of node v directly with information of the current layer, but not the previous layer. (2) The acyclic graph neural network aggregates only the precursor nodes of node v, not the entire neighbor node.
In a specific embodiment, step S2, a common practice in the prior art is to input, for a single job DAG, the topology of the single job DAG and the flow information of each coflow together into GNN, to obtain an ebedding vector of each node. However, this embodiment is directed to a plurality of different job DAGs, and arrival times of different jobs and coflows are random. It is assumed that the order of coflow arrival in different jobs is based on their respective dependencies. A job set comprises J jobs, each job is composed of a plurality of dependent tasks (coflow). A waiting queue W of capacity m and a waiting queue D of capacity n are constructed, where m is much greater than n. With the arrival of the tasks, the tasks are sequentially arranged in a waiting queue W, and the first n tasks are taken out and put into a waiting queue D when the number of the tasks is greater than or equal to n. The n tasks (coflow) come from different jobs, but the sequence of arrival meets the respective dependency relationship, so that J incomplete jobs DAG are formed. The J incomplete job DAGs (which may be considered as a large, connectionless, directed acyclic graph as a whole) are then taken together as inputs to the DAGNN, through which the emmbedding vector for each node is output.
In a specific embodiment, n enabling vectors are obtained in step S2, n enabling vectors output by DAGNN are used as input of a Policy-Network in deep reinforcement learning, and score of each node is obtained through training by a deep reinforcement learning agent. In short, the policy network maps the ebadd vector to a scalar value (score).
Because each job corresponds to a task weight, the task weights in the same job are the same, and the node corresponding weight is multiplied by the score of each node to obtain the weighted score value of each node.
In a specific embodiment, step S5, specifically:
s1: performing priority scheduling tasks based on the coflow scheduling priority list; after each time a coflow is scheduled, removing corresponding nodes and edges in the DAG graph, updating a node set with the ingress degree of 0, and further updating a coflow scheduling priority list and a coflow waiting list;
s2: and continuously repeating the operation until the n coflow scheduling is completed, and feeding back the environmental feedback to evaluate whether the action is good or bad.
In a specific embodiment, a priority scheduling task (coflow) is performed based on a coflow scheduling priority list, and the coflow scheduling priority list and the coflow waiting list are updated every time one coflow is scheduled until n coflow scheduling is completed, and environmental feedback is performed to evaluate whether the action is good or bad. In this embodiment, the coflow scheduling priority list and the coflow waiting list are updated, and since n coflow is divided into two lists according to whether the ingress is 0, each time a coflow is processed, it is deleted from the working DAG, which results in that the node whose ingress is not 0 is now ingress 0, so that the coflow scheduling priority list and the coflow waiting list need to be updated.
The embodiment abstracts the data center network into a huge non-blocking switch, and the task only competes bandwidth resources at ports. Each coflow is converted into a demand matrix, each value representing the size of the stream to be transmitted from an ingress port to an egress port. For each partial job DAG graph, it can be seen as a complete DAG graph.
The present embodiment is therefore based on obtaining the sum of the weighted job completion times for the current entire schedule, i.e., reorder. Reward evaluates the quality of the action, guides the Agent to advance towards the wanted direction, and finally learns the optimal strategy by continuously and directly interacting with the environment. The agent performs the action (i.e. gets the priority list) on the environment, changing the state of the environment, and the environment feeds back a report to evaluate the current action, which aims to maximize the sum of the jackpots. At the beginning of training, the selected actions are certainly bad, the selected actions are slowly trained, and the intelligent body learns good strategies through feedback of the environment. Through training, the intelligent agent can execute an optimal strategy every time J working DAGs arrive, so that the sum of the completion time of the weighted operation is minimum.
In this embodiment, after the execution of the coflow scheduling priority list is completed, the pending queue D (state transition) is updated, that is, the first n tasks are selected again from the pending queue W and put into the pending queue D, and the step 2 is returned to continue the execution.
This embodiment optimizes the sum of weighted job completion times, with different jobs having different importance and assigned different weights. On the basis of the policy network output to obtain the score of each node, the influence of the weight is considered, and the input of softmax is also a weighted score value, so that the sum of the optimization target-minimized weighted operation completion time is consistent with that of the optimization target.
It is to be understood that the above examples of the present invention are provided by way of illustration only and not by way of limitation of the embodiments of the present invention. Other variations or modifications of the above teachings will be apparent to those of ordinary skill in the art. It is not necessary here nor is it exhaustive of all embodiments. Any modification, equivalent replacement, improvement, etc. which come within the spirit and principles of the invention are desired to be protected by the following claims.

Claims (6)

1. A high-efficiency link scheduling method of dynamic workflow in a data center network is characterized in that: the method comprises the following steps:
s1: using a directed acyclic graph neural network to process n arriving coflows simultaneously, and forming a job by a plurality of dependent tasks, wherein the job is expressed by DAG;
s2: forming J incomplete operation DAGs, inputting the J incomplete operation DAGs as a large non-connected directed acyclic graph into a directed acyclic graph neural network, and outputting to obtain an emmbedding vector of each node;
s3: s2, obtaining n subedding vectors, using the n subedding vectors as the input of a strategy network in deep reinforcement learning, obtaining score of each node, and calculating to obtain a weighted score value of each node;
s4: according to the partial DAG graphs of the current different jobs, finding out all nodes with the current ingress degree of 0, calculating the probability of each node based on the weighted score value through a softmax operation, and obtaining a coflow scheduling priority list according to the probability arrangement of the nodes; nodes with current ingress degree not being 0 are temporarily stored in a reflow waiting list;
s5: performing priority scheduling tasks based on the coflow scheduling priority list; updating a coflow scheduling priority list and a coflow waiting list every time one coflow is scheduled until n coflow scheduling is completed, and feeding back the quality of a report evaluation action by the environment;
step S2, constructing J incomplete operation DAGs, which are specifically as follows:
constructing a waiting queue W with the capacity of m and a waiting queue D with the capacity of n, wherein m > n; sequentially arranging tasks in a waiting queue W until the number of the tasks is greater than or equal to n, taking out the first n tasks and putting the first n tasks into a waiting queue D; because the n tasks come from different jobs and the sequence coming successively meets the respective dependency relationship, J incomplete operation DAGs are formed.
2. The efficient link scheduling method for dynamic workflows in a data center network according to claim 1, wherein: the directed acyclic graph neural network is used for calculating global information and node characteristics of the operation DAG, and the formula is as follows:
Figure FDA0004005892980000011
Figure FDA0004005892980000012
wherein ,
Figure FDA0004005892980000013
representation of all nodes v representing layer i, h g A representation representing the entire DAG graph; p (v) represents a set of direct predecessor nodes of node v; t represents a set of nodes that are not directly successor; g l ,F l And R are both parameterized neural networks.
3. The efficient link scheduling method for dynamic workflows in a data center network according to claim 2, wherein: the weighted score value of each node is obtained by multiplying the task weight corresponding to the node by the score of each node; each job corresponds to a task weight, and the task weights in the same job are the same.
4. A method of efficient link scheduling of dynamic workflows in a data center network according to claim 3, wherein: step S5, specifically:
s1: performing priority scheduling tasks based on the coflow scheduling priority list; after each time a coflow is scheduled, removing corresponding nodes and edges in the DAG graph, updating a node set with the ingress degree of 0, and further updating a coflow scheduling priority list and a coflow waiting list;
s2: and continuously repeating the operation until the n coflow scheduling is completed, and feeding back the environmental feedback to evaluate whether the action is good or bad.
5. The efficient link scheduling method for dynamic workflows in a data center network of claim 4, wherein: and S5, after the execution of the coflow scheduling priority list is finished, updating the queue D to be processed, namely, selecting the first n tasks from the queue W again and putting the tasks into the queue D to be processed, and returning to the step 2 to continue the execution.
6. The efficient link scheduling method for dynamic workflows in a data center network of claim 5, wherein: and S5, evaluating the quality of the action by calculating the sum of the completion time of the weighted operation as a reward, and optimizing the agent by the reward so as to minimize the sum of the completion time of the weighted operation.
CN202110373804.6A 2021-04-07 2021-04-07 Efficient link scheduling method for dynamic workflow in data center network Active CN113127169B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110373804.6A CN113127169B (en) 2021-04-07 2021-04-07 Efficient link scheduling method for dynamic workflow in data center network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110373804.6A CN113127169B (en) 2021-04-07 2021-04-07 Efficient link scheduling method for dynamic workflow in data center network

Publications (2)

Publication Number Publication Date
CN113127169A CN113127169A (en) 2021-07-16
CN113127169B true CN113127169B (en) 2023-05-02

Family

ID=76775168

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110373804.6A Active CN113127169B (en) 2021-04-07 2021-04-07 Efficient link scheduling method for dynamic workflow in data center network

Country Status (1)

Country Link
CN (1) CN113127169B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113518012B (en) * 2021-09-10 2021-12-10 之江实验室 Distributed cooperative flow simulation environment construction method and system
CN114691342B (en) * 2022-05-31 2022-09-20 蓝象智联(杭州)科技有限公司 Method and device for realizing priority scheduling of federated learning algorithm component and storage medium
CN114756358B (en) * 2022-06-15 2022-11-04 苏州浪潮智能科技有限公司 DAG task scheduling method, device, equipment and storage medium
CN116996443B (en) * 2023-09-25 2024-01-23 之江实验室 Network collaborative traffic scheduling method and system combining GNN and SAC models

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101267452A (en) * 2008-02-27 2008-09-17 华为技术有限公司 A conversion method and application server for WEB service mixing scheme
CN111756653A (en) * 2020-06-04 2020-10-09 北京理工大学 Multi-coflow scheduling method based on deep reinforcement learning of graph neural network

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105227488B (en) * 2015-08-25 2018-05-08 上海交通大学 A kind of network flow group scheduling method for distributed computer platforms
CN111131080B (en) * 2019-12-26 2021-09-07 电子科技大学 Distributed deep learning flow scheduling method, system and equipment

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101267452A (en) * 2008-02-27 2008-09-17 华为技术有限公司 A conversion method and application server for WEB service mixing scheme
CN111756653A (en) * 2020-06-04 2020-10-09 北京理工大学 Multi-coflow scheduling method based on deep reinforcement learning of graph neural network

Also Published As

Publication number Publication date
CN113127169A (en) 2021-07-16

Similar Documents

Publication Publication Date Title
CN113127169B (en) Efficient link scheduling method for dynamic workflow in data center network
CN109039942B (en) Network load balancing system and balancing method based on deep reinforcement learning
CN109753751B (en) MEC random task migration method based on machine learning
CN111756653B (en) Multi-coflow scheduling method based on deep reinforcement learning of graph neural network
Sun et al. Deepweave: Accelerating job completion time with deep reinforcement learning-based coflow scheduling
CN113098714B (en) Low-delay network slicing method based on reinforcement learning
CN111325356A (en) Neural network search distributed training system and training method based on evolutionary computation
CN114697229B (en) Construction method and application of distributed routing planning model
CN108111335A (en) A kind of method and system dispatched and link virtual network function
CN110990140B (en) Method for scheduling distributed machine learning flow in photoelectric switching network
CN111917642B (en) SDN intelligent routing data transmission method for distributed deep reinforcement learning
CN114707575A (en) SDN multi-controller deployment method based on AP clustering
CN114710439B (en) Network energy consumption and throughput joint optimization routing method based on deep reinforcement learning
CN116112488A (en) Fine-grained task unloading and resource allocation method for MEC network
Zhou et al. Multi-task deep learning based dynamic service function chains routing in SDN/NFV-enabled networks
CN117670005A (en) Super-computing internet multi-objective workflow optimization method and system based on ant colony algorithm
CN115150335B (en) Optimal flow segmentation method and system based on deep reinforcement learning
Wei et al. Drl-deploy: adaptive service function chains deployment with deep reinforcement learning
CN115756646A (en) Industrial internet-based edge computing task unloading optimization method
Wang et al. Sa-ddqn: Self-attention mechanism based ddqn for sfc deployment in nfv/mec-enabled networks
CN113891401A (en) Heterogeneous network slice scheduling method based on deep reinforcement learning
CN114915665B (en) Heterogeneous task scheduling method based on hierarchical strategy
CN117527590B (en) Method, system and medium for micro-service deployment and request routing based on edge network
CN116909717B (en) Task scheduling method
Guo et al. Multidimensional Resource and Load Collaborative Scheduling Algorithm Based on Reinforcement Learning for Cloud Data Centers

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant