CN114050961B - Large-scale network simulation system and resource dynamic scheduling and distributing method - Google Patents

Large-scale network simulation system and resource dynamic scheduling and distributing method Download PDF

Info

Publication number
CN114050961B
CN114050961B CN202111314346.5A CN202111314346A CN114050961B CN 114050961 B CN114050961 B CN 114050961B CN 202111314346 A CN202111314346 A CN 202111314346A CN 114050961 B CN114050961 B CN 114050961B
Authority
CN
China
Prior art keywords
network
simulation
node
virtual
plane
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111314346.5A
Other languages
Chinese (zh)
Other versions
CN114050961A (en
Inventor
王健
贾东睿
杨程竣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University
Original Assignee
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University filed Critical Nanjing University
Priority to CN202111314346.5A priority Critical patent/CN114050961B/en
Publication of CN114050961A publication Critical patent/CN114050961A/en
Application granted granted Critical
Publication of CN114050961B publication Critical patent/CN114050961B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/04Network management architectures or arrangements
    • H04L41/046Network management architectures or arrangements comprising network management agents or mobile agents therefor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/0813Configuration setting characterised by the conditions triggering a change of settings
    • H04L41/082Configuration setting characterised by the conditions triggering a change of settings the condition being updates or upgrades of network functionality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/145Network analysis or design involving simulating, designing, planning or modelling of a network

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

The invention discloses a large-scale network simulation system and a resource dynamic scheduling allocation method, wherein the simulation system comprises a simulation scene logic plane, a simulation control plane, a simulation data plane, a simulation storage plane and a simulation decision plane; the simulation scene logic plane generates an underlying network model according to an actual service scene, abstracts a service request of a user into a virtual network model and sends the virtual network model to a simulation decision plane; the simulation decision plane combines the virtual network request test set and the underlying network model to train a virtual node mapping model and train a virtual link mapping model; then, generating a decision for the virtual network request according to the latest underlying network resource state matrix of the simulation storage plane, and sending a decision scheme to the simulation control plane; the simulation control plane updates the resource state matrix of the underlying network according to the mapping scheme, and stores the latest resource state to the simulation storage plane, so that the problems of dynamic scheduling and management of large-scale network resources can be effectively solved.

Description

Large-scale network simulation system and resource dynamic scheduling and distributing method
Technical Field
The invention relates to the technical field of communication network simulation, in particular to a large-scale network simulation system and a resource dynamic scheduling and distributing method.
Background
The existing network simulation means mainly comprise three types of simulation software based on discrete events, such as NS-3, OPNET and the like, a physical simulation experiment bed and a network simulation platform based on virtualization. Network simulation software such as NS3 is more vivid, but the real-time performance of simulation cannot be guaranteed by an event-driven mode, and the provided model only supports the research on the mainstream protocol and cannot test and evaluate new technologies. The test experiment bed can realize vivid network simulation experiment, but the simulation experiment bed is not only high in cost and difficult to construct, but also poor in flexibility, and can not meet the reproducibility of the experiment. The network simulation platform based on virtualization can guarantee the authenticity of simulation, generate real and interactive flow, simultaneously can meet experimental scenes of different scales, and has the characteristics of flexibility and expandability.
On the other hand, the simulation platform based on the network virtualization technology realizes the sharing of bottom resources, and the simulation platform based on the virtualization is easy to improve the simulation scale of the network due to the distributed architecture and the expandable characteristic. However, with the expansion of network scale, the service requirements in the network become more complex and changeable, and in order to support heterogeneous virtual network requests as much as possible by the underlying network, the underlying network resources need to be reasonably allocated by means of a virtual network mapping algorithm, so as to realize efficient scheduling and management of the resources.
Although the existing large-scale network simulation platform can simulate a network scene more vividly, most simulation platforms do not dynamically feed back simulation results to a simulation network, and the dynamic adaptability of the simulation network is difficult to embody. The invention introduces the reinforcement learning technology into the virtualization simulation platform, dynamically updates the decision strategy of the reinforcement learning agent network and the resource state of the simulation network through the dynamic interaction of the reinforcement learning model and the simulation network, realizes the dynamic simulation of the large-scale network based on resource scheduling, and solves the problems of the dynamic scheduling and management of the large-scale network resources.
Disclosure of Invention
The purpose of the invention is as follows: aiming at the problems in the background technology, the invention provides a large-scale network simulation system and a resource dynamic scheduling allocation method, which are used for solving the problems that the simulation process of the existing simulation platform cannot form a closed loop, the historical simulation result cannot be dynamically acted into the simulation network by the simulation platform, the dynamic adaptability of the simulation network is difficult to embody, most of the simulation platforms lack a decision part, the decision process of the real network to the network change is not considered, and more complicated and changeable network scenes are difficult to simulate.
The technical scheme is as follows: in order to realize the purpose, the invention adopts the technical scheme that:
a large-scale network simulation system comprises a simulation scene logic plane, a simulation control plane, a simulation data plane, a simulation storage plane and a simulation decision plane;
the simulation scene logic plane generates a corresponding underlying network model according to an actual scene, and the underlying network model is used for simulating a real network in the actual scene; the underlying network model comprises network nodes, links associated with the network nodes, network topology relations, software systems used by the network nodes and protocol stacks; the logic plane of the simulation scene counts the service requests of the users at a certain moment, and all the service requests at a certain moment can be abstracted into a virtual network because the corresponding calculation resources and communication resources of the underlying network nodes are needed for responding to a certain service request. The node values and link values in the virtual network topology represent underlying network computing resources and communication resources required to respond to the corresponding service request, respectively. The process of mapping the virtual network to the underlying network is the process of scheduling and distributing underlying network resources; specifically, the following are shown:
G V →G S
the virtual network is denoted G V =(N V ,L V ),N V Representing virtual network nodes, L V Representing a virtual network link; the underlying network is denoted G S =(N S ,L S ),N S Indicating underlying network nodes, L S Representing an underlying network path;
the simulation control plane comprises an SDN controller and a node controller; the SDN controller is used for maintaining a topological structure of a simulation network in a simulation data plane and controlling topological linkage of the simulation network by dynamically updating flow table contents of an SDN switch; the node controller is used for issuing a service simulation instruction and monitoring the resource attribute of each node; the node controller dynamically switches the simulation service types of the simulation network nodes according to the decision result of the simulation decision plane;
the simulation data plane constructs the same number of simulation nodes according to the data of the underlying network model, and the resource attributes of the simulation nodes, the software system and the protocol stack used by the simulation nodes are consistent with the underlying network model and are used for instantiating the underlying network model generated in the simulation scene logic plane and simulating real network nodes; the simulation data plane simulates wireless links among the nodes according to the model data of the underlying network, and the uplink and downlink bandwidths, the error rate and the packet loss rate of the links among the nodes are consistent with those of the underlying network model; the simulation data plane supports communication between the virtual nodes and the physical nodes;
the simulation storage plane captures and stores node flow information, network situation information and resource occupation information generated by the simulation data plane in the simulation process, and transmits the node flow information, the network situation information and the resource occupation information to the simulation decision plane for decision making and integration links; the simulation storage plane changes an information collection task in real time through monitoring an information collection item, and meanwhile, the simulation storage plane comprises a visualization module used for inquiring and retrieving historical node information stored in the simulation storage plane;
and the simulation decision plane receives the virtual network request of the simulation scene logic plane, generates a decision scheme, sends a decision result to the simulation control plane, and updates the resource state of the simulation data plane.
A dynamic scheduling and distributing method for the underlying network resources by adopting the large-scale network simulation system comprises the following steps:
s1, a logic plane of a simulation scene counts user service requests at a certain moment, and all service requests at a certain moment can be abstracted into a virtual network because corresponding underlying network node computing resources and communication resources are needed for responding to a certain service request. CPU resources required to be consumed by each node in the virtual network and bandwidth values of adjacent links represent underlying network computing resources and communication resources required by a certain service; sending the virtual network request to a simulation decision plane;
s2, the simulation control plane issues a control command to the simulation data plane according to the underlying network topology data provided by the simulation scene logic plane, instantiates the underlying network, and simultaneously the simulation storage plane updates the underlying network node state matrix M at the current moment n And underlying network link state matrix M p
S3, a simulation decision plane needs to combine a virtual network request test set and an underlying network model to train a virtual node mapping model; sequentially defining environment state parameters of the underlying network nodes, an underlying network resource matrix and a reward function by adopting a reinforcement learning method, training a reinforcement learning agent network by using a gradient descent method, then sending a virtual network request test set and underlying network topology into a virtual node mapping model for training, and training the reinforcement learning agent network according to the reward function value of each iteration;
s4, training the virtual link mapping model again on the basis of finishing the training of the virtual node mapping model; sending the virtual network request test set, the underlying network and the trained virtual node mapping model into a virtual link mapping model, and training a virtual link agent decision network by using a gradient descent method;
s5, extracting an environment state matrix of the simulation storage plane by the simulation decision plane, and sequentially performing node mapping and link mapping on the virtual network request by the reinforcement learning agent network of the simulation decision plane according to the latest environment state matrix; when the virtual network request cannot complete mapping, the simulation decision plane rejects the request and informs the user;
s6, when the simulation decision plane successfully answers a virtual network request, the decision result is sent to the simulation control plane, the simulation control plane issues a control command accordingly, the simulation data plane updates the resource state of the underlying network, and finally the updated underlying network node state matrix M is used for updating the resource state of the underlying network n And underlying network link state matrix M p And storing to the simulation storage plane.
Further, the training of the virtual node mapping model in step S3 specifically includes:
s3.1, training an agent network by using a strategy gradient method, and adjusting a mapping strategy according to feedback of an environment, namely an incentive function, wherein the higher the environment incentive value is, the better the mapping scheme is; setting the current overall profit-to-cost ratio as a training reward function, and obtaining a group of complete virtual node mapping schemes in each iteration, wherein the aim is to maximize the long-term accumulated profit; the yield is the sum of the CPU values successfully mapped to the underlying network by the virtual nodes, and the cost is the underlying network resources consumed by mapping the virtual network request, namely:
Figure BDA0003343149360000041
wherein R is 1 Representing the above gain, the function CpuSum (n) v ) CPU value n for calculating success of a virtual node on a bottom node v Representing a respective virtual node; c 1 Representing the above cost, the function CpuCost (n) v ) Underlying network resources, n, for calculating consumption of a virtual node in a mapping process v Representing the corresponding virtual node.
The reward function is expressed as follows:
Figure BDA0003343149360000042
s3.2, selecting four parameters of computing resources, node link resources, node degree centrality and approximate centrality of each node to carry out node mapping strategy optimization of the proxy network; in particular, the amount of the solvent to be used,
(1) And computing resources: with C i To indicate the currently available CPU resources of the ith node on the underlying network:
Figure BDA0003343149360000043
wherein the function CPU () is used to calculate the remaining available CPU resources of the current node,
Figure BDA0003343149360000044
representing the corresponding underlying node.
(2) Node link resource: BW (l) is the bandwidth of link l of the ith node, and the bandwidths of all the adjacent links of node i are added together to use BW i Node link resources representing the ith node on the underlying network:
Figure BDA0003343149360000045
(3) Node degree centrality: by D i Reflecting the number of adjacent links owned by a node:
Figure BDA0003343149360000046
wherein
Figure BDA0003343149360000047
The degree of the ith node is, and m is the number of the nodes of the underlying network;
(4) And the centering property: this attribute reflects the proximity of one node to other nodes, denoted J i Represents:
Figure BDA0003343149360000048
wherein d is ij Representing the number of edges on the shortest path between the node i and the node j;
after extracting the environmental state parameters of the nodes, representing the characteristic vector as V for the ith node of the underlying network i =(C i ,BW i ,D i ,J i ) T The node state matrix of the underlying network is denoted as M n =(V 1 ,V 2 ,…,V N ) T
Further, the training of the virtual link mapping model in step S4 specifically includes:
s4.1, training the agent network by using a strategy gradient method, and adjusting a mapping strategy according to feedback of an environment, namely an incentive function, wherein the higher the environment incentive value is, the better the mapping scheme is; setting the current overall profit-to-cost ratio as a training reward function, and obtaining a group of complete virtual node mapping schemes in each iteration, wherein the aim is to maximize the long-term accumulated profit; the profit is the sum of the link bandwidth values successfully mapped to the underlying network by the virtual network; the cost is the underlying network resources consumed by mapping the virtual network request, i.e.:
Figure BDA0003343149360000051
wherein R is 2 Representing the above benefit, function BWSum (l) v ) For calculating the bandwidth value, l, of a successful link to the underlying network v Representing the corresponding virtual link; c 2 Representing the above cost, function BWCost (l) v ) Underlying network resources, hop (l), used to compute the consumption of a virtual link in the mapping process v ) Indicates the number of the bottom layer links occupied by a virtual link, l v Representing the corresponding virtual link.
The reward function is expressed as follows:
Figure BDA0003343149360000052
s4.2, when the virtual link is mapped to the bottom link, the link segmentation condition exists; the reachable path characteristics between node pairs in the underlying network are used as environmental state parameters, and two parameters of bandwidth and medium centrality are selected for each underlying path to be optimized; in particular, the amount of the solvent to be used,
(1) And bandwidth: for all links between certain underlying paths, selecting the smallest link bandwidth value as the bandwidth of the physical path, that is:
Figure BDA0003343149360000053
wherein,
Figure BDA0003343149360000054
indicating the kth underlay path of the underlay network.
(2) The medium centrality is as follows: for all links in the k-th underlying path, use d l Representing the number of shortest paths through link l, by D k Representing the number of all shortest paths passing through the k-th underlying path, the intermediary centrality of the k-th underlying path is represented as:
Figure BDA0003343149360000055
wherein count represents the number of shortest path edges on path k;
after extracting the environmental state parameters of the underlying network path, expressing the characteristic vector of the k-th underlying network path as
Figure BDA0003343149360000061
The link state matrix of the underlying network is denoted as M p =(P 1 ,P 2 ,…,P N ) T
Has the beneficial effects that:
the large-scale network simulation system and the resource dynamic scheduling and distributing method provided by the invention introduce the reinforcement learning technology into the virtualization simulation platform, dynamically update the decision strategy of the reinforcement learning agent network and the resource state of the simulation network through the dynamic interaction of the reinforcement learning model and the simulation network, and make an active decision on a service request. On one hand, the dynamic simulation of the large-scale network based on resource scheduling is realized, and the problems of dynamic scheduling and management of large-scale network resources are solved; on the other hand, the simulation system supports simulation of more complex and real network scenes, can actively make decisions on network changes, and overcomes the defect that the conventional simulation platform lacks active decisions.
Drawings
FIG. 1 is a schematic diagram of a large-scale network simulation platform system architecture provided by the present invention;
FIG. 2 is a functional diagram of a large-scale network simulation platform system provided by the present invention;
FIG. 3 is a schematic diagram of virtual network mapping to an underlying network in an embodiment of the invention;
FIG. 4 is a diagram of a virtual node mapping model of a simulation decision plane in an embodiment of the present invention;
FIG. 5 is a diagram of a virtual link mapping model of an emulation decision plane in an embodiment of the present invention;
Detailed Description
The invention will be further described with reference to the following drawings and specific embodiments.
The invention firstly provides a large-scale network simulation system based on resource scheduling, which has a specific architecture as shown in figure 1 and a specific functional schematic diagram as shown in figure 2 and comprises a simulation scene logic plane, a simulation control plane, a simulation data plane, a simulation storage plane and a simulation decision plane.
The simulation scene logic plane generates a corresponding underlying network model according to an actual scene, and the underlying network model is used for simulating a real network in the actual scene; the underlying network model comprises network nodes, links associated with the network nodes, a network topology relation, a software system used by the network nodes and a protocol stack; the simulation scene logic plane counts the service requests of users at a certain time, and all the service requests at a certain time can be abstracted into a virtual network because corresponding underlying network nodes are required to calculate resources and communication resources when responding to a certain service request. The node values and link values in the virtual network topology represent underlying network computing resources and communication resources required to respond to the corresponding service request, respectively. The process of mapping the virtual network to the underlying network, that is, the process of scheduling and allocating the underlying network resources, is illustrated in fig. 3; specifically, the following are shown:
G V →G S
the virtual network is denoted G V =(N V ,L V ),N V Representing virtual network nodes, L V Representing a virtual network link; the underlying network is denoted G S =(N S ,L S ),N S Indicating underlying network nodes, L S Representing an underlying network path;
the simulation control plane comprises an SDN controller and a node controller; the SDN controller is used for maintaining a topological structure of a simulation network in a simulation data plane and controlling topological linkage of the simulation network by dynamically updating flow table contents of an SDN switch; the node controller is used for issuing a service simulation instruction and monitoring the resource attribute of each node; the node controller dynamically switches the simulation service types of the simulation network nodes according to the decision result of the simulation decision plane;
the simulation data plane constructs the same number of simulation nodes according to the data of the underlying network model, and the resource attributes of the simulation nodes, the software system and the protocol stack used by the simulation nodes are consistent with the underlying network model and are used for instantiating the underlying network model generated in the simulation scene logic plane and simulating real network nodes; the simulation data plane simulates wireless links among the nodes according to the model data of the underlying network, and the uplink and downlink bandwidths, the error rate and the packet loss rate of the links among the nodes are consistent with those of the underlying network model; the simulation data plane supports communication between the virtual nodes and the physical nodes;
the simulation storage plane captures and stores node flow information, network situation information and resource occupation information generated by the simulation data plane in the simulation process, and transmits the node flow information, the network situation information and the resource occupation information to the simulation decision plane for decision making and integration links; the simulation storage plane changes an information collection task in real time through monitoring an information collection item, and meanwhile, the simulation storage plane comprises a visualization module which is used for inquiring and retrieving historical node information stored in the simulation storage plane;
and the simulation decision plane receives the virtual network request of the simulation scene logic plane, generates a decision scheme, sends a decision result to the simulation control plane, and updates the resource state of the simulation data plane.
Based on the large-scale network simulation system, the bottom layer resources are dispatched and distributed, and the specific method comprises the following steps:
step S1, a logic plane of a simulation scene counts user service requests at a certain moment, and all service requests at a certain moment can be abstracted into a virtual network because corresponding underlying network nodes need to calculate resources and communication resources when responding to a certain service request. CPU resources required to be consumed by each node in the virtual network and bandwidth values of adjacent links represent underlying network computing resources and communication resources required by a certain service; sending the virtual network request to a simulation decision plane;
s2, the simulation control plane issues a control command to the simulation data plane according to the underlying network topology data provided by the simulation scene logic plane, instantiates the underlying network, and simultaneously the simulation storage plane updates the underlying network node state matrix M at the current moment n And underlying network link state matrix M p
S3, a simulation decision plane needs to combine a virtual network request test set and an underlying network model to train a virtual node mapping model; the virtual node mapping model designed by the invention is shown in figure 4, and adopts a reinforcement learning method to define the environmental state parameters of the underlying network nodes, the underlying network resource matrix and the reward function in turn, trains the reinforcement learning agent network by using a gradient descent method, then sends the virtual network request test set and the underlying network topology into the virtual node mapping model for training, and trains the reinforcement learning agent network according to the reward function value of each iteration; in particular, the amount of the solvent to be used,
s3.1, training the agent network by using a strategy gradient method, and adjusting a mapping strategy according to feedback of an environment, namely a reward function, wherein the higher the environment reward value is, the better the mapping scheme is; setting the current overall profit-to-cost ratio as a training reward function, and obtaining a group of complete virtual node mapping schemes in each iteration, wherein the aim is to maximize the long-term accumulated profit; the profit is the sum of the CPU values successfully mapped to the underlying network by the virtual node, and the cost is the underlying network resources consumed by the request for mapping the virtual node, namely:
Figure BDA0003343149360000081
wherein R is 1 Representing the above gain, the function CpuSum (n) v ) For calculating the CPU value, n, of a virtual node successfully arriving at the underlying node v Representing a respective virtual node; c 1 Representing the above cost, the function CpuCost (n) v ) Underlying network resources, n, for calculating consumption of a virtual node in a mapping process v Representing the corresponding virtual node.
The reward function is expressed as follows:
Figure BDA0003343149360000082
s3.2, selecting four parameters of computing resources, node link resources, node degree centrality and approximate centrality of each node to carry out node mapping strategy optimization of the proxy network; in particular, the amount of the solvent to be used,
(1) And computing resources: with C i To indicate the current available CPU resources of the ith node on the underlying network:
Figure BDA0003343149360000083
wherein, the function CPU () is used to calculate the remaining available CPU resources of the current node,
Figure BDA0003343149360000084
representing the corresponding underlying node.
(2) Node link resource: BW (l) is the bandwidth of link l of the ith node, and the bandwidths of all the adjacent links of node i are added together to use BW i Node link resources representing the ith node on the underlying network:
Figure BDA0003343149360000085
(3) Node degree centrality: by D i Reflecting the number of adjacent links owned by a node:
Figure BDA0003343149360000091
wherein
Figure BDA0003343149360000092
The degree of the ith node is, and m is the number of the nodes of the underlying network;
(4)、the approach to centrality: this attribute reflects the proximity of one node to other nodes, denoted J i Represents:
Figure BDA0003343149360000093
wherein d is ij Representing the number of edges on the shortest path between the node i and the node j;
after extracting the environmental state parameters of the nodes, representing the characteristic vector as V for the ith node of the underlying network i =(C i ,BW i ,D i ,J i ) T The node state matrix of the underlying network is denoted as M n =(V 1 ,V 2 ,…,V N ) T
S4, training the virtual link mapping model again on the basis of the completion of the training of the virtual node mapping model, as shown in FIG. 5; sending the virtual network request test set, the underlying network and the trained virtual node mapping model into a virtual link mapping model, and training a virtual link agent decision network by using a gradient descent method; in particular, the amount of the solvent to be used,
s4.1, training the agent network by using a strategy gradient method, and adjusting a mapping strategy according to feedback of an environment, namely a reward function, wherein the higher the environment reward value is, the better the mapping scheme is; setting the current overall profit-to-cost ratio as a training reward function, and obtaining a group of complete virtual node mapping schemes in each iteration, wherein the aim is to maximize the long-term accumulated profit; the profit is the sum of the link bandwidth values successfully mapped to the underlying network by the virtual network; the cost is the underlying network resources consumed by mapping the virtual network request, i.e.:
Figure BDA0003343149360000094
wherein R is 2 Representing the above benefit, function BWSum (l) v ) For calculating the bandwidth value, l, of a successful link to the underlying network v Representing respective virtual links;C 2 Representing the above cost, function BWCost (l) v ) Underlying network resources, hop (l), used to compute the consumption of a virtual link in the mapping process v ) Indicates the number of the bottom layer links occupied by a virtual link, l v Representing the corresponding virtual link.
The reward function is expressed as follows:
Figure BDA0003343149360000095
s4.2, when the virtual link is mapped to the bottom link, the link segmentation condition exists; the reachable path characteristics between each node pair in the underlying network are used as environment state parameters, and two parameters of bandwidth and medium centrality are selected for each underlying path to be optimized; in particular, the amount of the solvent to be used,
(1) Bandwidth: for all links between certain underlying paths, selecting the smallest link bandwidth value as the bandwidth of the physical path, that is:
Figure BDA0003343149360000101
wherein,
Figure BDA0003343149360000102
representing the kth underlay path of the underlay network.
(2) The center of the medium is as follows: for all links in the k-th underlying path, use d l Representing the number of shortest paths through link l, by D k Representing the number of all shortest paths through the kth underlying path, the intermediaries centrality of the kth underlying path is represented as:
Figure BDA0003343149360000103
wherein count represents the number of shortest path edges on path k;
after extracting the environmental state parameters of the underlying network path, the k-th itemUnderlying network paths, the feature vectors being represented as
Figure BDA0003343149360000104
The link state matrix of the underlying network is denoted as M p =(P 1 ,P 2 ,…,P N ) T
S5, extracting an environment state matrix of the simulation storage plane by the simulation decision plane, and sequentially performing node mapping and link mapping on the virtual network request by the reinforcement learning agent network of the simulation decision plane according to the latest environment state matrix; when the virtual network request cannot complete mapping, the simulation decision plane rejects the request and informs the user;
s6, when the simulation decision plane successfully answers a virtual network request, the decision result is sent to the simulation control plane, the simulation control plane issues a control command accordingly, the simulation data plane updates the resource state of the underlying network, and finally the updated underlying network node state matrix M is used for updating the resource state of the underlying network n And underlying network link state matrix M p And storing to the simulation storage plane.
For subsequently arriving virtual network requests, the mapping will be done in the steps described above.
The above description is only of the preferred embodiments of the present invention, and it should be noted that: it will be apparent to those skilled in the art that various modifications and adaptations can be made without departing from the principles of the invention and these are intended to be within the scope of the invention.

Claims (2)

1. A large-scale network simulation system is characterized by comprising a simulation scene logic plane, a simulation control plane, a simulation data plane, a simulation storage plane and a simulation decision plane;
the simulation scene logic plane generates a corresponding underlying network model according to an actual scene, and the underlying network model is used for simulating a real network in the actual scene; the underlying network model comprises network nodes, links associated with the network nodes, network topology relations, software systems used by the network nodes and protocol stacks; the logic plane of the simulation scene counts the service requests of the users at a certain moment, and all the service requests at a certain moment can be abstracted into a virtual network because the corresponding calculation resources and communication resources of the underlying network nodes are needed for responding to a certain service request; the node values and the link values in the virtual network topology respectively represent underlying network computing resources and communication resources required by responding to corresponding service requests; the process of mapping the virtual network to the underlying network is the process of scheduling and distributing underlying network resources; specifically, the following are shown:
G V →G S
the virtual network is denoted G V =(N V ,L V ),N V Representing virtual network nodes, L V Representing a virtual network link; the underlying network is denoted G S =(N S ,L S ),N S Indicating underlying network nodes, L S Representing an underlying network path;
the simulation control plane comprises an SDN controller and a node controller; the SDN controller is used for maintaining a topological structure of a simulation network in a simulation data plane and controlling topological linkage of the simulation network by dynamically updating flow table contents of an SDN switch; the node controller is used for issuing a service simulation instruction and monitoring the resource attribute of each node; the node controller dynamically switches the simulation service types of the simulation network nodes according to the decision result of the simulation decision plane;
the simulation data plane constructs the same number of simulation nodes according to the network model data, and the resource attributes of the simulation nodes, the software system and the protocol stack used by the simulation nodes are consistent with the network model and are used for instantiating the network model generated in the simulation scene logic plane and simulating the real network nodes; the simulation data plane simulates wireless links among nodes according to network model data, and the uplink and downlink bandwidths, the error rate and the packet loss rate of the links among the nodes are consistent with those of a network model; the simulation data plane supports communication between the virtual nodes and the physical nodes;
the simulation storage plane captures and stores node flow information, network situation information and resource occupation information generated by the simulation data plane in the simulation process, and transmits the node flow information, the network situation information and the resource occupation information to the simulation decision plane for decision making and integration links; the simulation storage plane changes an information collection task in real time through monitoring an information collection item, and meanwhile, the simulation storage plane comprises a visualization module which is used for inquiring and retrieving historical node information stored in the simulation storage plane;
and the simulation decision plane receives the virtual network request of the simulation scene logic plane, generates a decision scheme, sends a decision result to the simulation control plane, and updates the resource state of the simulation data plane.
2. A method for dynamically scheduling and allocating underlying network resources by using the large-scale network simulation system according to claim 1, comprising the steps of:
s1, a logic plane of a simulation scene counts user service requests at a certain moment, and all service requests at a certain moment can be abstracted into a virtual network because corresponding underlying network node computing resources and communication resources are needed for responding to a certain service request; CPU resources required to be consumed by each node in the virtual network and bandwidth values of adjacent links represent underlying network computing resources and communication resources required by a certain service; sending the virtual network request to a simulation decision plane;
s2, the simulation control plane issues a control command to the simulation data plane according to the underlying network topology data provided by the simulation scene logic plane, instantiates the underlying network, and simultaneously the simulation storage plane updates the underlying network node state matrix M at the current moment n And underlying network link state matrix M p
S3, the simulation decision plane needs to combine a virtual network request test set and an underlying network model to train a virtual node mapping model; sequentially defining environment state parameters of the underlying network nodes, an underlying network resource matrix and a reward function by adopting a reinforcement learning method, training a reinforcement learning agent network by using a gradient descent method, then sending a virtual network request test set and underlying network topology into a virtual node mapping model for training, and training the reinforcement learning agent network according to the reward function value of each iteration; the training of the virtual node mapping model specifically comprises:
s3.1, training the agent network by using a strategy gradient method, and adjusting a mapping strategy according to feedback of an environment, namely a reward function, wherein the higher the environment reward value is, the better the mapping scheme is; setting the current overall profit-to-cost ratio as a training reward function, and obtaining a group of complete virtual node mapping schemes in each iteration, wherein the aim is to maximize the long-term accumulated profit; the yield is the sum of the CPU values successfully mapped to the underlying network by the virtual nodes, and the cost is the underlying network resources consumed by mapping the virtual network request, namely:
Figure FDA0003725293830000021
wherein R is 1 Representing the above gain, the function CpuSum (n) v ) For calculating the CPU value, n, of a virtual node successfully mapped onto an underlying node v Representing the respective virtual node; cost 1 Representing the above cost, the function CpuCost (n) v ) Underlying network resources, n, consumed by a virtual node in a computational mapping process v Representing a respective virtual node;
the reward function is expressed as follows:
Figure FDA0003725293830000022
s3.2, selecting four parameters of computing resources, node link resources, node degree centrality and approximate centrality of each node to carry out node mapping strategy optimization of the proxy network; in particular, the amount of the solvent to be used,
(1) And computing resources: with C i To indicate the current available CPU resources of the ith node on the underlying network:
Figure FDA0003725293830000031
wherein the function CPU () is used to calculate the remaining available CPU resources of the current node,
Figure FDA0003725293830000032
representing the corresponding underlying node;
(2) Node link resources: BW (l) is the bandwidth of link l of the ith node, and the bandwidths of all adjacent links of node i are added together using BW i Node link resources representing the ith node on the underlying network:
Figure FDA0003725293830000033
(3) Node degree centrality: by D i Reflecting the number of adjacent links owned by a node:
Figure FDA0003725293830000034
wherein
Figure FDA0003725293830000035
The degree of the ith node is, and m is the number of the nodes of the underlying network;
(4) And the centering property: this attribute reflects the proximity of one node to other nodes, denoted J i Represents:
Figure FDA0003725293830000036
wherein d is ij Representing the number of edges on the shortest path between the node i and the node j;
after extracting the environmental state parameters of the nodes, representing the characteristic vector as V for the ith node of the underlying network i =(C i ,BW i ,D i ,J i ) T The node state matrix of the underlying network is denoted as M n =(V 1 ,V 2 ,…,V N ) T
S4, training the virtual link mapping model again on the basis of finishing the training of the virtual node mapping model; sending the virtual network request test set, the underlying network and the trained virtual node mapping model into a virtual link mapping model, and training a virtual link agent decision network by using a gradient descent method; the training of the virtual link mapping model specifically includes:
s4.1, training the agent network by using a strategy gradient method, and adjusting a mapping strategy according to feedback of an environment, namely an incentive function, wherein the higher the environment incentive value is, the better the mapping scheme is; setting the current overall profit-to-cost ratio as a training reward function, and obtaining a group of complete virtual node mapping schemes in each iteration, wherein the aim is to maximize the long-term accumulated profit; the profit is the sum of the link bandwidth values successfully mapped to the underlying network by the virtual network; the cost is the underlying network resources consumed by mapping the virtual network request, i.e.:
Figure FDA0003725293830000037
wherein R is 2 Representing the above benefit, function BWSum (l) v ) Used for calculating the link bandwidth value l of a virtual link successfully mapped to the underlying network v Representing the corresponding virtual link; cost 2 Representing the above cost, function BWCost (l) v ) Underlying network resources, hop (l), used to compute the consumption of a virtual link in the mapping process v ) Indicates the number of the bottom layer links occupied by a virtual link, l v Representing the corresponding virtual link;
the reward function is expressed as follows:
Figure FDA0003725293830000041
s4.2, when the virtual link is mapped to the bottom link, the link segmentation condition exists; the reachable path characteristics between each node pair in the underlying network are used as environment state parameters, and two parameters of bandwidth and medium centrality are selected for each underlying path to be optimized; in particular, the amount of the solvent to be used,
(1) Bandwidth: for all links between certain underlying paths, selecting the smallest link bandwidth value as the bandwidth of the physical path, that is:
Figure FDA0003725293830000042
wherein,
Figure FDA0003725293830000043
a kth underlay path representing an underlay network;
(2) The center of the medium is as follows: for all links in the k-th underlying path, use d l Representing the number of shortest paths through link l, by D k Representing the number of all shortest paths through the kth underlying path, the intermediaries centrality of the kth underlying path is represented as:
Figure FDA0003725293830000044
wherein count represents the number of shortest path edges on path k;
after extracting the environmental state parameters of the underlying network path, expressing the characteristic vector of the k-th underlying network path as
Figure FDA0003725293830000045
The link state matrix of the underlying network is denoted as M p =(P 1 ,P 2 ,…,P N ) T
S5, extracting an environment state matrix of the simulation storage plane by the simulation decision plane, and sequentially performing node mapping and link mapping on the virtual network request by the reinforcement learning agent network of the simulation decision plane according to the latest environment state matrix; when the virtual network request cannot complete mapping, the simulation decision plane rejects the request and informs the user;
s6, when the simulation decision plane successfully answers a virtual network request, the decision result is sent to the simulation control plane, the simulation control plane issues a control command according to the decision result, the simulation data plane updates the resource state of the underlying network, and finally the updated underlying network node state matrix M is used for updating the resource state of the underlying network n And underlying network link state matrix M p And storing to the simulation storage plane.
CN202111314346.5A 2021-11-08 2021-11-08 Large-scale network simulation system and resource dynamic scheduling and distributing method Active CN114050961B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111314346.5A CN114050961B (en) 2021-11-08 2021-11-08 Large-scale network simulation system and resource dynamic scheduling and distributing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111314346.5A CN114050961B (en) 2021-11-08 2021-11-08 Large-scale network simulation system and resource dynamic scheduling and distributing method

Publications (2)

Publication Number Publication Date
CN114050961A CN114050961A (en) 2022-02-15
CN114050961B true CN114050961B (en) 2022-10-18

Family

ID=80207489

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111314346.5A Active CN114050961B (en) 2021-11-08 2021-11-08 Large-scale network simulation system and resource dynamic scheduling and distributing method

Country Status (1)

Country Link
CN (1) CN114050961B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114615183B (en) * 2022-03-14 2023-09-05 广东技术师范大学 Routing method, device, computer equipment and storage medium based on resource prediction
CN117693059A (en) * 2022-08-30 2024-03-12 中兴通讯股份有限公司 Resource scheduling method, device and system, storage medium and electronic system
CN115729714A (en) * 2023-01-06 2023-03-03 之江实验室 Resource allocation method, device, storage medium and electronic equipment
CN117293807B (en) * 2023-09-25 2024-07-26 上海能优网电力科技有限公司 Multi-time scale optimization method and system for information side model of power distribution network
CN117714306B (en) * 2024-02-06 2024-04-16 湖南天冠电子信息技术有限公司 Ethernet performance optimization method based on scheduling algorithm

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102223281A (en) * 2011-06-24 2011-10-19 清华大学 Method for remapping resource demand dynamic change based on virtual network
CN109889930A (en) * 2019-03-26 2019-06-14 重庆邮电大学 A kind of virtual optical network mapping method and device for combining energy consumption and load balancing
CN110365514A (en) * 2019-05-24 2019-10-22 北京邮电大学 SDN multistage mapping method of virtual network and device based on intensified learning
CN113193999A (en) * 2021-04-29 2021-07-30 东北大学 Virtual network mapping method based on depth certainty strategy gradient
CN113328938A (en) * 2021-05-25 2021-08-31 电子科技大学 Network autonomous intelligent management and control method based on deep reinforcement learning

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11153229B2 (en) * 2018-01-19 2021-10-19 Ciena Corporation Autonomic resource partitions for adaptive networks
CN112651130A (en) * 2020-12-28 2021-04-13 航天科工网络信息发展有限公司 Decision support-oriented virtual-real mapping parallel simulation system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102223281A (en) * 2011-06-24 2011-10-19 清华大学 Method for remapping resource demand dynamic change based on virtual network
CN109889930A (en) * 2019-03-26 2019-06-14 重庆邮电大学 A kind of virtual optical network mapping method and device for combining energy consumption and load balancing
CN110365514A (en) * 2019-05-24 2019-10-22 北京邮电大学 SDN multistage mapping method of virtual network and device based on intensified learning
CN113193999A (en) * 2021-04-29 2021-07-30 东北大学 Virtual network mapping method based on depth certainty strategy gradient
CN113328938A (en) * 2021-05-25 2021-08-31 电子科技大学 Network autonomous intelligent management and control method based on deep reinforcement learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
A Dynamic and Collaborative Multi-Layer Virtual Network Embedding Algorithm in SDN Based on Reinforcement Learning;M. Lu, Y. Gu and D. Xie;《IEEE Transactions on Network and Service Management》;20200729;全文 *
基于数据驱动的虚拟网络映射算法;陈旭;《中国优秀硕士学位论文全文数据库(电子期刊)》;20181115(第11期);全文 *

Also Published As

Publication number Publication date
CN114050961A (en) 2022-02-15

Similar Documents

Publication Publication Date Title
CN114050961B (en) Large-scale network simulation system and resource dynamic scheduling and distributing method
CN109714219A (en) A kind of virtual network function fast mapping algorithm based on satellite network
CN113364850B (en) Software-defined cloud-edge collaborative network energy consumption optimization method and system
CN113794494B (en) Edge computing system and computing unloading optimization method for low-orbit satellite network
CN105515987B (en) A kind of mapping method based on SDN framework Virtual optical-fiber networks
CN108566659A (en) A kind of online mapping method of 5G networks slice based on reliability
CN108874525A (en) A kind of service request distribution method towards edge calculations environment
CN108111335B (en) A kind of method and system of scheduling and link virtual network function
CN113660681B (en) Multi-agent resource optimization method applied to unmanned aerial vehicle cluster auxiliary transmission
Lei et al. Congestion control in SDN-based networks via multi-task deep reinforcement learning
Guan et al. An intelligent wireless channel allocation in HAPS 5G communication system based on reinforcement learning
CN115580882A (en) Dynamic network slice resource allocation method and device, storage medium and electronic equipment
CN112199154B (en) Reinforced learning training system and method based on distributed collaborative sampling center type optimization
Bouzidi et al. Dynamic clustering of software defined network switches and controller placement using deep reinforcement learning
CN116489708B (en) Meta universe oriented cloud edge end collaborative mobile edge computing task unloading method
CN110247795A (en) A kind of cloud net resource service chain method of combination and system based on intention
CN116893861A (en) Multi-agent cooperative dependency task unloading method based on space-ground cooperative edge calculation
CN111343095B (en) Method for realizing controller load balance in software defined network
Na et al. An evolutionary game approach on IoT service selection for balancing device energy consumption
Hu et al. Dynamic task offloading in MEC-enabled IoT networks: A hybrid DDPG-D3QN approach
CN112073237A (en) Large-scale target network construction method in cloud edge architecture
EP4024212A1 (en) Method for scheduling interference workloads on edge network resources
CN116755867B (en) Satellite cloud-oriented computing resource scheduling system, method and storage medium
Zhu et al. Deep reinforcement learning-based edge computing offloading algorithm for software-defined IoT
CN115225512B (en) Multi-domain service chain active reconfiguration mechanism based on node load prediction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant