CN113641462B - Virtual network hierarchical distributed deployment method and system based on reinforcement learning - Google Patents

Virtual network hierarchical distributed deployment method and system based on reinforcement learning Download PDF

Info

Publication number
CN113641462B
CN113641462B CN202111195085.XA CN202111195085A CN113641462B CN 113641462 B CN113641462 B CN 113641462B CN 202111195085 A CN202111195085 A CN 202111195085A CN 113641462 B CN113641462 B CN 113641462B
Authority
CN
China
Prior art keywords
virtual network
physical host
module
action
deployment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111195085.XA
Other languages
Chinese (zh)
Other versions
CN113641462A (en
Inventor
陈曦
吴涛
邓伟健
黄�俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southwest Minzu University
Original Assignee
Southwest Minzu University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southwest Minzu University filed Critical Southwest Minzu University
Priority to CN202111195085.XA priority Critical patent/CN113641462B/en
Publication of CN113641462A publication Critical patent/CN113641462A/en
Application granted granted Critical
Publication of CN113641462B publication Critical patent/CN113641462B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5077Logical partitioning of resources; Management or configuration of virtualized resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45595Network integration; Enabling network access in virtual machine instances

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

The invention discloses a virtual network hierarchical distributed deployment method and system based on reinforcement learning, which are oriented to network simulation based on Docker containerized virtual network, aim at optimizing block cutting and deployment problems of the virtual network, and design a reward mechanism giving consideration to both physical host machine resource consumption and virtual network block cutting cross-host machine communication performance loss through a reinforcement learning framework under the distributed environment of physical host machines with limited calculation, network and storage resources
Figure DEST_PATH_IMAGE002
According to
Figure 796612DEST_PATH_IMAGE002
Continuously calculating the supply state of each physical host machine in different resources
Figure DEST_PATH_IMAGE004
Take different actions
Figure DEST_PATH_IMAGE006
Long term benefits of
Figure DEST_PATH_IMAGE008
Make the algorithm according to
Figure 838386DEST_PATH_IMAGE008
Dynamically and autonomously continuously learning the optimized virtual network block size and the appropriate deployment time, and introducing certain dynamic randomness through a randomness strategy to avoid excessively rigid selection
Figure 211599DEST_PATH_IMAGE008
The maximum action causes the problem of resource over consumption or block over fragmentation, thereby achieving the purpose of layering and distributed optimal deployment of the virtual network.

Description

Virtual network hierarchical distributed deployment method and system based on reinforcement learning
Technical Field
The invention relates to the technical field of network virtualization, in particular to a virtual network hierarchical distributed deployment method and system based on reinforcement learning.
Background
Network simulation is a key support for computer network architecture, protocol and algorithm research. Because the Docker container reserves a basically complete TCP/IP protocol stack, and has higher start efficiency and less performance overhead compared with a virtual machine, the containerization technology is gradually popular, and a new idea is provided for network simulation, namely, the Docker container is used as a core to construct virtual network elements (such as virtual routers, virtual switches, virtual end systems and the like), and technologies such as veth-pair, OVS (Open vSwitch), vxlan (virtual eXtensible Local Area network) and the like are used in cooperation to generate virtual links for connection, so that a virtual network is formed to be deployed on a physical host machine for simulation. The Docker container runs a TCP/IP protocol stack of a Linux kernel, is efficient and low in consumption, and has an open programming interface, so that network simulation based on the Docker container has the characteristics of high fidelity and easiness in programming. In a physical host with limited computing, network and storage resources, to implement large-scale deployment of containerized virtual networks based on technologies such as Docker and OVS, reasonable, automatic and efficient mapping and necessary block deployment need to be performed between the virtual networks and the computing clusters of the physical host, so that the resource demand of the virtual networks and the resource supply of the physical host are relatively balanced in a distributed scene, and the performance of network simulation is improved. Therefore, virtual network distributed optimization deployment is the key to network simulation based on the Docker technology.
The academic community has related research on related problems, namely, Virtual Network mapping (VNE), and the solution of the problem has relatively high complexity and even NP difficulty. Early learners solve the problem by a pure heuristic method, but local optimal solutions can be obtained by the pure heuristic method, and the problem of the local optimal solutions can be effectively solved by using meta-heuristic solving, for example, FAJJARI et al propose an expandable mapping strategy based on an ant colony meta-heuristic algorithm; also, for example, ARA Ú JO et al propose a hybrid algorithm incorporating meta-heuristics, and an online policy that takes into account the execution speed of the virtual network map to ensure minimal latency, providing a fast solution in a multi-domain environment. However, the mapping algorithm in the existing literature is mainly designed for application scenarios based on virtual machines, and mainly from the aspects of virtual machine resource allocation efficiency and mapping success rate. The virtual network is constructed by the Docker technology, and besides the consideration of the high efficiency of resource allocation and the success rate of mapping, the mapping algorithm needs to be designed and optimized by combining the technical characteristics of the Docker: (1) virtual network elements simulated by the Docker container are represented as low-overhead processes on a host machine, the granularity is finer, the time variation is more obvious, and therefore the mapping algorithm is required to have better dynamic property and adaptability and needs to be sensitive and agile to resource consumption and change; (2) docker is used as a lightweight virtualization technology, on one hand, the Docker is beneficial to constructing a virtual network with a larger scale, on the other hand, the Docker is expected to be deployed on a plurality of low-profile X86 host machines, and both Docker and Docker require full consideration of the resource limitation characteristics of the host machines and flexibly perform automatic optimization, block cutting and deployment on the virtual network; (3) after the virtual network is cut into blocks and is respectively deployed on a plurality of host machines, cross-host machine communication needs to be achieved through modes such as OVS + VxLAN and the like so as to transparently present a uniform virtual network to a user, therefore, the mapping algorithm needs to be cooperatively optimized by combining technical characteristics of virtual switches such as OVS + VxLAN tunnels and the like, and under the condition of considering host machine resource consumption, the number of virtual network cuts is reduced as much as possible so as to control performance loss caused by cross-host machine communication among the cuts.
Specifically, fig. 1 is a macro process for performing blocking, mapping, and deployment on a virtual network, which includes the following steps:
1. topology description of virtual networks: assuming that a user wants to deploy a virtual network as shown in the top of fig. 1 (the size of the virtual network may be large, and the embodiment of the present invention is illustrated for convenience, and only a topology of 2 end systems and 2 routers is drawn), the virtual network may be described in JSON file format.
2. And (3) slicing and mapping of the virtual network: (1) inputting an algorithm: the JSON file is used as the input of the algorithm, after the topological structure of the virtual network is read by the algorithm, whether the virtual network is cut into blocks or mapped is determined according to the conditions of the residual calculation, network and storage resources of the existing physical hosts (if a single host can accommodate the whole virtual network, the cutting and mapping are not needed). (2) And (3) outputting an algorithm: if dicing and mapping are required, several diced JSON files are generated, as shown in the "dice a" and "dice B" portions of fig. 1.
3. Deployment of the virtual network: and each physical host machine receives the block JSON file, and generates various virtual network elements by using techniques such as Docker and OVS according to the JSON description. This involves network virtualization, and mainly includes two aspects: node virtualization and link virtualization. Node virtualization: simulating equipment such as an end system, a router and the like by using a Docker container; the OVS technique was used to simulate a two-layer switching device. Link virtualization: and connecting various virtual network elements obtained by node virtualization by using a veth-pair technology.
4. Reconnection of virtual network: after the virtual network is subjected to blocking and mapping, different blocks are respectively deployed on different host machines, and the original topological structure of the virtual network is damaged on some links. Therefore, it is necessary to reconnect the original topology across hosts, mainly implemented by using OVS + VxLAN through tunneling technique, as shown by OVS + VxLAN between "cut-block a" and "cut-block B" in fig. 1. Considering that the OVS + VxLAN tunnel causes certain network performance loss, in order to enable the virtual network to have higher fidelity, the algorithm design enables the virtual network deployed at the same time to be compact as much as possible, and the number of physical host links crossing the bottom layer is reduced, namely the original virtual network is not easy to be cut into pieces.
In fact, summarizing, the key to the problem lies in how to optimally block and deploy the virtual network, which faces the following technical difficulties: (a) the block is too large, which easily consumes the resources of the deployed physical hosts, and the residual resources of some physical hosts are supplied too little after long-term operation, so that the virtual network or the block cannot be deployed effectively. The supply among the physical host machines is unbalanced, and when a new virtual network needs to be deployed, the balanced deployment is difficult. (b) The block cutting is too small, so that the virtual network is easily cut to be more than fragmented, the number of blocks is too many, and when the virtual network is distributed and deployed on a plurality of physical host machines (especially under the condition that the physical host machines need to be communicated with each other through multiple hops), the OVS + VxLAN tunnel-based cross-host machine communication causes too much performance loss, and the simulation effect and the fidelity are influenced. The requirement of the blocking and deployment method needs to be able to dynamically and autonomously learn, adapt to the resource consumption requirement of the virtual network and the resource supply situation of the physical host, form a virtual network block with a suitable scale, and perform hierarchical and distributed optimized deployment based on the network block. However, the existing work mostly abstracts the distributed deployment of the virtual network into a mathematical programming problem (generally NP is difficult), and a heuristic method is utilized to balance the solving efficiency and the optimization degree. However, the heuristic method has no advantages in the aspects of dynamic and timeliness characteristics, and the self-adaptive capability and the learning and evolution capability in the face of a complex network environment are also weak.
Disclosure of Invention
The invention aims at the problems existing in the prior artThe utility model provides a virtual network hierarchical distributed deployment method and system based on reinforcement learning. The method is oriented to network simulation based on Docker containerized virtual network, aims at the problem of optimizing block cutting and deployment of the virtual network, and designs a reward mechanism which gives consideration to resource consumption of physical host machines and communication performance loss of virtual network block cutting and host machine crossing through a reinforcement learning framework under the distributed environment of the physical host machines with limited computing, network and storage resources
Figure 499783DEST_PATH_IMAGE001
According to
Figure 712196DEST_PATH_IMAGE001
Continuously calculating different actions taken by each physical host machine under different resource supply states s
Figure 226354DEST_PATH_IMAGE002
Long term benefits of
Figure 418301DEST_PATH_IMAGE003
Make the algorithm according to
Figure 91728DEST_PATH_IMAGE003
Dynamically and autonomously continuously learning the optimized virtual network block size and the appropriate deployment time, and introducing certain dynamic randomness through a randomness strategy to avoid excessively rigid selection
Figure 383032DEST_PATH_IMAGE003
The maximum action causes the problem of resource over consumption or block over fragmentation, thereby achieving the purpose of layering and distributed optimal deployment of the virtual network.
The specific technical scheme of the invention is as follows:
a virtual network hierarchical distributed deployment method based on reinforcement learning comprises the following steps:
step 1: according to each physical host machine
Figure 446803DEST_PATH_IMAGE004
Establishing an action value function to form an action value function table;
wherein,
Figure 6222DEST_PATH_IMAGE004
represents a physical host machine, the superscript p represents physical, the subscript r represents the number of the physical host machine, and the value range is
Figure 471839DEST_PATH_IMAGE005
And R is the total number of the physical host machines.
Step 2: waiting for a new virtual network deployment request, and jumping to the step 3 when the new virtual network deployment request arrives;
and step 3: based on observations of resource supply of physical hosts
Figure 199623DEST_PATH_IMAGE006
Finding the physical host with the largest resource supply
Figure 547428DEST_PATH_IMAGE007
And 4, step 4: judging the physical host machine
Figure 81177DEST_PATH_IMAGE007
Whether or not the virtual network can be accommodated,
if so, jumping to step 5,
if the data can not be accommodated, jumping to step 6;
and 5: direct deployment, setting current actions as
Figure 338983DEST_PATH_IMAGE008
Deploying and skipping to step 8;
step 6: deployment of blocks according to action cost function
Figure 81679DEST_PATH_IMAGE009
Selecting an action if the action is
Figure 182359DEST_PATH_IMAGE010
Deploying, skipping to step 8, if the action is
Figure 254220DEST_PATH_IMAGE011
Expanding and jumping to step 7;
and 7: virtual network element with maximum out-degree in undeployed part of virtual network
Figure 366533DEST_PATH_IMAGE012
As the center, the expansion of the blocks is carried out, and the virtual network element set in the virtual network blocks is gradually constructed
Figure 29595DEST_PATH_IMAGE013
Skipping to step 8;
wherein,
Figure 56719DEST_PATH_IMAGE014
representing a virtual network element, superscript
Figure 994588DEST_PATH_IMAGE015
Represents a local, subscript
Figure 289303DEST_PATH_IMAGE016
The number representing the virtual network element has a value range of
Figure 795371DEST_PATH_IMAGE017
IIs the total number of the virtual network elements.
Figure 339485DEST_PATH_IMAGE018
The virtual network block is represented by a virtual network block, the superscript b represents a block, the subscript m represents the serial number of the virtual network block, the total block number of the virtual network block is undetermined, and the virtual network block is dynamically determined by an algorithm according to the resource supply of a physical host machine and other multi-aspect conditions.
And 8: calculating the prize according to the formula
Figure 517263DEST_PATH_IMAGE019
Figure 604168DEST_PATH_IMAGE020
In the formula,
Figure 343454DEST_PATH_IMAGE021
at time t, the physical host
Figure 640443DEST_PATH_IMAGE022
The number of deployed virtual network tiles,
Figure 529901DEST_PATH_IMAGE023
for cutting blocks from virtual networks
Figure 799209DEST_PATH_IMAGE018
The sum of the multi-dimensional resources consumed,
Figure 273178DEST_PATH_IMAGE024
is the largest physical host
Figure 995146DEST_PATH_IMAGE022
Observations of resource provisioning;
and step 9: according to the reward
Figure 953875DEST_PATH_IMAGE019
Updating action cost function in current action cost function table
Figure 812109DEST_PATH_IMAGE025
Step 10: judging whether the current action is
Figure 955515DEST_PATH_IMAGE010
Deploying actions, if yes, jumping to the step 11; if not, skipping to the step 3;
step 11: deploying virtual network elements in a current virtual network
Figure 669174DEST_PATH_IMAGE026
Or virtual network element set in virtual network tiles
Figure 431593DEST_PATH_IMAGE027
To the currently selected physical host
Figure 409913DEST_PATH_IMAGE028
Updating the state S of the physical host according to the attribute value;
step 12: judging whether the virtual network is completely deployed or not, and if so, skipping to the step 2; if not, skipping to the step 3.
Preferably, in said step 3, according to
Figure 458641DEST_PATH_IMAGE029
Finding the physical host with the largest resource supply
Figure 155201DEST_PATH_IMAGE030
Preferably, in step 7, the virtual network element with the largest out-degree in the undeployed part of the virtual network is used
Figure 190154DEST_PATH_IMAGE031
As a center, the block is expanded, and a virtual network element set in the virtual network block is gradually constructed by adopting breadth-first search
Figure 852341DEST_PATH_IMAGE032
And skipping to step 8.
Preferably, in step 9: updating the action cost function in the current action cost function table according to the following formula
Figure 9653DEST_PATH_IMAGE033
The formula is expressed as:
Figure 131193DEST_PATH_IMAGE034
wherein the prize is awarded
Figure 297732DEST_PATH_IMAGE035
Acting on behalf of the current state s
Figure 47382DEST_PATH_IMAGE036
The short-term benefit is obtained by the method,
Figure 139710DEST_PATH_IMAGE037
representing all optional actions in the current state s
Figure 482966DEST_PATH_IMAGE038
Maximum long term benefit obtainable in (1)
Figure 453197DEST_PATH_IMAGE039
Indicates that the action is selected
Figure 322932DEST_PATH_IMAGE038
After that, a jump is made to a new state, max denotes taking the maximum value,
Figure 87626DEST_PATH_IMAGE040
represents the summation of the short-term benefit and the long-term benefit, and is the subsequent maximum benefit which can be obtained in the current state, wherein
Figure 918179DEST_PATH_IMAGE041
For discount rate, representing long-term benefit
Figure 724723DEST_PATH_IMAGE042
The influence rate of the benefit in the current state is closer to 1, which means that the long-term benefit is emphasized more, and conversely, the short-term benefit is emphasized more,
Figure 386649DEST_PATH_IMAGE043
indicates this iteration to select a new action
Figure 587823DEST_PATH_IMAGE044
With prime mover
Figure 640092DEST_PATH_IMAGE045
A return gain formed therebetween, wherein
Figure 748863DEST_PATH_IMAGE046
The learning rate represents the speed of reinforcement learning, and the closer to 1 represents the faster learning, and the slower learning is vice versa; the whole formula
Figure 23550DEST_PATH_IMAGE047
The representative continuously updates each action taken in each state s by iteratively calculating the return gain
Figure 67729DEST_PATH_IMAGE048
Long term benefits that can be obtained
Figure 731929DEST_PATH_IMAGE049
Thereby enabling the system to autonomously select the optimal action by learning.
Preferably, in the step 8,
Figure 582073DEST_PATH_IMAGE050
denoted by physical host at time t
Figure 156274DEST_PATH_IMAGE051
The provided multidimensional resources mainly comprise CPU resources of a processor, RAM resources of a memory and DISK resources of a DISK.
A reinforcement learning based hierarchical distributed deployment system for a virtual network, comprising:
the action value function table building module: for according to each physical host
Figure 699251DEST_PATH_IMAGE051
Establishing an action value function to form an action value function table;
the virtual network deployment request processing module: the system comprises a physical host searching module, a resource allocation module and a resource allocation module, wherein the physical host searching module is used for sending a signal to control the resource allocation module to work when a new virtual network allocation request arrives;
the physical host search module with the largest resource supply connected with the virtual network deployment request processing module: for observing resource supply according to physical host machine
Figure 289894DEST_PATH_IMAGE052
Finding the physical host with the largest resource supply
Figure 6046DEST_PATH_IMAGE053
And a first judgment module of the searching module of the physical host with the largest resource supply: a physical host for determining that the resource supply is maximum
Figure 497070DEST_PATH_IMAGE054
Whether or not the virtual network can be accommodated,
if the module can be accommodated, the direct deployment module is controlled to work,
if the block can not be accommodated, controlling the block deployment module to work;
the direct deployment module is connected with the first judgment module: for direct deployment of virtual networks, setting current actions as
Figure 617473DEST_PATH_IMAGE055
Deploying, and sending a signal to control the calculation module to start working;
the dicing deployment module is connected with the first judgment module: for selecting an action according to an action cost function if the action is
Figure 521844DEST_PATH_IMAGE055
Deploying, and sending a signal to control the computing module to start working, if the action is
Figure 212326DEST_PATH_IMAGE056
Expanding and sending a signal to control a virtual network element set building module to start working;
a virtual network element set constructing module connected with the block deployment module:for maximizing virtual network element in undeployed part of virtual network
Figure 495540DEST_PATH_IMAGE057
As the center, the expansion of the blocks is carried out, and the virtual network element set in the virtual network blocks is gradually constructed
Figure 177057DEST_PATH_IMAGE058
Sending a signal to control the calculation module to start working;
the computing module is connected with the direct deployment module and the virtual network element set constructing module: for calculating a prize according to the formula
Figure 240828DEST_PATH_IMAGE059
Figure 174149DEST_PATH_IMAGE060
In the formula,
Figure 374186DEST_PATH_IMAGE061
the physical host with maximum resource supply at the time of t
Figure 993649DEST_PATH_IMAGE062
The number of deployed virtual network tiles,
Figure 216820DEST_PATH_IMAGE063
for cutting blocks from virtual networks
Figure 750569DEST_PATH_IMAGE058
The sum of the multi-dimensional resources consumed,
Figure 133009DEST_PATH_IMAGE064
is the largest physical host
Figure 359591DEST_PATH_IMAGE062
Observations of resource provisioning;
and the placeThe updating module is connected with the action value function table building module: for according to the reward
Figure 70058DEST_PATH_IMAGE059
Updating action cost function in current action cost function table
Figure 911893DEST_PATH_IMAGE065
And the second judgment module is connected with the action value function table construction module: for judging whether the current action is
Figure 414418DEST_PATH_IMAGE066
Deploying, if so, sending a signal to control the deployment processing module to start working; if not, sending a signal to control the physical host search module with the maximum resource supply to start working;
a deployment processing module connected to the second determination module: method for deploying virtual network elements in current virtual network
Figure 218426DEST_PATH_IMAGE067
Or virtual network element set in virtual network tiles
Figure 744086DEST_PATH_IMAGE058
To the currently selected physical host
Figure 681955DEST_PATH_IMAGE062
Updating the state S of the physical host according to the attribute value;
a third judgment module connected with the deployment processing module, the virtual network deployment request processing module and the physical host search module with the maximum resource supply: judging whether the virtual network is completely deployed, if so, sending a signal to control the virtual network deployment request processing module to work; if not, sending a signal to control the physical host search module with the maximum resource supply to start working.
Preferably, the physical host search module with the largest resource supplyA block according to
Figure 648774DEST_PATH_IMAGE068
Finding the physical host with the largest resource supply
Figure 453044DEST_PATH_IMAGE069
Preferably, the updating module is configured to update the action cost function in the current action cost function table according to the following formula
Figure 528316DEST_PATH_IMAGE065
The formula is expressed as:
Figure 879663DEST_PATH_IMAGE034
wherein the prize is awarded
Figure 28885DEST_PATH_IMAGE035
Acting on behalf of the current state s
Figure 830487DEST_PATH_IMAGE036
The short-term benefit is obtained by the method,
Figure 737264DEST_PATH_IMAGE037
representing all optional actions in the current state s
Figure 453153DEST_PATH_IMAGE038
Maximum long term benefit obtainable in (1)
Figure 784778DEST_PATH_IMAGE039
Indicates that the action is selected
Figure 429386DEST_PATH_IMAGE038
After that, a jump is made to a new state, max denotes taking the maximum value,
Figure 89037DEST_PATH_IMAGE040
represents the summation of the short-term benefit and the long-term benefit, and is the subsequent maximum benefit which can be obtained in the current state, wherein
Figure 906820DEST_PATH_IMAGE041
For discount rate, representing long-term benefit
Figure 532099DEST_PATH_IMAGE042
The influence rate of the benefit in the current state is closer to 1, which means that the long-term benefit is emphasized more, and conversely, the short-term benefit is emphasized more,
Figure 550871DEST_PATH_IMAGE043
indicates this iteration to select a new action
Figure 760135DEST_PATH_IMAGE044
With prime mover
Figure 381609DEST_PATH_IMAGE045
A return gain formed therebetween, wherein
Figure 32033DEST_PATH_IMAGE046
The learning rate represents the speed of reinforcement learning, and the closer to 1 represents the faster learning, and the slower learning is vice versa; the whole formula
Figure 284023DEST_PATH_IMAGE047
The representative continuously updates each action taken in each state s by iteratively calculating the return gain
Figure 559014DEST_PATH_IMAGE048
Long term benefits that can be obtained
Figure 593966DEST_PATH_IMAGE049
Thereby enabling the system to autonomously select the optimal action by learning.
Has the advantages that:
the invention aims at network simulation based on Docker containerized virtual network, aims at the problems of optimizing, cutting and deploying virtual network, and achieves the purpose of layering and optimizing and deploying virtual network by a reinforcement learning framework under the distributed environment of physical host machines with limited computing, network and storage resources. The beneficial effects mainly include:
the method is based on the lightweight virtualization technology: the deployment algorithm enables the reinforcement learning framework to be effectively adapted to the hierarchical and distributed deployment scenes of the virtual network according to the characteristics of low consumption and fine granularity of Docker and the light-weight characteristics of network virtualization technologies such as OVS and VxLAN.
Autonomous dynamic learning: the process of determining the block size of the virtual network mainly comprises the long-term benefit of the system according to the state and the action
Figure 426793DEST_PATH_IMAGE065
The system is determined by the algorithm which is not designed in advance basically, so that subjective interference of artificial algorithm design is less, the system has better dynamic and autonomous learning capability, the resource supply of a physical host machine and the resource consumption requirement of a virtual network are dynamically matched, and optimal block cutting and deployment are realized.
Resource consumption balancing: the design of reinforcement learning reward considers the resource consumption of a physical host and the communication performance loss of the virtual network block across the hosts, on one hand, the number of blocks of the virtual network block is controlled as much as possible, and on the other hand, the scale of the virtual network block is controlled as much as possible. Macroscopically, the difference of the scales of the virtual networks is not too large, so that the resource consumption of the physical host machine is balanced.
Drawings
In order to more clearly illustrate the detailed description of the invention or the technical solutions in the prior art, the drawings that are needed in the detailed description of the invention or the prior art will be briefly described below. Throughout the drawings, like elements or portions are generally identified by like reference numerals. In the drawings, elements or portions are not necessarily drawn to scale.
Fig. 1 is a schematic view of a hierarchical distributed deployment of a virtual network in the present invention;
FIG. 2 is a framework of a virtual network hierarchical distributed deployment system based on reinforcement learning in the present invention;
FIG. 3 is a flow of hierarchical distributed deployment of a reinforcement learning-based virtual network in the present invention;
fig. 4 is schematic diagrams of two virtual networks to be deployed in the present invention, and in fig. 4, (a) shows a smaller-scale virtual network to be deployed and (b) shows a larger-scale virtual network to be deployed.
Detailed Description
The technical solutions in the embodiments of the present invention will be described clearly and completely below, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that all the directional indicators (such as upper, lower, left, right, front and rear … …) in the embodiment of the present invention are only used to explain the relative position relationship between the components, the movement situation, etc. in a specific posture (as shown in the drawing), and if the specific posture is changed, the directional indicator is changed accordingly.
In addition, the descriptions related to "first", "second", etc. in the present invention are only for descriptive purposes and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include at least one such feature. In the description of the present invention, "a plurality" means at least two, e.g., two, three, etc., unless specifically limited otherwise.
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The invention will now be further described with reference to the accompanying drawings.
In order to facilitate the subsequent introduction of a specific technical scheme, concepts used in the invention are defined, and the hierarchical distributed deployment problem of the virtual network is modeled.
Logical virtual network: is an overall logic description of the virtual network to be deployed, input by the user, and directly faces the user, and is an undirected graph
Figure 912001DEST_PATH_IMAGE070
Wherein
Figure 95857DEST_PATH_IMAGE071
A set of virtual network elements is represented,
Figure 934500DEST_PATH_IMAGE072
a set of virtual links is represented as a set of virtual links,
Figure 388878DEST_PATH_IMAGE073
can also be recorded as
Figure 44987DEST_PATH_IMAGE074
Representing virtual network elements
Figure 450561DEST_PATH_IMAGE075
And
Figure 358474DEST_PATH_IMAGE076
a virtual link between them. As shown at the top of fig. 1.
Virtual network dicing: when the logic virtual network can not be integrally deployed on a host machine, the logic virtual network is cut into a plurality of virtual network blocks according to a certain algorithm
Figure 165893DEST_PATH_IMAGE077
That is, the virtual network partition is a local description of the logical virtual network after being partitioned, and is transparent to the user. The mth virtual network is cut into undirected graphs
Figure 491439DEST_PATH_IMAGE078
Wherein
Figure 384308DEST_PATH_IMAGE079
A set of virtual network elements is represented,
Figure 564754DEST_PATH_IMAGE080
a set of virtual links is represented as a set of virtual links,
Figure 288996DEST_PATH_IMAGE081
can also be recorded as
Figure 490170DEST_PATH_IMAGE082
Representing virtual network elements
Figure 542440DEST_PATH_IMAGE083
And
Figure 90358DEST_PATH_IMAGE084
a virtual link between them. As shown in the middle "cut a" and "cut B" of fig. 1.
Physical network: is a network formed by physical hosts for deploying virtual networks and is an undirected graph
Figure 934686DEST_PATH_IMAGE085
Wherein
Figure 41183DEST_PATH_IMAGE086
Represents a collection of physical host machines that are,
Figure 580748DEST_PATH_IMAGE087
a set of physical links is represented as,
Figure 493210DEST_PATH_IMAGE088
can also be recorded as
Figure 634122DEST_PATH_IMAGE089
Denotes a physical host
Figure 849203DEST_PATH_IMAGE090
And
Figure 938381DEST_PATH_IMAGE091
the physical link between them. As shown at the bottom of fig. 1.
Additional auxiliary networks: obviously, the virtual network segment may include a plurality of additional virtual network elements and virtual links transparent to the user, and is used for purposes of virtual network segment cross-host communication, and the like, so the following mathematical relationship holds:
Figure 388954DEST_PATH_IMAGE092
. The additional auxiliary network is an auxiliary virtual network for connecting the virtual network blocks
Figure 817661DEST_PATH_IMAGE093
Transparent to the user, automatically generated by an algorithm on demand, wherein
Figure 265960DEST_PATH_IMAGE094
A set of additional virtual network elements is represented,
Figure 406217DEST_PATH_IMAGE095
representing a set of additional virtual links. As shown in fig. 1, OVS1, OVS2 are additional virtual network elements, and veth-pair2, veth-pair3, VxLAN are additional virtual links.
Resource supply: at time t, by a physical host
Figure 535847DEST_PATH_IMAGE096
The provided multidimensional resources mainly comprise resources such as a CPU (central processing unit), an RAM (memory access memory), a DISK (DISK drive), and the like, and are marked as
Figure 881378DEST_PATH_IMAGE097
And by physical links
Figure 562895DEST_PATH_IMAGE098
The provided multidimensional resources mainly comprise resources such as bandwidth BW and the like, and are marked as
Figure 564349DEST_PATH_IMAGE099
Resource consumption: at the time of t, the virtual network elements in the virtual network block are collected
Figure 294407DEST_PATH_IMAGE100
The sum of the consumed multidimensional resources mainly includes resources such as a processor CPU, a memory RAM, a DISK DISK and the like, and is marked as
Figure 320876DEST_PATH_IMAGE101
And the virtual link set in the virtual network block
Figure 376557DEST_PATH_IMAGE102
The sum of the consumed multidimensional resources, mainly including bandwidth BW and other resources, is marked as
Figure 599727DEST_PATH_IMAGE103
Distributed deployment of virtual networks: the method can be modeled as a 0-1 planning problem, the optimization goal is to minimize the number M of virtual network blocks so as to reduce the performance loss caused by cross-host communication, and the constraint condition is that all virtual network blocks of a logical virtual network are mapped and deployed on a certain physical host and the relationship between resource supply and resource consumption is correctly matched. Solved by
Figure 399056DEST_PATH_IMAGE104
Is a variable of 0-1 and is used for determining virtual network blocks
Figure 781496DEST_PATH_IMAGE105
Whether or not to be deployed in a physical host
Figure 243964DEST_PATH_IMAGE106
And (4) the following steps. The mathematical expression is as follows:
Figure 220010DEST_PATH_IMAGE107
the invention combines virtualization according to a reinforcement learning frameworkThe concrete requirement of hierarchical and distributed deployment of the network, the quintuple for strengthening the learning
Figure 557450DEST_PATH_IMAGE108
Modeling and design were performed, where:
s: finite set of states, here observations of the supply of physical host resources, i.e.
Figure 794397DEST_PATH_IMAGE109
Due to the fact that
Figure 660721DEST_PATH_IMAGE110
The contained attribute of each dimension is a continuous space, on one hand, fitting can be carried out by matching a Deep reinforcement learning framework such as DQN (Deep Q-Network) with a convolution neural Network, and on the other hand, fitting can be carried out by matching the Deep reinforcement learning framework such as DQN (Deep Q-Network) with the convolution neural Network
Figure 944636DEST_PATH_IMAGE111
The continuous state space of the multi-dimensional attributes is discretized, finite states are constructed based on the attributes, and a lightweight Q-learning algorithm can be adopted for solving. The invention adopts a second, more lightweight method, and the discretization is specifically referred to the subsequent step 1.
A: a finite set of movements, consisting essentially of two movements, (1)
Figure 351346DEST_PATH_IMAGE112
Deployment action: deploying entire logical virtual network or current virtual network tiles
Figure 911641DEST_PATH_IMAGE113
Physical hosts with maximum resource supply
Figure 309386DEST_PATH_IMAGE114
Wherein
Figure 211090DEST_PATH_IMAGE115
And will correspond to
Figure 828016DEST_PATH_IMAGE116
1, placing; (2)
Figure 497944DEST_PATH_IMAGE117
expanding action: continuing to use the virtual network element with the largest out-degree in the undeployed part of the virtual network
Figure 237229DEST_PATH_IMAGE118
Is a center (wherein
Figure 268639DEST_PATH_IMAGE119
) Expanding by breadth-first search to construct virtual network blocks
Figure 905901DEST_PATH_IMAGE120
P: the finite set of transition probabilities between states, which is not involved in the algorithm, is model-free reinforcement learning and can be ignored.
R: the concrete modeling calculation mode of the set of the rewards corresponding to each action is shown in the subsequent step 8.
Figure 175208DEST_PATH_IMAGE121
: the discount factor is a factor of the discount,
Figure 649177DEST_PATH_IMAGE122
and indicates the influence degree of the reward of the follow-up action on the current action.
The method system framework constructed in this way is shown in fig. 2, and comprises 4 core modules and 1 core database: the system comprises a virtual network analysis module, a cutting judgment module, an optimized cutting module, an optimized deployment module and a Q table database.
The specific technical scheme and flow of the virtual network hierarchical distributed deployment based on reinforcement learning are as follows, and refer to fig. 3.
Step 1 is that each physical host machine
Figure 636725DEST_PATH_IMAGE123
Establishing independent action pricesThe value function table (i.e., the Q table, whose structure is shown in Table 1), row state, and column action. Each row state represents resource provisioning of a physical host
Figure 657770DEST_PATH_IMAGE124
A number of subrows are included, i.e. the state is a linear combination of attributes. Because each dimension attribute is a continuous state space, the dimension attributes are discretized according to a certain rule (for example, the memory RAM can be segmented on the basis of 4 GB), and finite states are constructed on the basis of the attribute segments. Within a cell is a corresponding action cost function
Figure 47163DEST_PATH_IMAGE125
And is initialized to 0.
Table 1 physical host machine
Figure 800356DEST_PATH_IMAGE126
Structure of (Q) table
Figure 338233DEST_PATH_IMAGE127
And 2, entering a main loop of the virtual network deployment algorithm, and waiting for a new virtual network deployment request. And when a new virtual network deployment request arrives, skipping to the step 3.
Step 3, finding out the physical host with the maximum resource supply
Figure 959707DEST_PATH_IMAGE128
Wherein
Figure 875711DEST_PATH_IMAGE129
And switching to the corresponding Q table.
Step 4, judging the physical host machine
Figure 127701DEST_PATH_IMAGE130
Whether the virtual network can be accommodated. If the data can be accommodated, jumping to the step 5; if not, jumping to step 6.
Step 5, starting a direct deployment processSet the current action to
Figure 886578DEST_PATH_IMAGE131
And deploying and skipping to the step 8.
Step 6, starting a dicing deployment process by adopting
Figure 750891DEST_PATH_IMAGE132
Algorithm according to
Figure 521401DEST_PATH_IMAGE133
The selection of the action is made in such a way that,
Figure 944292DEST_PATH_IMAGE134
is a smaller value (e.g. of
Figure 924887DEST_PATH_IMAGE135
) The method is used for encouraging the network to be expanded as much as possible and reducing the number of blocks. If the action is
Figure 763530DEST_PATH_IMAGE136
Deploying and skipping step 8; if the action is
Figure 716442DEST_PATH_IMAGE137
Expanding and jumping to step 7; and setting the current action according to the selection result.
Figure 871086DEST_PATH_IMAGE138
Step 7, continuing to use the virtual network element with the largest out-degree in the undeployed part of the virtual network
Figure 276660DEST_PATH_IMAGE139
Is a center (wherein
Figure 918994DEST_PATH_IMAGE140
) Expanding the blocks by adopting breadth-first search and gradually constructing virtual network blocks
Figure 726413DEST_PATH_IMAGE141
And skipping to step 8.
Step 8 calculates the prize according to the following formula
Figure 553424DEST_PATH_IMAGE142
Figure 947758DEST_PATH_IMAGE143
a) The left side of the right half of the formula can be understood as: the dicing deployment reward encourages the dicing to be as large as possible, so that the dicing quantity is as small as possible, and the extra performance loss of cross-host communication is reduced. Wherein
Figure 128204DEST_PATH_IMAGE144
At time t, the physical host
Figure 321288DEST_PATH_IMAGE145
The number of deployed virtual network tiles. This part is a positive number, so it can be seen that there is a reward for successful deployment, but the more virtual networks deployed, the faster the reward decays, thus suppressing the excessive number of blocks.
b) The right side of the right half of the formula can be understood as: and the block expansion reward encourages the block to be as small as possible, and ensures that the physical host consumes resources as few as possible so as to accommodate the deployment of other subsequent virtual networks. It can be seen that the more resources are consumed by the virtual network in blocks, the faster the reward is decayed. When the cut piece is over-expanded, i.e.
Figure 319199DEST_PATH_IMAGE146
A deployment failure will result and a negative reward is formed to inhibit the dice from over-expanding.
The two parts of rewards are mutually restricted, and the virtual network block size and the deployment mode matched with the existing system resource supply are learned through the continuous operation of the reinforcement learning framework.
Step 9, updating the current Q table according to the following formula according to the Q-learning algorithm
Figure 433786DEST_PATH_IMAGE147
The formula is expressed as:
Figure 417922DEST_PATH_IMAGE148
wherein
Figure 766645DEST_PATH_IMAGE149
Is the maximum gain that can be obtained in the current state, often called target Q,
Figure 873141DEST_PATH_IMAGE147
the accumulated reward is now, and the two are subtracted, so that the reward gain, that is, the TD deviation (temporal difference error), is obtained. Wherein
Figure 412707DEST_PATH_IMAGE150
To learn the rate, show the gain in return for
Figure 262851DEST_PATH_IMAGE147
The degree of influence of (c);
specifically, awards
Figure 961686DEST_PATH_IMAGE035
Acting on behalf of the current state s
Figure 740548DEST_PATH_IMAGE036
The short-term benefit is obtained by the method,
Figure 767410DEST_PATH_IMAGE037
representing all optional actions in the current state s
Figure 217983DEST_PATH_IMAGE038
Maximum long term benefit obtainable in (1)
Figure 974587DEST_PATH_IMAGE039
Indicates that the action is selected
Figure 360569DEST_PATH_IMAGE038
After that, a jump is made to a new state, max denotes taking the maximum value,
Figure 937043DEST_PATH_IMAGE040
represents the summation of the short-term benefit and the long-term benefit, and is the subsequent maximum benefit which can be obtained in the current state, wherein
Figure 424263DEST_PATH_IMAGE041
For discount rate, representing long-term benefit
Figure 973056DEST_PATH_IMAGE042
The influence rate of the benefit in the current state is closer to 1, which means that the long-term benefit is emphasized more, and conversely, the short-term benefit is emphasized more,
Figure 592256DEST_PATH_IMAGE043
indicates this iteration to select a new action
Figure 452765DEST_PATH_IMAGE044
With prime mover
Figure 386086DEST_PATH_IMAGE045
A return gain formed therebetween, wherein
Figure 851702DEST_PATH_IMAGE046
The learning rate represents the speed of reinforcement learning, and the closer to 1 represents the faster learning, and the slower learning is vice versa; the whole formula
Figure 205586DEST_PATH_IMAGE047
The representative continuously updates each action taken in each state s by iteratively calculating the return gain
Figure 428757DEST_PATH_IMAGE048
Long term benefits that can be obtained
Figure 228085DEST_PATH_IMAGE049
Thereby enabling the system to autonomously select the optimal action by learning.
Step 10 of judging whether the current action is
Figure 610525DEST_PATH_IMAGE151
Deploying actions, if yes, jumping to the step 11; if not, skipping to the step 3.
Step 11 deploys the current virtual network
Figure 571528DEST_PATH_IMAGE152
Or cutting into pieces
Figure 547574DEST_PATH_IMAGE153
To the currently selected physical host
Figure 377690DEST_PATH_IMAGE128
And updating the state S of the physical host according to the attribute value.
Step 12 determines whether the virtual network has been completely deployed. If yes, skipping to the step 2; if not, skipping to the step 3.
This example will deploy two virtual networks topo1 and topo2 onto physical hosts H1 and H2 as in fig. 4. To simplify the discussion, the multidimensional index is simplified into a one-dimensional index of the RAM. H1 and H2 are physical hosts of the same configuration, and the RAM supply is 16 GB, i.e.
Figure 614637DEST_PATH_IMAGE154
. The RAM is segmented into 4 gears, and four states are constructed
Figure 12120DEST_PATH_IMAGE155
To
Figure 944304DEST_PATH_IMAGE156
H1 and H2 form a Q table as shown in table 2, and the Q value is initialized to 0. The memory consumption of each virtual network element in the virtual network is 0.5 GB, namely
Figure 351014DEST_PATH_IMAGE157
. According to the aboveConfiguration, see
Figure 943932DEST_PATH_IMAGE158
That is, when topo1 is deployed first, no cutting may be used, and when topo2 is deployed next, a cutting must be made. The parameters for the Q-learning reinforcement learning calculation are:
Figure 512317DEST_PATH_IMAGE159
TABLE 2Q-Table Structure of physical hosts H1 or H2
Figure 197376DEST_PATH_IMAGE160
First, deploying topo 1: the host H1 with the largest resource supply at this time is found to be
Figure 876619DEST_PATH_IMAGE161
State (see Table 2), the virtual network can be accommodated, selected according to step 5
Figure 822578DEST_PATH_IMAGE162
Skipping to step 8 to obtain the corresponding reward
Figure 499547DEST_PATH_IMAGE163
Is calculated according to step 9
Figure 232755DEST_PATH_IMAGE164
Fill its Q table and finally deploy directly to H1. After deployment, the number of virtual network blocks on the host
Figure 778005DEST_PATH_IMAGE165
As shown in table 3, line 1.
Second, topo2 is next deployed.
a) Deploying a first block of virtual network blocks: the host H2 with the largest resource supply at this time is found to be
Figure 781733DEST_PATH_IMAGE166
Status (see Table 2), unable to accommodate the virtual network, jump to step 6, based on
Figure 364025DEST_PATH_IMAGE167
Algorithm due to
Figure 148310DEST_PATH_IMAGE168
Selecting with great probability
Figure 670820DEST_PATH_IMAGE169
Act of
Figure 466738DEST_PATH_IMAGE170
And expanding the existing network blocks by using the R3 virtual network element with the maximum output as a center and adopting breadth-first search. Then skipping to step 8 to obtain the corresponding reward
Figure 547826DEST_PATH_IMAGE171
Is calculated according to step 9
Figure 553829DEST_PATH_IMAGE172
Fill in its Q table. Resource consumption at this point due to no actual deployment
Figure 316248DEST_PATH_IMAGE173
As shown in table 3, line 2. Subsequent rounds based on
Figure 294568DEST_PATH_IMAGE174
And
Figure 113270DEST_PATH_IMAGE175
(it is noted that,
Figure 216355DEST_PATH_IMAGE175
since the accumulation of each round of awards changes gradually), continuously and repeatedly selecting
Figure 579203DEST_PATH_IMAGE176
Continuing to expand existing network tiles, similar to Table 3 line 2, with specific changes such asTable 3, lines 3-11. The latter round according to
Figure 739926DEST_PATH_IMAGE174
(Note that
Figure 162817DEST_PATH_IMAGE177
) It is possible to select those having a non-maximum Q value
Figure 753198DEST_PATH_IMAGE178
Skipping to step 8 to obtain the corresponding reward
Figure 483519DEST_PATH_IMAGE179
Is calculated according to step 9
Figure 436432DEST_PATH_IMAGE180
Fill its Q table and finally deploy directly to H2. After deployment, the number of virtual network blocks on the host
Figure 702328DEST_PATH_IMAGE181
As shown in table 3, line 10. And finishing the deployment of the first block of virtual network blocks. It should be noted that although a virtual network block of 10 virtual network elements is constructed and deployed here (rows 3-12 of table 3), how many network elements a particular block contains is composed of
Figure 170219DEST_PATH_IMAGE182
(reflecting the long-term benefits of a particular action in a particular state) and
Figure 874869DEST_PATH_IMAGE174
(introduction of the randomness to avoid the action selection rigidity) is determined together, in the specific implementation, the number of the virtual network elements contained in the virtual network block is not necessarily 10, and 10 are only used as examples here.
b) Deploying a second virtual network block: after the deployment of the first virtual network partition is completed, the entire topo2 is not yet deployed, and the algorithm needs to continue to run. The host H1 with the largest resource supply at this time is found to be
Figure 619971DEST_PATH_IMAGE183
Status (see Table 2), unable to accommodate the virtual network, jump to step 6, based on
Figure 883200DEST_PATH_IMAGE174
Algorithm due to
Figure 572808DEST_PATH_IMAGE184
Selecting with great probability
Figure 18832DEST_PATH_IMAGE185
Act of
Figure 680758DEST_PATH_IMAGE186
And expanding the existing network blocks by using the R3 virtual network element with the maximum output as a center and adopting breadth-first search. Then skipping to step 8 to obtain the corresponding reward
Figure 678670DEST_PATH_IMAGE187
Is calculated according to step 9
Figure 730939DEST_PATH_IMAGE188
Fill in its Q table. Resource consumption at this point due to no actual deployment
Figure 544437DEST_PATH_IMAGE189
As shown in table 3, line 13. Subsequent rounds based on
Figure 123186DEST_PATH_IMAGE190
And
Figure 229682DEST_PATH_IMAGE191
(it is noted that,
Figure 769248DEST_PATH_IMAGE191
due to progressive release of the award in each round), continuously and repeatedly selecting
Figure 681709DEST_PATH_IMAGE192
The existing network cut was expanded continuously, similar to table 3 line 13, with specific changes as shown in table 3 lines 14-18. The latter round according to
Figure 834340DEST_PATH_IMAGE193
(Note that
Figure 49421DEST_PATH_IMAGE194
) It is possible to select
Figure 200916DEST_PATH_IMAGE195
Not of maximum value
Figure 589172DEST_PATH_IMAGE196
Skipping to step 8 to obtain the corresponding reward
Figure 752300DEST_PATH_IMAGE197
Is calculated according to step 9
Figure 731758DEST_PATH_IMAGE198
Fill its Q table and finally deploy directly to H1. After deployment, the number of virtual network blocks on the host
Figure 872014DEST_PATH_IMAGE199
As shown in table 3, line 19. And finishing the deployment of the second virtual network block. It should be noted that although a virtual network block of 7 virtual network elements is constructed and deployed here (rows 13-19 of table 3), how many network elements a particular block contains is composed of
Figure 63961DEST_PATH_IMAGE200
(reflecting the long-term benefits of a particular action in a particular state) and
Figure 612754DEST_PATH_IMAGE190
(introduction of the randomness to avoid the action selection rigidity) is determined together, in the specific implementation, the number of the virtual network elements contained in the virtual network block is not necessarily 7, and the 7 are only used as examples here.
c) Transition of state: after the second virtual network block is deployed, two blocks (17 virtual network elements) are deployed in the entire topo2, and the algorithm needs to be continuously run if the deployment is not completed. The host H2 with the largest resource supply at this time is found, and since the resource consumption reaches the critical condition of state transition after the virtual network is actually deployed, the state is changed from the original state
Figure 231954DEST_PATH_IMAGE201
Switch to
Figure 92463DEST_PATH_IMAGE202
(see Table 2) to iteratively calculate the new state based on how the first and second virtual network tiles are deployed
Figure 25784DEST_PATH_IMAGE202
Corresponding to
Figure 989936DEST_PATH_IMAGE203
And
Figure 576775DEST_PATH_IMAGE204
to thereby optimally select actions for subsequent processing
Figure 65525DEST_PATH_IMAGE205
Or
Figure 599274DEST_PATH_IMAGE206
The optimization cuts and deployments are made to provide a numerical basis, as shown in rows 20-21 of table 3. Not all subsequent iteration steps are listed here, limited to space.
Table 3 deployment virtual network topo1 and topo2 examples
Figure 981714DEST_PATH_IMAGE207
The above embodiments are only used to illustrate the technical solution of the present invention, and not to limit the same; while the invention has been described in detail and with reference to the foregoing embodiments, it will be understood by those skilled in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; such modifications and substitutions do not depart from the spirit and scope of the present invention, and they should be construed as being included in the following claims and description.

Claims (8)

1. A virtual network hierarchical distributed deployment method based on reinforcement learning is characterized by comprising the following steps:
step 1: for each physical host
Figure 738706DEST_PATH_IMAGE001
Establishing an independent action value function table, wherein corresponding action value functions are arranged in cells
Figure 28873DEST_PATH_IMAGE002
And is initialized to 0, wherein,
Figure 158372DEST_PATH_IMAGE001
represents a physical host machine, the superscript p represents physical, the subscript r represents the number of the physical host machine, and the value range is
Figure 533990DEST_PATH_IMAGE003
R is the total number of physical host machines; s represents a state in reinforcement learning,
Figure 485765DEST_PATH_IMAGE004
representing actions in reinforcement learning, action cost function
Figure 145286DEST_PATH_IMAGE002
Taking actions on behalf of a state s in reinforcement learning
Figure 196418DEST_PATH_IMAGE004
Long term benefits of;
step 2: waiting for a new virtual network deployment request, and jumping to the step 3 when the new virtual network deployment request arrives;
and step 3: based on observations of resource supply of physical hosts
Figure 308600DEST_PATH_IMAGE005
Finding the physical host with the largest resource supply
Figure 673853DEST_PATH_IMAGE006
(ii) a Wherein,
Figure 453459DEST_PATH_IMAGE005
representing at time t, by a physical host
Figure 3389DEST_PATH_IMAGE006
The multi-dimensional resources that are provided,
Figure 88020DEST_PATH_IMAGE007
supplying the number of the largest physical host to the resource;
and 4, step 4: judging the physical host machine
Figure 568548DEST_PATH_IMAGE006
Whether or not the virtual network can be accommodated,
if so, jumping to step 5,
if the data can not be accommodated, jumping to step 6;
and 5: direct deployment, setting current actions as
Figure 687814DEST_PATH_IMAGE008
Deploying and skipping to step 8;
step 6: deployment of blocks according to action cost function
Figure 533279DEST_PATH_IMAGE002
SelectingAction if the action is
Figure 370785DEST_PATH_IMAGE008
Deploying, skipping to step 8, if the action is
Figure 407004DEST_PATH_IMAGE009
Expanding and jumping to step 7;
and 7: virtual network element with maximum out-degree in undeployed part of virtual network
Figure 380776DEST_PATH_IMAGE010
As the center, the expansion of the blocks is carried out, and the virtual network element set in the virtual network blocks is gradually constructed
Figure 538088DEST_PATH_IMAGE011
Skipping to step 8; wherein,
Figure 49841DEST_PATH_IMAGE010
for virtual network elements in a virtual network, superscripts
Figure 622905DEST_PATH_IMAGE012
Represents a local, subscript
Figure 44659DEST_PATH_IMAGE013
The number representing the virtual network element has a value range of
Figure 294243DEST_PATH_IMAGE014
IThe total number of the virtual network elements;
Figure 44025DEST_PATH_IMAGE015
the number of the most out-dated virtual network element,
Figure 138888DEST_PATH_IMAGE016
representing the virtual network blocks, the superscript b representing the block, the subscript m representing the number of the virtual network blocks, and the valueIn the range of
Figure 352832DEST_PATH_IMAGE017
Dynamically determining the total number of blocks in the execution process;
and 8: calculating the prize according to the formula
Figure 851946DEST_PATH_IMAGE018
Figure 135029DEST_PATH_IMAGE019
In the formula,
Figure 784316DEST_PATH_IMAGE020
is composed of
Figure 305296DEST_PATH_IMAGE021
Physical host with maximum resource supply at any moment
Figure 647416DEST_PATH_IMAGE022
The number of deployed virtual network tiles,
Figure 152216DEST_PATH_IMAGE023
is represented in𝑡At the moment, the virtual network elements in the virtual network blocks are gathered
Figure 933090DEST_PATH_IMAGE024
The sum of the multi-dimensional resources consumed,
Figure 121626DEST_PATH_IMAGE025
is the largest physical host
Figure 821597DEST_PATH_IMAGE022
Observations of resource provisioning;
and step 9: according to the reward
Figure 830005DEST_PATH_IMAGE018
Updating action cost function in current action cost function table
Figure 414570DEST_PATH_IMAGE002
Step 10: judging whether the current action is
Figure 650422DEST_PATH_IMAGE008
Deploying actions, if yes, jumping to the step 11; if not, skipping to the step 3;
step 11: deploying a current entire virtual network
Figure 334344DEST_PATH_IMAGE026
Or virtual network element set in virtual network tiles
Figure 157944DEST_PATH_IMAGE024
To the currently selected physical host
Figure 405254DEST_PATH_IMAGE022
And updating the state of the physical host according to the attribute value
Figure 568382DEST_PATH_IMAGE027
(ii) a Wherein
Figure 485523DEST_PATH_IMAGE028
Superscript on behalf of the current entire virtual network
Figure 717790DEST_PATH_IMAGE012
Represents local;
step 12: judging whether the virtual network is completely deployed or not, and if so, skipping to the step 2; if not, skipping to the step 3.
2. The method of claim 1, wherein in step 3, the rootAccording to
Figure 253945DEST_PATH_IMAGE029
Finding the physical host with the largest resource supply
Figure 599475DEST_PATH_IMAGE022
And the superscript R is the total number of the physical hosts.
3. The method according to claim 1, wherein in step 7, the most out-of-range virtual network element in the undeployed part of the virtual network is taken as the virtual network element
Figure 874468DEST_PATH_IMAGE030
Is a center in which
Figure 282446DEST_PATH_IMAGE031
Upper label ofIThe total number of the virtual network elements is subjected to block expansion, breadth-first search is adopted, and a virtual network element set in the virtual network blocks is gradually constructed
Figure 12505DEST_PATH_IMAGE024
And skipping to step 8.
4. A method according to claim 1, characterized in that in step 9: updating the action cost function in the current action cost function table according to the following formula
Figure 133914DEST_PATH_IMAGE002
The formula is expressed as:
Figure 64961DEST_PATH_IMAGE032
wherein the prize is awarded
Figure 475082DEST_PATH_IMAGE033
Representing the action taken in the current state sMaking
Figure 743253DEST_PATH_IMAGE034
The short-term benefit is obtained by the method,
Figure 469900DEST_PATH_IMAGE035
representing all optional actions in the current state s
Figure 821116DEST_PATH_IMAGE036
Maximum long term benefit obtainable in (1)
Figure 203687DEST_PATH_IMAGE037
Indicates that the action is selected
Figure 196919DEST_PATH_IMAGE036
After that, a jump is made to a new state, max denotes taking the maximum value,
Figure 840390DEST_PATH_IMAGE038
represents the summation of the short-term benefit and the long-term benefit, and is the subsequent maximum benefit which can be obtained in the current state, wherein
Figure 644398DEST_PATH_IMAGE039
For discount rate, representing long-term benefit
Figure 492094DEST_PATH_IMAGE040
The influence rate of the benefit in the current state is closer to 1, which means that the long-term benefit is emphasized more, and conversely, the short-term benefit is emphasized more,
Figure 39750DEST_PATH_IMAGE041
indicates this iteration to select a new action
Figure 537728DEST_PATH_IMAGE042
With prime mover
Figure 761905DEST_PATH_IMAGE043
A return gain formed therebetween, wherein
Figure 853489DEST_PATH_IMAGE044
The learning rate represents the speed of reinforcement learning, and the closer to 1 represents the faster learning, and the slower learning is vice versa; the whole formula
Figure 188524DEST_PATH_IMAGE045
The representative continuously updates each action taken in each state s by iteratively calculating the return gain
Figure 806587DEST_PATH_IMAGE046
Long term benefits that can be obtained
Figure 952397DEST_PATH_IMAGE047
Thereby enabling the system to autonomously select the optimal action by learning.
5. The method of claim 1, wherein, in step 8,
Figure 780545DEST_PATH_IMAGE048
is shown as
Figure 670004DEST_PATH_IMAGE021
From time to time, by physical host
Figure 142573DEST_PATH_IMAGE049
The provided multidimensional resources mainly comprise CPU resources of a processor, RAM resources of a memory and DISK resources of a DISK.
6. A virtual network hierarchical distributed deployment system based on reinforcement learning is characterized by comprising:
the action value function table building module: for each physical host
Figure 708553DEST_PATH_IMAGE049
Establishing an independent action value function table, wherein corresponding action value functions are arranged in cells
Figure 774729DEST_PATH_IMAGE050
And is initialized to 0, wherein,
Figure 451567DEST_PATH_IMAGE051
represents a physical host machine, the superscript p represents physical, the subscript r represents the number of the physical host machine, and the value range is
Figure 44222DEST_PATH_IMAGE052
R is the total number of physical host machines; s represents a state (state) in reinforcement learning,
Figure 677DEST_PATH_IMAGE053
representing actions in reinforcement learning, action cost function
Figure 600154DEST_PATH_IMAGE054
Acting on a particular state s in a representation reinforcement study
Figure 831415DEST_PATH_IMAGE053
Long term benefits of;
the virtual network deployment request processing module: the system comprises a physical host searching module, a resource allocation module and a resource allocation module, wherein the physical host searching module is used for sending a signal to control the resource allocation module to work when a new virtual network allocation request arrives;
the physical host search module with the largest resource supply connected with the virtual network deployment request processing module: for observing resource supply according to physical host machine
Figure 403211DEST_PATH_IMAGE055
Finding the physical host with the largest resource supply
Figure 61725DEST_PATH_IMAGE056
(ii) a Wherein,
Figure 961548DEST_PATH_IMAGE055
representing at time t, by a physical host
Figure 986048DEST_PATH_IMAGE057
The multi-dimensional resources that are provided,
Figure 428662DEST_PATH_IMAGE058
supplying the number of the largest physical host to the resource;
and a first judgment module of the searching module of the physical host with the largest resource supply: a physical host for determining that the resource supply is maximum
Figure 241766DEST_PATH_IMAGE059
Whether or not the virtual network can be accommodated,
if the module can be accommodated, the direct deployment module is controlled to work,
if the block can not be accommodated, controlling the block deployment module to work;
the direct deployment module is connected with the first judgment module: for direct deployment of virtual networks, setting current actions as
Figure 300989DEST_PATH_IMAGE008
Deploying, and sending a signal to control the calculation module to start working;
the dicing deployment module is connected with the first judgment module: for selecting an action according to an action cost function if the action is
Figure 123320DEST_PATH_IMAGE008
Deploying, and sending a signal to control the computing module to start working, if the action is
Figure 545074DEST_PATH_IMAGE009
Expanding and sending a signal to control a virtual network element set building module to start working;
a virtual network element set constructing module connected with the block deployment module: for maximizing virtual network element in undeployed part of virtual network
Figure 279812DEST_PATH_IMAGE060
As the center, the expansion of the blocks is carried out, and the virtual network element set in the virtual network blocks is gradually constructed
Figure 341178DEST_PATH_IMAGE061
Sending a signal to control the calculation module to start working;
the computing module is connected with the direct deployment module and the virtual network element set constructing module: for calculating a prize according to the formula
Figure 232780DEST_PATH_IMAGE018
Figure 446723DEST_PATH_IMAGE019
In the formula,
Figure 601630DEST_PATH_IMAGE062
is composed of
Figure 838707DEST_PATH_IMAGE021
Physical host with maximum resource supply at any moment
Figure 2841DEST_PATH_IMAGE063
The number of deployed virtual network tiles,
Figure 133608DEST_PATH_IMAGE023
for cutting blocks from virtual networks
Figure 475728DEST_PATH_IMAGE024
Aggregate of multidimensional resources consumedAnd,
Figure 466947DEST_PATH_IMAGE025
is the largest physical host
Figure 185504DEST_PATH_IMAGE022
Observations of resource provisioning;
the updating module is connected with the action value function table building module and comprises: for according to the reward
Figure 170778DEST_PATH_IMAGE018
Updating action cost function in current action cost function table
Figure 136329DEST_PATH_IMAGE002
And the second judgment module is connected with the action value function table construction module: for judging whether the current action is
Figure 410315DEST_PATH_IMAGE008
Deploying, if so, sending a signal to control a deployment processing module to start working; if not, sending a signal to control the physical host search module with the maximum resource supply to start working;
a deployment processing module connected to the second determination module: method for deploying virtual network elements in current virtual network
Figure 729301DEST_PATH_IMAGE064
Or virtual network element set in virtual network tiles
Figure 21611DEST_PATH_IMAGE065
To the currently selected physical host
Figure 643216DEST_PATH_IMAGE066
Updating the state S of the physical host according to the attribute value;
a third judgment module connected with the deployment processing module, the virtual network deployment request processing module and the physical host search module with the maximum resource supply: judging whether the virtual network is completely deployed, if so, sending a signal to control the virtual network deployment request processing module to work; if not, sending a signal to control the physical host search module with the maximum resource supply to start working.
7. The reinforcement learning-based virtual network hierarchical distributed deployment system according to claim 6, wherein the resource-supply-maximum physical host search module is based on
Figure 653767DEST_PATH_IMAGE029
Finding the physical host with the largest resource supply
Figure 776443DEST_PATH_IMAGE022
8. The reinforcement learning-based virtual network hierarchical distributed deployment system according to claim 6, wherein the updating module is configured to update the action cost function in the current action cost function table according to the following formula
Figure 611675DEST_PATH_IMAGE002
The formula is expressed as:
Figure 512504DEST_PATH_IMAGE067
wherein the prize is awarded
Figure 433187DEST_PATH_IMAGE033
Acting on behalf of the current state s
Figure 280926DEST_PATH_IMAGE034
The short-term benefit is obtained by the method,
Figure 360877DEST_PATH_IMAGE035
representing all optional actions in the current state s
Figure 121023DEST_PATH_IMAGE036
Maximum long term benefit obtainable in (1)
Figure 575007DEST_PATH_IMAGE037
Indicates that the action is selected
Figure 977170DEST_PATH_IMAGE036
After that, a jump is made to a new state, max denotes taking the maximum value,
Figure 36261DEST_PATH_IMAGE038
represents the summation of the short-term benefit and the long-term benefit, and is the subsequent maximum benefit which can be obtained in the current state, wherein
Figure 232887DEST_PATH_IMAGE039
For discount rate, representing long-term benefit
Figure 252796DEST_PATH_IMAGE040
The influence rate of the benefit in the current state is closer to 1, which means that the long-term benefit is emphasized more, and conversely, the short-term benefit is emphasized more,
Figure 448197DEST_PATH_IMAGE041
indicates this iteration to select a new action
Figure 112528DEST_PATH_IMAGE042
With prime mover
Figure 994902DEST_PATH_IMAGE043
A return gain formed therebetween, wherein
Figure 502107DEST_PATH_IMAGE044
The learning rate represents the speed of reinforcement learning, and the closer to 1 represents the faster learning, and the slower learning is vice versa; the whole formula
Figure 980493DEST_PATH_IMAGE045
The representative continuously updates each action taken in each state s by iteratively calculating the return gain
Figure 279756DEST_PATH_IMAGE046
Long term benefits that can be obtained
Figure 880501DEST_PATH_IMAGE047
Thereby enabling the system to autonomously select the optimal action by learning.
CN202111195085.XA 2021-10-14 2021-10-14 Virtual network hierarchical distributed deployment method and system based on reinforcement learning Active CN113641462B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111195085.XA CN113641462B (en) 2021-10-14 2021-10-14 Virtual network hierarchical distributed deployment method and system based on reinforcement learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111195085.XA CN113641462B (en) 2021-10-14 2021-10-14 Virtual network hierarchical distributed deployment method and system based on reinforcement learning

Publications (2)

Publication Number Publication Date
CN113641462A CN113641462A (en) 2021-11-12
CN113641462B true CN113641462B (en) 2021-12-21

Family

ID=78426774

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111195085.XA Active CN113641462B (en) 2021-10-14 2021-10-14 Virtual network hierarchical distributed deployment method and system based on reinforcement learning

Country Status (1)

Country Link
CN (1) CN113641462B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114827783B (en) * 2022-07-01 2022-10-14 西南民族大学 Aggregation tree-based bandwidth scheduling method for cross-domain distributed machine learning

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106411749A (en) * 2016-10-12 2017-02-15 国网江苏省电力公司苏州供电公司 Path selection method for software defined network based on Q learning
CN109947567A (en) * 2019-03-14 2019-06-28 深圳先进技术研究院 A kind of multiple agent intensified learning dispatching method, system and electronic equipment
CN110022230A (en) * 2019-03-14 2019-07-16 北京邮电大学 The parallel dispositions method of service chaining and device based on deeply study
CN110365514A (en) * 2019-05-24 2019-10-22 北京邮电大学 SDN multistage mapping method of virtual network and device based on intensified learning
CN110365568A (en) * 2019-06-18 2019-10-22 西安交通大学 A kind of mapping method of virtual network based on deeply study
CN110995619A (en) * 2019-10-17 2020-04-10 北京邮电大学 Service quality aware virtual network mapping method and device
CN111147307A (en) * 2019-12-30 2020-05-12 重庆邮电大学 Service function chain reliable deployment method based on deep reinforcement learning
JP2020127182A (en) * 2019-02-06 2020-08-20 日本電信電話株式会社 Control device, control method, and program

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106411749A (en) * 2016-10-12 2017-02-15 国网江苏省电力公司苏州供电公司 Path selection method for software defined network based on Q learning
JP2020127182A (en) * 2019-02-06 2020-08-20 日本電信電話株式会社 Control device, control method, and program
CN109947567A (en) * 2019-03-14 2019-06-28 深圳先进技术研究院 A kind of multiple agent intensified learning dispatching method, system and electronic equipment
CN110022230A (en) * 2019-03-14 2019-07-16 北京邮电大学 The parallel dispositions method of service chaining and device based on deeply study
CN110365514A (en) * 2019-05-24 2019-10-22 北京邮电大学 SDN multistage mapping method of virtual network and device based on intensified learning
CN110365568A (en) * 2019-06-18 2019-10-22 西安交通大学 A kind of mapping method of virtual network based on deeply study
CN110995619A (en) * 2019-10-17 2020-04-10 北京邮电大学 Service quality aware virtual network mapping method and device
CN111147307A (en) * 2019-12-30 2020-05-12 重庆邮电大学 Service function chain reliable deployment method based on deep reinforcement learning

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Low-Latency and Resource-Efficient Service Function Chaining Orchestration in Network Function Virtualization;Gang Sun等;《 IEEE Internet of Things Journal》;20190823;第7卷(第7期);第5760-5772页 *
MUVINE: Multi-Stage Virtual Network Embedding in Cloud Data Centers Using Reinforcement Learning-Based Predictions;Hiren Kumar Thakkar等;《 IEEE Journal on Selected Areas in Communications》;20200408;第38卷(第6期);第1058-1074页 *
Optimizing NFV Chain Deployment in Software-Defined Cellular Core;Jiaqi Zheng等;《IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS》;20191231;第38卷(第2期);第248-262页 *
基于强化学习和QoS感知的虚拟网络映射算法设计与实现;李蒙;《中国优秀硕士学位论文全文数据库 信息科技辑》;20210515;I139-8 *
基于改进深度强化学习的虚拟网络功能部署优化算法;唐伦等;《电子与信息学报》;20210615;第43卷(第6期);第1724-1732页 *
基于时变资源的容器化虚拟网络映射算法;邓伟健等;《https://kns.cnki.net/kcms/detail/51.1307.TP.20210628.1335.017.html》;20210628;第1-9页 *

Also Published As

Publication number Publication date
CN113641462A (en) 2021-11-12

Similar Documents

Publication Publication Date Title
CN111147307B (en) Service function chain reliable deployment method based on deep reinforcement learning
US10025892B2 (en) Simulation systems and methods
US10878146B2 (en) Handover techniques for simulation systems and methods
CN112600717B (en) Satellite network management and control protocol semi-physical test device based on SDN
EP2629490A1 (en) Optimizing traffic load in a communications network
CN114050961B (en) Large-scale network simulation system and resource dynamic scheduling and distributing method
CN113341712B (en) Intelligent hierarchical control selection method for unmanned aerial vehicle autonomous control system
CN105515987A (en) SDN framework based virtual optical network oriented mapping method
CN113641462B (en) Virtual network hierarchical distributed deployment method and system based on reinforcement learning
Lu et al. A cluster-tree-based energy-efficient routing protocol for wireless sensor networks with a mobile sink
Bouzidi et al. Dynamic clustering of software defined network switches and controller placement using deep reinforcement learning
WO2024077881A1 (en) Scheduling method and system for neural network training, and computer-readable storage medium
CN115329985B (en) Unmanned cluster intelligent model training method and device and electronic equipment
WO2023179180A1 (en) Network virtualization system structure and virtualization method
WO2023089350A1 (en) An architecture for a self-adaptive computation management in edge cloud
CN117707795B (en) Graph-based model partitioning side collaborative reasoning method and system
Afrasiabi et al. Reinforcement learning-based optimization framework for application component migration in NFV cloud-fog environments
Tyagi et al. GM-WOA: a hybrid energy efficient cluster routing technique for SDN-enabled WSNs
CN115879543A (en) Model training method, device, equipment, medium and system
CN106709597A (en) Parallel TSP problem optimizing method and device based on artificial bee colony algorithm
Shooshtarian et al. A maximally robustness embedding algorithm in virtual data centers with multi-attribute node ranking based on TOPSIS
Adewale Adaptive and Scalable Controller Placement in Software-Defined Networking
CN112234599A (en) Advanced dynamic self-adaptive partitioning method and system for multi-element complex urban power grid
Sun et al. Wireless sensor network path optimization based on hybrid algorithm
Steffenel et al. A framework for adaptive collective communications for heterogeneous hierarchical computing systems

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant