CN115052033B - Resource sharing-based micro-service effective containerized deployment method for intelligent factory - Google Patents

Resource sharing-based micro-service effective containerized deployment method for intelligent factory Download PDF

Info

Publication number
CN115052033B
CN115052033B CN202210619201.4A CN202210619201A CN115052033B CN 115052033 B CN115052033 B CN 115052033B CN 202210619201 A CN202210619201 A CN 202210619201A CN 115052033 B CN115052033 B CN 115052033B
Authority
CN
China
Prior art keywords
micro
service
server
services
deployed
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210619201.4A
Other languages
Chinese (zh)
Other versions
CN115052033A (en
Inventor
刘宇翔
杨博
吴雨
郑忠斌
陈彩莲
关新平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Original Assignee
Shanghai Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University filed Critical Shanghai Jiaotong University
Priority to CN202210619201.4A priority Critical patent/CN115052033B/en
Publication of CN115052033A publication Critical patent/CN115052033A/en
Application granted granted Critical
Publication of CN115052033B publication Critical patent/CN115052033B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1095Replication or mirroring of data, e.g. scheduling or transport for data synchronisation between network nodes
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

The invention discloses an effective containerized deployment method of micro-services based on resource sharing in an intelligent factory, which relates to the field of intelligent manufacturing. The invention can achieve the purpose of reducing mirror image pull delay and communication overhead and improving micro-service response efficiency through the micro-service deployment scheme based on resource sharing.

Description

Resource sharing-based micro-service effective containerized deployment method for intelligent factory
Technical Field
The invention relates to the field of intelligent manufacturing, in particular to an effective containerized deployment method of micro-services based on resource sharing in an intelligent factory.
Background
With the rapid development of intelligent manufacturing and flexible production, the flexibility of industrial production is greatly enhanced. Therefore, industrial software is required to rapidly redistribute and adjust production procedures according to changes of orders, and higher requirements are put on flexibility and expansibility of the industrial software. The traditional industrial software adopts a single-body type service architecture, the high coupling and high occupancy rate in the service enable the whole system to have higher complexity, and the expansibility, stability and fault tolerance of the whole system are difficult to meet the requirements of intelligent manufacturing, so that the industrial software architecture based on the micro-service is widely paid attention to. Through the micro-service architecture, a complete service can be split into a plurality of loosely coupled micro-services according to service and function requirements, and the loosely coupled micro-services cooperate with each other to complete production tasks. The logic between different micro services is independent, so that the method has high flexibility, expandability and fault tolerance, can well adapt to the requirements of intelligent manufacturing, and meets the requirements of production customization.
In intelligent manufacturing, a large number of computationally intensive production tasks exist, and in order to meet the high requirements of the computationally intensive production tasks on instantaneity and service efficiency, an edge-computing-oriented micro-service platform is emerging. Edge computing is a novel computing paradigm, and by providing low-latency computing services through small edge servers deployed near the device, computing resources of terminal devices, edge nodes, and cloud servers can be fully utilized. Currently, container technology represented by Docker and the container orchestration tool Kubernetes developed by google are becoming the dominant solution for micro-service deployment and maintenance on edge platforms. Through the container orchestration tool, each micro-service may be packaged into a Docker mirror, deployed to edge servers according to service requests and deployment policies.
In the process of micro-service containerized deployment, how to improve service efficiency so that deployed micro-services can be quickly started, run and obtain calculation results is a critical problem. The service efficiency is mainly influenced by two aspects, namely, the starting time of the micro service is too long, so that the service response speed is slow; and secondly, communication overhead among micro services, transmission of input data and calculation results can cause transmission delay, so that service efficiency is affected.
The start-up time of the micro-service is mainly dependent on the pull delay of the Docker mirror. Dock images generally include runtime tools, system tools, and system dependencies, stored in the cloud through different image layers. When a service needs to be provided locally, the edge server first pulls the non-local container image containing all necessary layers from the cloud and deploys it on the edge server. Because of the limited network bandwidth, image pulling can create a corresponding downlink delay that depends on the size of the image pulled and the link bandwidth. An integrated study of microservice images showed that at 100Mbps bandwidth, the average startup time of the individual images was about 20.7 seconds, while the average image pull delay was about 15.8 seconds, accounting for 76.6% of the average startup time. Mirror pull latency has become a non-negligible factor affecting the container startup time, and thus the efficiency of service response.
On the other hand, the communication overhead of the micro services depends on the communication data volume between the micro services. Because of the continuity of production, industrial applications may be co-ordinated by multiple micro-services deployed on one or more edge servers. For example, the equipment failure detection (Equipment Failure Detection, EFD) function of semiconductor manufacturing may be implemented by six input-output interconnected micro-services including source data access, feature extraction, k-Nearest Neighbor (knN) cluster-based learning services, XGBoost, support vector machine, and result aggregation services. These micro services may be referred to as a micro service chain, and frequent information interaction may occur between micro services located on the same micro service chain. For example, in the equipment failure detection, the acquired source data needs to be transmitted to the feature extraction micro-service for data processing, and the processed feature vector file needs to be transmitted to the next micro-service for clustering. The large amount of data transmission between micro services can lead to high transmission delays, which in turn affect service response efficiency.
One effective way to optimize service efficiency is resource sharing. Although the functions of the micro services are different from each other, there is also commonality between the micro services. Through the repeated utilization of the common resources, the response efficiency and the service quality of the micro service can be improved.
One policy for resource sharing is layer sharing. Although production facilities (robotic arms, AGVs, cameras, etc.) are diverse, the same type of micro-services can be implemented based on the same base layer. For example, fault detection, face recognition and quality detection of micro-service images can be realized through a Python public base layer; the micro-service mirror image related to the production data storage and the query can be realized through the MySQL public base layer. Sharing of the Docker native support layer, if the micro services deployed on the same edge server use the same base layer, then the layer will be pulled only once without repeated pulling when mirror pulling is performed, and the layer can be shared by all the micro services. By sharing the same base layer, the delay of mirror pulling can be effectively reduced, thereby improving the starting speed and service response efficiency of micro-service.
Another policy for resource sharing is chain sharing. In the micro-service chain, frequent information transmission exists between two adjacent micro-services, and if the two micro-services are deployed on different edge servers, the information needs to be transmitted through multiple hops; if two micro-services are deployed on the same server, the information to be transmitted can be directly accessed by the next micro-service by sharing the same memory address, and multi-hop transmission of data is not needed, so that the problems of delay, packet loss and the like caused by data transmission are reduced.
In the above two resource sharing schemes, layer sharing deploys the same type of micro services on different micro service chains together so as to recycle the same base layer, and chain sharing tends to deploy the micro services on the same micro service chain together, so that data transmission is reduced through shared memory. However, because of the limited resources of the edge servers, the micro services cannot be deployed on the same edge server.
Aiming at the problem of resource sharing among micro services, the deployment of the micro services mainly faces the following two difficulties. 1) How to model the hierarchical structure of the micro-service mirror to accurately describe the relationship between the micro-service mirror and the container layer. 2) How to describe the chain structure of the micro-services and the amount of communication data between the micro-services. 3) How to simultaneously consider layer sharing and chain sharing to establish an optimization problem, and realize optimal resource sharing balance based on a corresponding solving method, thereby improving the efficiency of service response.
Accordingly, those skilled in the art are working to develop an optimal micro-service deployment strategy to make the trade-off between layer sharing and chain sharing.
Disclosure of Invention
In view of the above-mentioned drawbacks of the prior art, the present invention is directed to how to improve the service response rate in the production process. In order to achieve the above purpose, the invention provides an effective containerized deployment method of a micro-service based on resource sharing in an intelligent factory, which is characterized in that a mirror image pull delay model and a communication overhead model of a system are built based on a layered structure of the micro-service and a micro-service chain, and an optimization problem is built into an integer quadratic programming problem and solved by a business solver.
Further, the method comprises the steps of:
step 1, system modeling, which aims at an intelligent manufacturing system, and performs detailed system modeling to describe the relationship among a mirror layer, the micro service chain and service response efficiency;
step 2, constructing an integer quadratic programming problem to optimize mirror image pull delay and communication overhead;
step 3, solving by using a solver to obtain an optimal deployment strategy;
and 4, performing micro-service deployment by using the obtained deployment strategy.
Further, the step 1 further includes the following steps:
step 1.1, modeling of an edge server;
step 1.2, modeling of a micro-service mirror image and a mirror image layer;
step 1.3, modeling of data transmission and a multi-hop model;
and step 1.4, modeling the micro service chain.
Further, in the step 1.1, the intelligent manufacturing system includes
Figure BDA0003674459630000031
The edge server and a cloud server deployed at the far end are used for storing the micro-service mirror image; each edge server is provided with limited computing and storage resources, and a certain number of micro services can be deployed; the computing and storage resources of the edge server n are denoted +.>
Figure BDA0003674459630000032
and />
Figure BDA0003674459630000033
The bandwidth between the cloud server and the edge server n is +.>
Figure BDA0003674459630000034
Further, in the step 1.2, all the micro-service images are stored in a micro-service image registry of a cloud server, and pulled by the edge server according to the deployed micro-service; each complete micro service image consists of a shareable base layer and an unshared layer; all layers were combined with l= {1,2, L to show that, by means of the method,
Figure BDA0003674459630000035
indicate->
Figure BDA0003674459630000036
The size of the layer; thus, each micro-service can be composed of +.>
Figure BDA0003674459630000037
One or more layers of (C) may be composed of E kil E {0,1} represents ms ki Whether or not the first layer is included.
Further, in step 1.3, each server receives different service requests and data, if the server is deployed with the expected micro service, the service requests can be directly processed, and if the server is not deployed with the expected micro service, the requests and data need to be transmitted to another edge server with the expected micro service through multi-hops; because of different geographic positions among servers, the multi-hop times of communication among different servers also have differences, and D is defined nn′ For the number of hops to transmit a request or data from server n to server n', a matrix D is used to represent the multi-hop connections between all servers;
wherein ,
Figure BDA0003674459630000041
further, in the step 1.4, modeling the micro service chain as a directed weighted acyclic graph; every two micro services ms ki and mskj All use interactive weight
Figure BDA0003674459630000042
To represent the magnitude of the traffic between two micro-services; the interaction graph is written in the form of a matrix, which for application k is defined as
Figure BDA0003674459630000043
Wherein only on the micro-service chain
Figure BDA0003674459630000044
There are specific values for +.>
Figure BDA0003674459630000045
Based on the interaction graph, communication overhead can be calculated; if two adjacent micro services on one micro service chain are deployed on the same edge server, as the two micro services can share hardware resources, the data needing to be interacted can be directly accessed through a memory, and multi-hop data transmission among servers is not needed; if two micro services are deployed on different servers, then multi-hop transmission over the communication link between the servers is required.
Further, the step 2 further includes the following steps:
step 2.1, defining an optimization variable; to represent deployment of micro services, definition
Figure BDA0003674459630000046
Expressed in ms ki Is (are) deployment case>
Figure BDA0003674459630000047
Representing that the micro-service is deployed at the edge server n; because of the hierarchical structure of micro services, once a micro service is deployed at the edge server n, all layers contained by the micro service image are neededExist in the server n; with variable->
Figure BDA0003674459630000048
To represent download of micro-service layer, < >>
Figure BDA0003674459630000049
Indicating that layer i needs to be downloaded to the edge server n;
step 2.2, constructing an objective function
Figure BDA00036744596300000410
wherein />
Figure BDA00036744596300000411
S=[S 1 ,…,S L ] T ,/>
Figure BDA00036744596300000412
x=[(x 1 ) T ,…,(x K ) T ] T ,/>
Figure BDA00036744596300000413
Figure BDA00036744596300000414
The sum of the Hadamard products of the two matrices; md is the mirror pull delay, x, of the microservice T Wx is communication overhead among micro services, θ is a weight factor, and is used for balancing download delay and communication data volume;
step 2.3, constructing a micro-service deployment constraint qx=b, which ensures that each micro-service can only be deployed on one server and must be deployed, wherein
Figure BDA00036744596300000415
q=[1,1,...,1] 1×N ,/>
Figure BDA0003674459630000051
Step 2.4, constructing between two optimization variablesConstraint; because of the existence of layer sharing, if micro services deployed on the same server can share the same base layer, the base layer only needs to be downloaded once; thus, the first and second substrates are bonded together,
Figure BDA0003674459630000052
and />
Figure BDA0003674459630000053
The following constraints need to be met: />
Figure BDA0003674459630000054
Wherein Z is an arbitrarily large constant greater than 1, y= [ Y ] 1 ,…,Y N ] T
Figure BDA0003674459630000055
wherein
Figure BDA0003674459630000056
The n-th bit is 1->
Figure BDA0003674459630000057
With a first position of 1
Figure BDA0003674459630000058
E ki =[E ki1 ,…,E kiL ] T
Step 2.5, constructing resource constraint; the storage space occupied by all the micro-service mirror layers cannot exceed the maximum storage of the server, and the constraint is constructed as follows
Fd≤C S, wherein
Figure BDA0003674459630000059
S=[S 1 ,…,S L ] T ,/>
Figure BDA00036744596300000510
The computing resources of all microservice requests cannot exceed the maximum computing power of the server, and the constraint is built as
Gx≤C C Wherein g= [ G 1 ,…,G N ] T
Figure BDA00036744596300000511
Figure BDA00036744596300000512
Step 2.6, constructing variable constraints, wherein the optimization variables are binary variables, and the constraints are constructed as follows
Figure BDA00036744596300000513
Further, in step 3, the expression of the optimization problem is obtained through the step 2 as follows
Figure BDA00036744596300000514
s.t.Qx=b
Figure BDA00036744596300000515
Fd≤C S
Gx≤C C
Figure BDA00036744596300000516
The problem is an integer quadratic programming problem and is directly input into a commercial solver Gurobi to obtain a numerical solution.
Further, in the step 4, the numerical solution of x and d obtained in the step 3 is used to obtain the micro-service deployment strategy for balancing the pulling delay and the communication overhead of the mirror image optimally, and the engineering deployment is performed next step.
The invention has the following technical effects:
1. the invention simultaneously considers the layered structure of the micro-service mirror image and the chained structure among the micro-service, and establishes the integer quadratic programming problem of minimizing mirror image pulling delay and communication overhead, thereby obtaining the optimal micro-service deployment scheme to improve service response efficiency
2. The similar technology is to use an iterative approximation algorithm, solve a plurality of optimization problems, or use algorithms such as a neural network and reinforcement learning, and the like, has a complex structure, and can obtain a deployment scheme by only obtaining parameters related to a system structure and constructing a parameter matrix to solve an integer quadratic programming problem.
The conception, specific structure, and technical effects of the present invention will be further described with reference to the accompanying drawings to fully understand the objects, features, and effects of the present invention.
Drawings
FIG. 1 is a flowchart of an algorithm of a preferred embodiment of the present invention.
Detailed Description
The following description of the preferred embodiments of the present invention refers to the accompanying drawings, which make the technical contents thereof more clear and easy to understand. The present invention may be embodied in many different forms of embodiments and the scope of the present invention is not limited to only the embodiments described herein.
In the drawings, like structural elements are referred to by like reference numerals and components having similar structure or function are referred to by like reference numerals. The dimensions and thickness of each component shown in the drawings are arbitrarily shown, and the present invention is not limited to the dimensions and thickness of each component. The thickness of the components is exaggerated in some places in the drawings for clarity of illustration.
As shown in fig. 1, in order to optimize service response efficiency in a production process, the present invention attempts to propose a resource sharing micro-service deployment scheme for an edge computing scenario. And establishing a mirror image pull delay model and a communication overhead model of the system based on the layered structure of the micro service and the micro service chain, converting the optimization problem into an integer quadratic programming problem through model reconstruction, and solving through a business solver. The final micro-service deployment scheme can achieve the purpose of reducing mirror image pull delay and communication overhead so as to improve the micro-service response efficiency.
The resource sharing micro-service deployment scheme for the edge computing scene comprises the following steps:
step one: system modeling, detailed system modeling is performed for intelligent manufacturing systems to describe the relationship between the mirror layer, the micro-service chain, and the service response efficiency. The method mainly comprises the following steps:
s1: modeling of edge servers. The intelligent manufacturing system comprises
Figure BDA0003674459630000061
The system comprises a platform edge server and a cloud server deployed at a remote end for storing micro-service images. Each edge server has limited computing and storage resources, and a certain number of micro services can be deployed. The computing and storage resources of edge server n are denoted +.>
Figure BDA0003674459630000062
and />
Figure BDA0003674459630000063
The bandwidth between the cloud server and the edge server n is +.>
Figure BDA0003674459630000064
S2: modeling of micro-service mirroring and mirroring layers. All the micro-service images are stored in a micro-service image registry of the cloud server, and pulled by the edge server according to the deployed micro-service. Each complete micro-service image consists of some shareable base layers and some unshared layers. The different layers are of different sizes, we use all layers
Figure BDA0003674459630000071
To indicate (I)>
Figure BDA0003674459630000072
Indicate->
Figure BDA0003674459630000073
Layer size. Thus, each micro-service can be composed of +.>
Figure BDA0003674459630000074
One or more layers of (C) may be composed of E kil E {0,1} represents ms ki Whether or not the first layer is included.
S3: modeling of data transfer and multi-hop models. Each server receives different service requests and data, if the server is deployed with the expected micro service, the service requests can be directly processed, and if the server is not deployed with the expected micro service, the requests and the data need to be transmitted to another edge server with the expected micro service through multiple hops. Because of different geographic locations between servers, there is also a difference in the number of hops communicated between different servers, we define D nn′ To transmit a request or data from server n to server n', the number of hops may be obtained through the shortest communication path between the two servers. Obviously D nn′ =D n′n ,D nn =0, we can represent the multi-hop connection between all servers with one matrix D.
Figure BDA0003674459630000075
And S4, modeling a micro service chain. In general, a micro-service chain can be modeled as a directed acyclic graph. However, because of communication interactions between micro services, to more clearly show the impact of communication data on micro service deployment, the micro service chain may be modeled as a directionally weighted acyclic graph. Every two micro services ms ki and mskj All use interactive weight
Figure BDA0003674459630000076
To indicate the magnitude of traffic between two micro-services. For ease of understanding, the interaction graph may be written in the form of a matrix, which for application k is defined as
Figure BDA0003674459630000077
Wherein only on the micro-service chain
Figure BDA0003674459630000078
There are specific values for +.>
Figure BDA0003674459630000079
Based on the interaction map, a calculation of the communication overhead can be performed. If two adjacent micro services on one micro service chain are deployed on the same edge server, as the two micro services can share hardware resources, data needing to be interacted can be directly accessed through a memory, and multi-hop data transmission between the servers is not needed. If two micro services are deployed on different servers, then multi-hop transmission over the communication link between the servers is required.
Step two: and constructing an integer quadratic programming problem to optimize mirror image pull delay and communication overhead. The invention constructs the integer quadratic programming problem aiming at the system modeling completed in the step one and gives corresponding constraint conditions. The method mainly comprises the following steps:
s1: optimization variables are defined. To represent the deployment of micro-services, we define
Figure BDA00036744596300000710
Expressed in ms ki Is (are) deployment case>
Figure BDA00036744596300000711
Representing that the micro-service is deployed at edge server n. Because of the hierarchical structure of the micro-services, once a micro-service is deployed at edge server n, all layers contained by the micro-service imageNeeds to exist on the server n. Variable for us
Figure BDA00036744596300000712
To represent download of micro-service layer, < >>
Figure BDA00036744596300000713
Indicating that the first layer needs to be downloaded to the edge server n.
S2: construction of objective functions
Figure BDA0003674459630000081
wherein />
Figure BDA0003674459630000082
S=[S 1 ,…,S L ] T ,/>
Figure BDA0003674459630000083
x=[(x 1 ) T ,…,(x K ) T ] T ,/>
Figure BDA0003674459630000084
Figure BDA0003674459630000085
The sum is defined as the Hadamard product of the two matrices. Md is the mirror pull delay, x, of the microservice T Wx is the communication overhead between micro services, θ is the weight factor, used to trade-off between download delay and communication data volume.
S3: constructing a micro-service deployment constraint qx=b that ensures that each micro-service can only be deployed on one server and must be deployed, wherein
Figure BDA0003674459630000086
q=[1,1,...,1] 1×N
Figure BDA0003674459630000087
S4: constraints between two optimization variables are constructed. Storage due to layer sharingIn the event that micro services deployed on the same server can share the same base layer, the base layer need only be downloaded once. Thus, the first and second substrates are bonded together,
Figure BDA0003674459630000088
and />
Figure BDA0003674459630000089
The following constraints need to be met: />
Figure BDA00036744596300000810
Wherein Z is an arbitrarily large constant greater than 1, y= [ Y ] 1 ,…,Y N ] T
Figure BDA00036744596300000811
wherein
Figure BDA00036744596300000812
The n-th bit is 1
Figure BDA00036744596300000813
With a first position of 1
Figure BDA00036744596300000814
E ki =[E ki1 ,…,E kiL ] T
S5: and constructing resource constraints. The storage space occupied by all micro-service mirror layers cannot exceed the maximum storage of the server, and the constraint is constructed as follows
Fd≤C S, wherein
Figure BDA00036744596300000815
S=[S 1 ,…,S L ] T ,/>
Figure BDA00036744596300000816
The computing resources of all micro-service requests cannot exceed the maximum computing capacity of the server, and the constraint is constructed to Gx less than or equal to C C Wherein g= [ G 1 ,…,G N ] T
Figure BDA00036744596300000817
Figure BDA0003674459630000091
S6, constructing variable constraints, wherein the optimization variables are binary variables, and the constraints are constructed as follows
Figure BDA0003674459630000092
Step three: and solving by using a solver to obtain an optimal deployment strategy. The expression of the optimization problem can be obtained through the second step
Figure BDA0003674459630000093
s.t.Qx=b
Figure BDA0003674459630000094
Fd≤C S
Gx≤C C
Figure BDA0003674459630000095
The problem is an integer quadratic programming problem and can be directly input into a commercial solver Gurobi to obtain a numerical solution.
Step four: and performing micro-service deployment by using the obtained deployment strategy. And D, using the numerical solution of x and d obtained in the step three to obtain an optimal micro-service deployment strategy for balancing the mirror image pulling delay and the communication overhead, and carrying out engineering deployment.
The foregoing describes in detail preferred embodiments of the present invention. It should be understood that numerous modifications and variations can be made in accordance with the concepts of the invention without requiring creative effort by one of ordinary skill in the art. Therefore, all technical solutions which can be obtained by logic analysis, reasoning or limited experiments based on the prior art by the person skilled in the art according to the inventive concept shall be within the scope of protection defined by the claims.

Claims (2)

1. An intelligent factory resource sharing-based micro-service effective containerized deployment method is characterized in that a mirror image pull delay model and a communication overhead model of a system are established based on a layered structure of micro-services and a micro-service chain, an optimization problem is modeled as an integer quadratic programming problem, and the integer quadratic programming problem is solved through a business solver;
the method comprises the following steps:
step 1, system modeling, which aims at an intelligent manufacturing system, and performs detailed system modeling to describe the relationship among a mirror layer, the micro service chain and service response efficiency;
step 2, constructing an integer quadratic programming problem to optimize mirror image pull delay and communication overhead;
step 3, solving by using a solver to obtain an optimal deployment strategy;
step 4, performing micro-service deployment by using the obtained deployment strategy;
wherein, the step 1 further comprises the following steps:
step 1.1, modeling of an edge server;
step 1.2, modeling of a micro-service mirror image and a mirror image layer;
step 1.3, modeling of data transmission and a multi-hop model;
step 1.4, modeling the micro service chain;
in the step 1.1, the intelligent manufacturing system comprises
Figure FDA0004145423920000011
The edge server and a deploymentThe cloud server at the far end is used for storing the micro-service mirror image; each edge server is provided with limited computing and storage resources, and a certain number of micro services are deployed; the computing and storage resources of the edge server n are denoted +.>
Figure FDA0004145423920000012
and />
Figure FDA0004145423920000014
The bandwidth between the cloud server and the edge server n is +.>
Figure FDA0004145423920000013
In the step 1.2, all the micro-service images are stored in a micro-service image registry of a cloud server, and pulled by the edge server according to the deployed micro-service; each complete micro service image consists of a shareable base layer and an unshared layer; applying all layers to
Figure FDA0004145423920000015
To indicate (I)>
Figure FDA0004145423920000016
Representing the size of the L-th e L layer; thus, each microservice is composed of +.>
Figure FDA0004145423920000017
One or more layers of (E) kil E {0,1} represents ms ki Whether or not to include a first layer;
in step 1.3, each server receives different service requests and data, if the server is deployed with an expected micro-service, the service requests are directly processed, and if the server is not deployed with the expected micro-service, the requests and the data need to be transmitted to another edge server with the expected micro-service through multi-hops; because of the difference in geographic locations between serversThe number of multi-hops of communication with the server is different, and D is defined nn′ For the number of hops to transmit a request or data from server n to server n', a matrix D is used to represent the multi-hop connections between all servers;
wherein ,
Figure FDA0004145423920000021
in the step 1.4, modeling the micro service chain as a directed weighted acyclic graph; every two micro services ms ki and mskj All use interactive weight
Figure FDA0004145423920000022
To represent the magnitude of the traffic between two micro-services; the interaction map is written in the form of a matrix, which interaction matrix is defined as +.>
Figure FDA0004145423920000023
Wherein only on the micro-service chain
Figure FDA0004145423920000024
With specific values for not located on the micro-service chain
Figure FDA0004145423920000025
Based on the interaction diagram, calculating communication overhead; if two adjacent micro services on one micro service chain are deployed on the same edge server, as the two micro services share hardware resources, the data needing to be interacted are directly accessed through a memory, and multi-hop data transmission among the servers is not needed; if two micro services are deployed on different servers, then multi-hop transmission over the communication link between the servers is required;
the step 2 further comprises the following steps:
step 2.1, fixingSense optimization variables; to represent deployment of micro services, definition
Figure FDA0004145423920000026
Expressed in ms ki Is (are) deployment case>
Figure FDA0004145423920000027
Representing that the micro-service is deployed at the edge server n; because of the hierarchical structure of micro services, once a micro service is deployed at the edge server n, all layers contained by the micro service image need to exist at server n; by variables
Figure FDA0004145423920000028
To represent download of micro-service layer, < >>
Figure FDA0004145423920000029
Indicating that layer i needs to be downloaded to the edge server n;
step 2.2, constructing an objective function
Figure FDA00041454239200000210
wherein />
Figure FDA00041454239200000211
S=[S 1 ,…,S L ] T ,/>
Figure FDA00041454239200000212
x=[(x 1 ) T ,…,(x K ) T ] T ,/>
Figure FDA00041454239200000213
Figure FDA00041454239200000214
Figure FDA00041454239200000215
Figure FDA00041454239200000216
The sum of the Hadamard products of the two matrices; md is the mirror pull delay, x, of the microservice T Wx is communication overhead among micro services, θ is a weight factor, and is used for balancing download delay and communication data volume;
step 2.3, constructing a micro-service deployment constraint qx=b, which ensures that each micro-service can only be deployed on one server and must be deployed, wherein
Figure FDA00041454239200000217
Figure FDA0004145423920000031
Step 2.4, constructing constraints between two optimization variables; because of the existence of layer sharing, if the micro services deployed on the same server share the same base layer, the base layer only needs to be downloaded once; thus, the first and second substrates are bonded together,
Figure FDA0004145423920000032
and />
Figure FDA0004145423920000033
The following constraints need to be met: />
Figure FDA0004145423920000034
Wherein Z is an arbitrarily large constant greater than 1, y= [ Y ] 1 ,…,Y N ] T ,/>
Figure FDA0004145423920000035
wherein
Figure FDA0004145423920000036
The n-th bit is 1
Figure FDA0004145423920000037
With a first position of 1
Figure FDA0004145423920000038
E ki =[E ki1 ,...,E kiL ] T
Step 2.5, constructing resource constraint; the storage space occupied by all the micro-service mirror layers cannot exceed the maximum storage of the server, and the constraint is constructed as follows
Fd≤C S, wherein
Figure FDA0004145423920000039
S=[S 1 ,…,S L ] T ,/>
Figure FDA00041454239200000310
The computing resources of all microservice requests cannot exceed the maximum computing power of the server, and the constraint is built as
Gx≤C C Wherein g= [ G 1 ,…,G N ] T
Figure FDA00041454239200000311
Figure FDA00041454239200000312
Step 2.6, constructing variable constraints, wherein the optimization variables are binary variables, and the constraints are constructed as follows
Figure FDA00041454239200000313
In the step 3, the expression of the optimization problem is obtained through the step 2
Figure FDA00041454239200000314
s.t.Qx=b
Figure FDA00041454239200000315
Fd≤C S
Gx≤C C
Figure FDA00041454239200000316
The problem is an integer quadratic programming problem and is directly input into a commercial solver Gurobi to obtain a numerical solution.
2. The method for efficiently and containerized deployment of micro-services based on resource sharing in an intelligent factory according to claim 1, wherein in the step 4, the numerical solution of x and d obtained in the step 3 is used to obtain an optimal micro-service deployment strategy for balancing mirror image pull delay and communication overhead, and in the next step, engineering deployment is performed.
CN202210619201.4A 2022-06-01 2022-06-01 Resource sharing-based micro-service effective containerized deployment method for intelligent factory Active CN115052033B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210619201.4A CN115052033B (en) 2022-06-01 2022-06-01 Resource sharing-based micro-service effective containerized deployment method for intelligent factory

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210619201.4A CN115052033B (en) 2022-06-01 2022-06-01 Resource sharing-based micro-service effective containerized deployment method for intelligent factory

Publications (2)

Publication Number Publication Date
CN115052033A CN115052033A (en) 2022-09-13
CN115052033B true CN115052033B (en) 2023-04-28

Family

ID=83159807

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210619201.4A Active CN115052033B (en) 2022-06-01 2022-06-01 Resource sharing-based micro-service effective containerized deployment method for intelligent factory

Country Status (1)

Country Link
CN (1) CN115052033B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024086344A1 (en) * 2022-10-20 2024-04-25 Fisher-Rosemount Systems, Inc. Compute fabric functionalities for a process control or automation system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108388472A (en) * 2018-03-01 2018-08-10 吉林大学 A kind of elastic task scheduling system and method based on Docker clusters
CN110058950A (en) * 2019-04-17 2019-07-26 上海沄界信息科技有限公司 Distributed cloud computing method and equipment based on serverless backup framework

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10893100B2 (en) * 2015-03-12 2021-01-12 International Business Machines Corporation Providing agentless application performance monitoring (APM) to tenant applications by leveraging software-defined networking (SDN)
US11765225B2 (en) * 2019-03-18 2023-09-19 Reliance Jio Infocomm Limited Systems and methods for microservice execution load balancing in virtual distributed ledger networks
CN113542503B (en) * 2020-03-31 2022-07-15 华为技术有限公司 Method, electronic device and system for creating application shortcut
CN114338504B (en) * 2022-03-15 2022-07-08 武汉烽火凯卓科技有限公司 Micro-service deployment and routing method based on network edge system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108388472A (en) * 2018-03-01 2018-08-10 吉林大学 A kind of elastic task scheduling system and method based on Docker clusters
CN110058950A (en) * 2019-04-17 2019-07-26 上海沄界信息科技有限公司 Distributed cloud computing method and equipment based on serverless backup framework

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Almiani Muder.Resilient Back Propagation Neural Network Security Model For Containerized Cloud Computing.《Simulation Modelling Practice and Theory 》.2022,全文. *
刘为 ; .微服务架构及相应云平台解析.科教导刊(下旬).2017,(01),全文. *
陆志刚 ; 徐继伟 ; 黄涛 ; .基于分片复用的多版本容器镜像加载方法.软件学报.2020,(06),全文. *

Also Published As

Publication number Publication date
CN115052033A (en) 2022-09-13

Similar Documents

Publication Publication Date Title
CN115052033B (en) Resource sharing-based micro-service effective containerized deployment method for intelligent factory
Mouradian et al. Application component placement in NFV-based hybrid cloud/fog systems
CN113835899B (en) Data fusion method and device for distributed graph learning
CN105049353A (en) Method for configuring routing path of business and controller
Rikos et al. Optimal CPU scheduling in data centers via a finite-time distributed quantized coordination mechanism
CN111813539A (en) Edge computing resource allocation method based on priority and cooperation
Tran et al. Dependable control systems with Internet of Things
CN114465900B (en) Data sharing delay optimization method and device based on federal edge learning
Santos et al. Reinforcement learning for service function chain allocation in fog computing
Henna et al. Distributed and collaborative high-speed inference deep learning for mobile edge with topological dependencies
CN117118497A (en) Controller and gateway joint deployment method suitable for satellite-ground integrated network
CN116367190A (en) Digital twin function virtualization method for 6G mobile network
CN114024894B (en) Dynamic calculation method and system in software-defined heaven-earth integrated network
CN114710200B (en) Satellite network resource arrangement method and system based on reinforcement learning
CN116582407A (en) Containerized micro-service arrangement system and method based on deep reinforcement learning
CN112217652A (en) Network topology device and method based on central communication mode
Zhang et al. Accelerate deep learning in IoT: Human-interaction co-inference networking system for edge
CN115713009A (en) Dynamic aggregation federal learning method based on satellite and ground station connection density
CN114745386A (en) Neural network segmentation and unloading method under multi-user edge intelligent scene
CN114826820A (en) FC-AE-1553 network service scheduling method and equipment under centralized control
CN113949666A (en) Flow control method, device, equipment and system
CN113708982A (en) Service function chain deployment method and system based on group learning
CN112153147A (en) Method for placing chained service entities based on entity sharing in mobile edge environment
CN114090306B (en) Pluggable block chain layered consensus method, system, device and storage medium
CN113132435B (en) Distributed training network system with separated storage and service network and communication method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant