CN113037794B - Method, device and system for computing resource allocation scheduling - Google Patents

Method, device and system for computing resource allocation scheduling Download PDF

Info

Publication number
CN113037794B
CN113037794B CN201911355811.2A CN201911355811A CN113037794B CN 113037794 B CN113037794 B CN 113037794B CN 201911355811 A CN201911355811 A CN 201911355811A CN 113037794 B CN113037794 B CN 113037794B
Authority
CN
China
Prior art keywords
application
node
resource
proxy service
event
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911355811.2A
Other languages
Chinese (zh)
Other versions
CN113037794A (en
Inventor
刘虢
余万水
黄浩
刘洪政
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mashang Xiaofei Finance Co Ltd
Original Assignee
Mashang Xiaofei Finance Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mashang Xiaofei Finance Co Ltd filed Critical Mashang Xiaofei Finance Co Ltd
Priority to CN201911355811.2A priority Critical patent/CN113037794B/en
Publication of CN113037794A publication Critical patent/CN113037794A/en
Application granted granted Critical
Publication of CN113037794B publication Critical patent/CN113037794B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
    • H04L67/025Protocols based on web technology, e.g. hypertext transfer protocol [HTTP] for remote control or remote monitoring of applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/06Protocols specially adapted for file transfer, e.g. file transfer protocol [FTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/2866Architectures; Arrangements
    • H04L67/30Profiles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/56Provisioning of proxy services
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/60Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The embodiment of the invention provides a method, a device and a system for allocating and scheduling computing resources, wherein the method comprises the following steps: the resource manager judges whether a computing resource configuration event for modifying the application is monitored; if the event of modifying the computing resource configuration of the application is monitored, the resource manager calls a proxy service on a node so as to modify the computing resource configuration of the application through the proxy service; the resource manager acquires a modification result returned by the proxy service for modifying the computing resource configuration of the application; the resource manager operates the data storage system according to the modification result to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof, so that the re-allocation of the computing resources is realized under the condition of not restarting the application example, and the use experience of a user is not influenced.

Description

Computing resource allocation scheduling method, device and system
Technical Field
The embodiment of the invention relates to the technical field of computers, in particular to a method, a device and a system for computing resource configuration scheduling.
Background
Referring to fig. 1, a Vertical Pod auto scaler (VPA) scheme of kubernets (a container cluster management system) is illustrated, which inherits a basic flow of kubernets cluster management that an application resource definition change triggers redeployment (restart).
The method comprises the steps that a recommender in a VPA controller obtains resource usage information from a cluster index server, continuously calculates a recommended value and sends the recommended value to a VPA access controller, the access controller judges whether a condition for triggering and modifying resource definition is met or not according to VPA configuration, the condition for triggering and modifying the resource definition is judged to be met and then the original resource definition is covered, and an updater in the VPA controller monitors that an application instance is changed and restarted to enable configuration change to take effect.
Implementation of the VPA scheme of Kubernetes requires a restart of the application instance. Restarting an application instance may have an impact on the user using the application, for example, restarting an application instance in a production environment may cause a reduction in quality of service, and a test environment may affect the work under test and affect the test results.
Disclosure of Invention
An object of the embodiments of the present invention is to provide a method, an apparatus, and a system for scheduling computing resource allocation, which solve the problem that modifying application computing resource allocation by using a container cluster management system will cause restarting of an application instance.
In a first aspect, an embodiment of the present invention provides a method for computing resource configuration scheduling, which is applied to a resource manager, and includes:
judging whether a computing resource configuration event of an application is monitored to be modified;
if the event of modifying the computing resource configuration of the application is monitored, calling a proxy service on a node to modify the computing resource configuration of the application through the proxy service;
obtaining a modification result returned by the proxy service for modifying the computing resource configuration of the application;
and operating the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof.
Optionally, if the event of modifying the computing resource configuration of the application is monitored, invoking a proxy service on a node to modify the computing resource configuration of the application through the proxy service, including:
determining the name of each first node deployed by the application instance if a computing resource configuration event modifying the application is monitored;
counting currently available computing resources for scheduling on each first node through an application program interface server API server;
if the current available resource of each first node is larger than the added value of the resource request, the proxy service on each first node is called, and the group resource limit value is controlled on each first node by the proxy service modification application instance container.
Optionally, if the event of modifying the computing resource configuration of the application is monitored, invoking a proxy service on a node to modify the computing resource configuration of the application through the proxy service includes:
determining a type of a computing resource configuration event that modifies the application if the event is heard;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event limited by the computing resources of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources currently available for scheduling on each node through the API server;
if the current resource available for scheduling of each node is larger than the added value of the resource request, the proxy service on each node is called, and the group resource limit value is controlled on each node by the application instance container modified through the proxy service.
Optionally, the operating the data storage system according to the modification result to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof includes:
and calling a key-value-pair data storage system interface according to the modification result of the computing resource configuration of the application, and modifying the value of the resource request quantity of the application in the key-value-pair data storage system so as to keep the control group resource limit value in each node consistent with the resource limit value stored in the key-value-pair data storage system.
In a second aspect, an embodiment of the present invention provides a resource manager, including:
the judging module is used for judging whether a computing resource configuration event of the modified application is monitored or not;
the calling module is used for calling the proxy service on the node to modify the computing resource configuration of the application through the proxy service if the computing resource configuration event of the application is monitored to be modified;
an obtaining module, configured to obtain a modification result returned by the proxy service for modifying the computing resource configuration of the application;
and the synchronous updating module is used for operating the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result.
Optionally, the invoking module is further configured to:
determining the name of each first node deployed by the application instance if a computing resource configuration event modifying the application is monitored;
counting currently available computing resources for scheduling on each first node through an API server;
if the current resource available for scheduling of each first node is larger than the added value of the resource request, invoking the proxy service on each first node, and modifying the control group resource limit value of the application instance container on each first node through the proxy service.
Optionally, the calling module is further configured to:
determining a type of a computing resource configuration event that modifies the application if the event is heard;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event of the computing resource limit of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in a cluster, and counting the computing resources which can be currently scheduled on each node through the API server;
if the current resource available for scheduling of each node is larger than the added value of the resource request, then the proxy service on each node is invoked, and the proxy service modifies the group resource limit value controlled by the application instance container on each node.
Optionally, the synchronization update module is further configured to: and calling a key-value-pair data storage system interface according to the modification result of the computing resource configuration of the application, and modifying the value of the resource request quantity of the application in the key-value-pair data storage system so as to keep the control group resource limit value in each node consistent with the resource limit value stored in the key-value-pair data storage system.
In a third aspect, an embodiment of the present invention provides a computing resource allocation scheduling system, including: the system comprises a monitoring system, a control system, a container scheduling platform and a resource manager:
the monitoring system sends a request message to a management and control system, wherein the request message is used for requesting to modify the computing resource configuration of the application;
the management and control system informs an API server of the container scheduling platform to modify the computing resource configuration of the application according to the request message;
if the resource manager monitors an event for modifying the computing resource configuration of the application, the resource manager calls proxy service on a node and modifies the computing resource configuration of the application through the proxy service;
and the resource manager operates the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof, wherein the data storage system is used for storing the information of the container scheduling platform.
Optionally, the resource manager determines a name of each first node deployed by the application instance;
the resource manager counts the computing resources currently available for scheduling on each first node through the API server;
if the current available resource for scheduling of each first node is greater than the added value of the resource request, the resource manager invokes a proxy service on each first node, which modifies the group resource limit value controlled by the application instance container on each first node.
Optionally, the proxy service determines process identifiers of the containers running on the nodes according to the identifiers of the applications;
and finding a path corresponding to the corresponding process in the control group through the process identifier, and modifying the content of the control file of the corresponding control group under the path so as to enable the changed resource limit value of the control group to take effect on each node.
Optionally, the resource manager calls a key-value-pair data storage system interface according to a modification result of the computing resource configuration of the application, and modifies a value of a resource request amount of the application in the key-value-pair data storage system, so that the control group resource limit value in each node is consistent with the resource limit value stored in the key-value-pair data storage system.
Optionally, if the resource manager listens for a computing resource configuration event that modifies the application, determining the type of the event;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event of the computing resource limit of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in a cluster, and counting the computing resources which can be currently scheduled on each node through the API server;
if the current resource available for scheduling of each node is greater than the added value of the resource request, the resource manager invokes a proxy service on each node that modifies the group resource limit value controlled by the application instance container on each node.
In a fourth aspect, an embodiment of the present invention provides a computer-readable storage medium, on which a computer program is stored, where the computer program, when executed by a processor, implements steps including the method for scheduling a computing resource configuration according to the first aspect.
In the embodiment of the invention, the resource limit configuration modification of the control group is realized by calling the proxy service on the node through the resource manager, and simultaneously, the data is written into the data storage system for cluster information storage through the interface, so that the re-allocation of the computing resources is realized under the condition of not restarting the application example, and the use experience of a user is not influenced.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the invention. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
FIG. 1 is a schematic diagram of the longitudinal expansion of a prior Kubernets;
FIG. 2 is a flowchart of a computing resource allocation scheduling method according to an embodiment of the present invention;
FIG. 3 is a second flowchart of a computing resource allocation scheduling method according to an embodiment of the present invention;
FIG. 4 is a diagram illustrating a computing resource allocation schedule according to an embodiment of the present invention;
FIG. 5 is a second illustration of a computing resource allocation schedule according to an embodiment of the present invention;
FIG. 6 is a diagram illustrating a computing resource allocation scheduling system according to an embodiment of the present invention;
FIG. 7 is a diagram of a resource manager according to an embodiment of the invention.
Detailed Description
To facilitate understanding of embodiments of the present invention, the following technical terms are introduced.
(1) Kubernetes: the cross-host cluster open-source container scheduling platform can automatically deploy, expand and operate application containers and provides a container-centered infrastructure.
(2) Kubernets CRD (Custom Resource Definition): the Application Programming Interface (API) is expanded by self-defining the secondary development capability of the resources, a new resource type can be added into the Kubernets API through the CRD without modifying a Kubernets source code to create a self-defined API server (server), and the expansion capability of the Kubernets is greatly improved by the function.
(3) API Server: the kubernets API Server provides HyperText Transfer Protocol (HTTP) representation State Transfer (Rest) interfaces such as addition, deletion, modification, query and monitoring (watch) of various resource objects (e.g., pod, etc.) of kubernets, and the interfaces are data buses and data centers of the whole cluster.
(4) Pod: the smallest unit that can be created and managed in the kubernets system is the smallest resource object model created or deployed by a user in the resource object model, and is also a resource object for running containerized applications on kubernets, other resource objects are all used for supporting or extending the functions of Pod objects, such as a controller object for managing Pod objects, a Service or Ingress resource object for exposing Pod reference objects, a PersistentVolume resource object for providing storage for pods, and the like, k8s (an open development platform) does not directly process the pods, and the pods can directly process the pods, the pods are composed of one or more containers, and each Pod can set a quota of computing resources including CPUs and memories.
(5) Cgroups: the control groups (control groups) are a function of the Linux kernel, and are used to limit, control and separate resources (such as CPU, memory, and disk input/output) of a process group.
(6) A container: (container) a portable, lightweight operating system level virtualization technique. It uses namespace (namespace) to isolate different software running environments and by mirroring the running environment of self-contained software, so that the container can be conveniently run anywhere. Compared with the traditional virtual machine technology, because the virtual operating system does not exist, the occupied resource is smaller, and the running efficiency is higher.
(7) Etcd: a highly available, distributed key-value pair data storage system. Kubernetes uses etcd as a back-end store to store state information of all network configurations and objects of the cluster.
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be obtained by a person skilled in the art without inventive step based on the embodiments of the present invention, are within the scope of protection of the present invention.
The terms "comprises," "comprising," or any other variation thereof, in the description and claims of this application, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus. Furthermore, the use of "and/or" in the specification and claims means that at least one of the connected objects, such as a and/or B, means that three cases, a alone, B alone, and both a and B, exist.
In the embodiments of the present invention, words such as "exemplary" or "for example" are used to mean serving as examples, illustrations or descriptions. Any embodiment or design described as "exemplary" or "e.g.," an embodiment of the present invention is not necessarily to be construed as preferred or advantageous over other embodiments or designs. Rather, use of the word "exemplary" or "such as" is intended to present concepts related in a concrete fashion.
Referring to fig. 2, an embodiment of the present invention provides a method for scheduling computing resource allocation, where the method includes: step 201, step 202, step 203 and step 204.
Step 201: the monitoring system sends a request message to the management and control system, wherein the request message is used for requesting to modify the computing resource configuration of the application;
in the embodiment of the present invention, the monitoring system is configured to monitor whether it is necessary to trigger modification of the computing resource configuration of the application, for example, when the monitoring system adds a corresponding configuration, and when a certain index exceeds a threshold, the management and control system is invoked to request modification of the computing resource of the application instance container.
In the embodiment of the present invention, the management and control system is configured to manage (for example, increase or decrease a resource limit) a computing resource on a node (which is a smallest computing hardware unit in the container cluster management system, and may be a physical machine or a virtual machine, for example).
For example, if a monitoring index (for example, the usage rate of a CPU or a memory, etc.) configured in the monitoring system exceeds a threshold, the request message is sent to the management and control system.
In the embodiment of the present invention, the request message may be used to request to decrease the computing resource configuration of the application, or may also request to increase the computing resource configuration of the application.
Step 202: the management and control system informs an application program interface server (API server) of the container scheduling platform to modify the computing resource configuration of the application according to the request message;
in the embodiment of the present invention, the container scheduling platform is used for managing containerized applications on multiple hosts in a cloud platform, for example, the container scheduling platform may be kubernets, but is not limited thereto.
Step 203: if the resource manager monitors the event of modifying the computing resource configuration of the application, the proxy service on the node is called, and the computing resource configuration of the application is modified through the proxy service;
in the embodiment of the present invention, the resource manager may be a custom resource manager, which is used for scheduling cluster resources, and the resource manager may count, through an API server interface, computing resources (for example, CPU resources or memory resources) currently available for scheduling on the application instance deployment node, for example, the resource manager may be a kubernets CRD, but is not limited thereto.
Step 204: and the resource manager operates the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof.
In this embodiment of the present invention, the data storage system is configured to store information of the container scheduling platform, for example, the information may include information (e.g., a limit value) related to a computing resource of the application instance deployment node, for example, the data storage system may be Etcd, but is not limited thereto.
In some embodiments, if a resource manager hears, through the API server, a computing resource configuration event that modifies the application, the resource manager determines a name of each first node deployed by the application instance; the resource manager counts the computing resources currently available for scheduling on each first node through the API server; if the current available resource for scheduling of each first node is greater than the added value of the resource request, the resource manager invokes a proxy service on each first node, which modifies the group resource limit value controlled by the application instance container on each first node.
In some embodiments, if a resource manager hears, through the API server, a computing resource configuration event that modifies the application, then determining a type of the event;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event limited by the computing resources of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources currently available for scheduling on each node through the API server;
if the currently available resources for scheduling of each of the nodes is greater than the added value of the resource request, the resource manager invokes a proxy service on each of the nodes that modifies the group resource limit value controlled on the node by the application instance container.
It is understood that the resource currently available for scheduling by the node is the amount of allocable resource on the node minus the amount of resource applied by each instance of all deployed applications.
In some embodiments, said proxy service modifying application instance container controls a group resource limit value at each of said nodes, comprising: the proxy service determines process identifiers of the containers running on the nodes according to the identifiers of the applications; and finding a path corresponding to the corresponding process in the control group through the process identifier, and modifying the content of the control file of the corresponding control group under the path so as to enable the changed resource limit value of the control group to take effect on each node.
In some embodiments, the resource manager performs a synchronous update on the data storage system according to a modification result of the computing resource configuration of the application, including: the resource manager calls a key-value-pair data storage system interface according to a modification result of the computing resource configuration of the application, and modifies values of resource request amounts of the application in the key-value-pair data storage system (for example, modifies resource limit values of a CPU (central processing unit) and a memory occupied by the application), so that a control group resource limit value in the node is consistent with a resource limit value stored in the key-value-pair data storage system. Therefore, the resource limit of the CPU and the memory occupied by the application can be automatically modified. For a single application, automatic longitudinal scalability is achieved. For the whole cluster, the utilization rate of the computing resources of the whole cluster can be effectively improved.
In the embodiment of the invention, the resource limit configuration modification of the control group is realized by calling the proxy service on the node through the resource manager, and simultaneously, the data is written into the data storage system for cluster information storage through the interface, so that the transparent reallocation of the computing resources is realized under the condition of not restarting the application example, and the use experience of a user is not influenced.
Referring to fig. 3, an embodiment of the present invention provides a method for scheduling computing resource allocation, where an execution subject of the method is a resource manager, and the method includes: step 301, step 302, step 303 and step 304.
Step 301: judging whether a computing resource configuration event for modifying the application is monitored, and executing a step 302 if the computing resource configuration event for modifying the application is monitored; otherwise, the flow ends.
Step 302: invoking a proxy service on a node to modify a computing resource configuration of the application through the proxy service;
step 303: obtaining a modification result returned by the proxy service for modifying the computing resource configuration of the application;
step 304: and operating the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof.
In some embodiments, in step 302, if a computing resource configuration event that modifies the application is monitored, a name of each first node deployed by the application instance is determined;
counting the computing resources currently available for scheduling on each first node through an application program interface server (API server);
if the current available resource of each first node is larger than the added value of the resource request, the proxy service on each first node is called, and the group resource limit value is controlled on each first node by the proxy service modification application instance container.
In some embodiments, in step 302, if a computing resource configuration event that modifies the application is monitored, a type of the event is determined;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event of the computing resource limit of the application, confirming a node where each instance of the application is located according to the deployment condition of the application in a cluster, and counting the computing resources which can be currently scheduled on each node through the APIserver;
if the current resource available for scheduling of each node is larger than the added value of the resource request, the proxy service on each node is called, and the group resource limit value is controlled on each node by the application instance container modified through the proxy service.
In some embodiments, in step 304, a key-value pair data storage system interface is invoked according to a modification result of the computing resource configuration of the application, and a value of a resource request amount of the application in the key-value pair data storage system is modified, so that a control group resource limit value in each of the nodes and a resource limit value stored in the key-value pair data storage system are kept consistent.
In the embodiment of the invention, the resource limit configuration modification of the control group is realized by calling the proxy service on the node through the resource manager, and simultaneously, the data is written into the data storage system for cluster information storage through the interface, so that the transparent reallocation of the computing resources is realized under the condition of not restarting the application example, and the use experience of a user is not influenced.
Referring to fig. 4, when the computing resource of the application needs to be modified, a corresponding configuration may be added to the monitoring system, and when a certain index exceeds a threshold, the management and control system is called to request that the computing resource of the application instance container be modified.
And after receiving the request, the management and control system calls an API server of a container scheduling platform Kubernetes to inform modification.
The resource manager exists in the form of kubernets CRD and listens for the above modifications. After monitoring a resource application modification event, a resource manager confirms the names of nodes deployed by each application example from a resource definition, then counts the currently available scheduling computing resources (such as cpu or memory) on the computing nodes through an API server interface, and directly gives up if the available scheduling resources of any host in the nodes are smaller than the added value of the modified resource request; otherwise, a proxy service (or referred to as a proxy program) which is pre-installed on each node is called, and the corresponding cgroup resource limit value of the specified container on the node (or referred to as a physical machine) is modified through the proxy service.
If the modification operation of each node is successful, the resource manager calls an etcd interface to modify the value of the resource request quantity corresponding to the POD in the etcd, so that the cgroup resource limit value in the physical machine is consistent with the value stored in the etcd by the corresponding object in the cluster.
Referring to fig. 5, the trigger management and control system modifies the computing resource configuration: request/limit (limit) configuration supporting manual modification of the cpu and memory of the application; the minimum resource requirement of the container is used as the judgment of resource allocation in the container scheduling.
Regarding the request:
(1) The container is allowed to be scheduled to the node only when the resource quantity available on the current node > = request;
(2) The request parameter does not limit the maximum available resources of the container.
Regarding Limit:
(1) Maximum value of available resources of the container;
(2) The setting of "0" means that the resource to be used is not limited and can be used indefinitely.
Meanwhile, a monitoring notification item can be configured in the monitoring system, for example, when the average CPU utilization rate of a certain application in a period of time is lower than 1%, the monitoring notification item notifies the management and control system, and the value of request/limit of CPU in the original resource definition of the application is changed to be smaller. The management and control system can determine a recommended value according to the preset configuration and algorithm and inform the recommended value to the API Server of Kubernets.
And monitoring modification: a user-defined resource manager is realized in a Kubernets CRD mode, and the user-defined resource manager monitors the event of application computing resource change through an API Server of a Kubernets master control node.
Check feasibility of resource change: if the change is a reduction in resource constraints, go directly to the next step to modify. If the resource limit is increased, whether expansion room exists on each computing node needs to be verified firstly, so that the computing node where each instance of the application is located is confirmed according to the deployment situation of the application in the event in the cluster, the API Server is inquired, the resource quantity applied by all deployed Pod is subtracted from the node allocable resource quantity, and when the value is larger than the increment of the change, the expansion is feasible, and the next actual modification is allowed.
Modifying node (physical machine) configuration through proxy: the resource manager calls the proxy service on the node, the proxy service determines the process ID number (PID) of the container running on the node (physical machine) according to the application identifier, then finds the path corresponding to the process in the cgroup through the PID, and modifies the content of the corresponding cgroup control file under the path, so that the changed value takes effect on the node (physical machine).
Although the configuration is valid at this time, the kubernets cluster is unaware of the change and needs to continuously modify the data of the relevant pod in the cluster object information store to match the actual situation with the metadata of the cluster management.
Synchronization to etcd is modified: and when the change of each computing node in the previous step is successful, the user-defined resource manager modifies the metadata information of all the Pod of the current change through the etcd interface. Thereby making the validated configuration consistent with the cluster-managed metadata to avoid subsequent negative impact on cluster scheduling.
By combining the monitoring system with the predefined configuration and algorithm, the resource limit of CPU and memory occupied by the application can be automatically modified. For a single application, automatic longitudinal scalability is achieved. For the whole cluster, the utilization rate of the computing resources of the whole cluster can be effectively improved.
The method avoids the basic process that the original application resource definition is modified by Kubernetes to cause restarting, realizes the resource limitation configuration modification of the control group by directly calling a user-defined proxy on a physical machine through a user-defined resource manager, and simultaneously writes data into a data storage system for cluster information storage through an interface, realizes the transparent re-allocation of computing resources under the condition of not restarting an application example, and does not influence the use experience of a user.
Referring to fig. 6, an embodiment of the present invention further provides a computing resource configuration scheduling system (or referred to as a cloud computing platform), including: monitoring system 601, management and control system 602, container scheduling platform 603, resource manager 604, data storage system 605 and node 606, optionally, management and control system 602 and resource manager 604 may be deployed on a cloud management and control system;
the monitoring system 601 sends a request message to the management and control system 602, where the request message is used to request to modify the computing resource configuration of the application;
the management and control system 602 notifies an API server of the container scheduling platform 603 to modify the computing resource configuration of the application according to the request message;
if the resource manager 604 monitors an event for modifying the computing resource configuration of the application, invoking a proxy service on a node 606, and modifying the computing resource configuration of the application through the proxy service;
the resource manager 604 operates the data storage system according to the modification result to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof, and the data storage system 605 is used for storing the information of the container scheduling platform.
In some embodiments, the resource manager 604 determines the name of each first node of the application instance deployment; the resource manager 604 counts the currently available computing resources for scheduling on each first node through the API server; if the currently available resources for scheduling at each of the first nodes is greater than the incremental value of the resource request, the resource manager 604 invokes a proxy service at each of the first nodes that modifies the group resource limit value controlled at each of the first nodes by the application instance container.
In some embodiments, the proxy service determines process identities for containers running on each of the nodes based on the identity of the application; and finding a path corresponding to the corresponding process in the control group through the process identifier, and modifying the content of the control file of the corresponding control group under the path so as to enable the changed resource limit value of the control group to take effect on each node.
In some embodiments, the resource manager 604 invokes a key-value pair data storage system interface according to the modification result of the computing resource configuration of the application, and modifies the value of the resource request amount of the application in the key-value pair data storage system, so that the control group resource limit value in each of the nodes and the resource limit value stored in the key-value pair data storage system are consistent.
In some embodiments, if the resource manager 604 listens for a computing resource configuration event that modifies the application, then the type of the event is determined;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event limited by the computing resources of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources currently available for scheduling on each node through the API server;
if the currently available resources for scheduling for each of the nodes is greater than the incremental value of the resource request, the resource manager 604 invokes a proxy service on each of the nodes that modifies the group resource limit value controlled by the application instance container on each of the nodes.
The computing resource allocation scheduling system provided in the embodiment of the present invention may implement the foregoing method embodiment, and the implementation principle and technical effect are similar, which are not described herein again.
Referring to fig. 7, an embodiment of the present invention further provides a resource manager, where the resource manager 700 includes:
a determining module 701, configured to determine whether a computing resource configuration event of an application is monitored;
a calling module 702, configured to, if a computing resource configuration event of the application is monitored to be modified, call a proxy service on a node to modify the computing resource configuration of the application through the proxy service;
an obtaining module 703, configured to obtain a modification result returned by the proxy service, where the modification result modifies the computing resource configuration of the application;
and a synchronous updating module 704, configured to operate the data storage system according to the modification result, so as to synchronously update and store the configuration scheduling information that is consistent with the current modification and the modification result.
In some embodiments, the invoking module 702 is further configured to:
determining the name of each first node deployed by the application instance if a computing resource configuration event modifying the application is monitored;
counting currently available computing resources for scheduling on each first node through an API server;
if the current resource available for scheduling of each first node is larger than the added value of the resource request, invoking the proxy service on each first node, and modifying the control group resource limit value of the application instance container on each first node through the proxy service.
In some embodiments, the invoking module 702 is further configured to:
determining a type of a computing resource configuration event that modifies the application if the event is heard;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event limited by the computing resources of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources currently available for scheduling on each node through the API server;
if the current resource available for scheduling of each node is larger than the added value of the resource request, then the proxy service on each node is invoked, and the proxy service modifies the group resource limit value controlled by the application instance container on each node.
In some embodiments, the synchronization update module 704 is further configured to: and calling a key-value-pair data storage system interface according to a modification result of the computing resource configuration of the application, and modifying the value of the resource request amount of the application in the key-value-pair data storage system so as to keep the control group resource limit value in each node consistent with the resource limit value stored in the key-value-pair data storage system.
The resource manager provided in the embodiment of the present invention may execute the method embodiments described above, and the implementation principle and the technical effect are similar, which are not described herein again.
The embodiment of the present invention further provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the computer program implements each process of the foregoing method for scheduling a computing resource allocation, and can achieve the same technical effect, and in order to avoid repetition, the details are not repeated here. The computer-readable storage medium may be a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk.
The steps of a method or algorithm described in connection with the disclosure herein may be embodied in hardware or may be embodied in software instructions executed by a processor. The software instructions may be comprised of corresponding software modules that may be stored in RAM, flash memory, ROM, EPROM, EEPROM, registers, hard disk, a removable disk, a compact disk, or any other form of storage medium known in the art. An exemplary storage medium is coupled to the processor such the processor can read information from, and write information to, the storage medium. Of course, the storage medium may also be integral to the processor. The processor and the storage medium may reside in an ASIC. In addition, the ASIC may be carried in a core network interface device. Of course, the processor and the storage medium may reside as discrete components in a core network interface device.
Those skilled in the art will recognize that, in one or more of the examples described above, the functions described in this invention may be implemented in hardware, software, firmware, or any combination thereof. When implemented in software, the functions may be stored on or transmitted over as one or more instructions or code on a computer-readable medium. Computer-readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one place to another. A storage media may be any available media that can be accessed by a general purpose or special purpose computer.
The above-mentioned embodiments, objects, technical solutions and advantages of the present invention are further described in detail, it should be understood that the above-mentioned embodiments are only examples of the present invention, and are not intended to limit the scope of the present invention, and any modifications, equivalent substitutions, improvements and the like made on the basis of the technical solutions of the present invention should be included in the scope of the present invention.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, embodiments of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, embodiments of the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
Embodiments of the present invention are described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It will be apparent to those skilled in the art that various changes and modifications may be made in the embodiments of the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the embodiments of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to encompass these modifications and variations.

Claims (14)

1. A computing resource allocation scheduling method is applied to a resource manager and is characterized by comprising the following steps:
judging whether a computing resource configuration event of an application is monitored to be modified;
if the computing resource configuration event of the application is monitored to be modified, proxy services on all nodes deployed by the application instance are called, the group resource limit value of the application instance container is controlled on all first nodes through the proxy services, or the group resource limit value of the application instance container is controlled on all the nodes through the proxy services;
obtaining a modification result returned by the proxy service for modifying the computing resource configuration of the application;
and operating the data storage system according to the modification result to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof.
2. The method of claim 1, wherein said invoking a proxy service on a node if it hears a computing resource configuration event that modifies the application, modifying a group resource limit value at each of the first nodes by the proxy service, comprises:
if the computing resource configuration event of the application is monitored to be modified, determining the name of each first node deployed by the application instance;
counting currently available computing resources for scheduling on each first node through an application program interface server API server;
if the current resource available for scheduling of each first node is larger than the added value of the resource request, invoking the proxy service on each first node, and modifying the control group resource limit value of the application instance container on each first node through the proxy service.
3. The method of claim 1, wherein said invoking a proxy service on a node if it hears a computing resource configuration event that modifies the application, modifying a group resource limit value on each of the nodes by the proxy service, comprises:
determining a type of a computing resource configuration event that modifies the application if the event is heard;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event of the computing resource limit of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources which can be currently scheduled on each node through an API server;
if the current resource available for scheduling of each node is larger than the added value of the resource request, the proxy service on each node is called, and the group resource limit value is controlled on each node by the application instance container modified through the proxy service.
4. The method according to claim 1, wherein the operating the data storage system according to the modification result to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof comprises:
and calling a key-value-pair data storage system interface according to the modification result of the computing resource configuration of the application, and modifying the value of the resource request quantity of the application in the key-value-pair data storage system so as to keep the control group resource limit value in each node consistent with the resource limit value stored in the key-value-pair data storage system.
5. A resource manager, comprising:
the judging module is used for judging whether a computing resource configuration event of the modified application is monitored;
a calling module, configured to, if a computing resource configuration event for modifying the application is monitored, call a proxy service on each node deployed by an application instance, modify a group resource restriction value on each first node by using the proxy service, or modify an application instance container by using the proxy service, and control a group resource restriction value on each node;
an obtaining module, configured to obtain a modification result returned by the proxy service for modifying the computing resource configuration of the application;
and the synchronous updating module is used for operating the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof.
6. The resource manager of claim 5, wherein the calling module is further configured to:
determining the name of each first node deployed by the application instance if a computing resource configuration event modifying the application is monitored;
counting the currently available computing resources for scheduling on each first node through an API server;
if the current available resource of each first node is larger than the added value of the resource request, the proxy service on each first node is called, and the group resource limit value is controlled on each first node by the proxy service modification application instance container.
7. The resource manager of claim 5, wherein the calling module is further configured to:
determining a type of a computing resource configuration event that modifies the application if the event is heard;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event limited by the computing resources of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources currently available for scheduling on each node through an API server;
if the current available resource for scheduling of each of the nodes is greater than the incremental value of the resource request, invoking a proxy service on each of the nodes, the proxy service modifying a control group resource limit value of an application instance container on each of the nodes.
8. The resource manager of claim 5, wherein the synchronization update module is further configured to: and calling a key-value-pair data storage system interface according to the modification result of the computing resource configuration of the application, and modifying the value of the resource request quantity of the application in the key-value-pair data storage system so as to keep the control group resource limit value in each node consistent with the resource limit value stored in the key-value-pair data storage system.
9. A computing resource allocation scheduling system, comprising: the system comprises a monitoring system, a control system, a container scheduling platform and a resource manager:
the monitoring system sends a request message to a management and control system, wherein the request message is used for requesting to modify the computing resource configuration of the application;
the management and control system informs an API server of the container scheduling platform to modify the computing resource configuration of the application according to the request message;
if the resource manager monitors a computing resource configuration event for modifying the application, invoking a proxy service on each node deployed by the application instance, modifying the group resource limit value on each first node by the proxy service, or modifying the group resource limit value on each node by the application instance container by the proxy service;
and the resource manager operates the data storage system according to the modification result so as to synchronously update and store the configuration scheduling information consistent with the current modification and the modification result thereof, wherein the data storage system is used for storing the information of the container scheduling platform.
10. The system of claim 9, wherein:
the resource manager determines the name of each first node deployed by the application instance;
the resource manager counts the currently available computing resources for scheduling on each first node through the API server;
if the current available resources for scheduling at each of the first nodes is greater than the incremental value of the resource request, the resource manager invokes a proxy service at each of the first nodes that modifies the application instance container to control the group resource limit value at each of the first nodes.
11. The system of claim 10, wherein:
the proxy service determines process identifiers of the containers running on the nodes according to the identifiers of the applications;
and finding a path corresponding to the corresponding process in the control group through the process identifier, and modifying the content of the control file of the corresponding control group under the path so as to enable the changed resource limit value of the control group to take effect on each node.
12. The system of claim 11, wherein:
and the resource manager calls a key-value-pair data storage system interface according to the modification result of the computing resource configuration of the application, and modifies the value of the resource request quantity of the application in the key-value-pair data storage system, so that the resource limit value of the control group in each node is consistent with the resource limit value stored in the key-value-pair data storage system.
13. The system of claim 12, wherein:
if the resource manager monitors a computing resource configuration event for modifying the application, determining the type of the event;
if the type of the event is the reduction event of the computing resource limit of the application, calling proxy service on each node, and modifying the computing resource configuration of the application through the proxy service;
if the type of the event is an increasing event limited by the computing resources of the application, confirming the node where each instance of the application is located according to the deployment condition of the application in the cluster, and counting the computing resources currently available for scheduling on each node through the API server;
if the current resource available for scheduling of each node is greater than the added value of the resource request, the resource manager invokes a proxy service on each node that modifies the group resource limit value controlled by the application instance container on each node.
14. A computer-readable storage medium, having stored thereon a computer program which, when being executed by a processor, carries out steps comprising the method for computing resource configuration scheduling according to any one of claims 1 to 4.
CN201911355811.2A 2019-12-25 2019-12-25 Method, device and system for computing resource allocation scheduling Active CN113037794B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911355811.2A CN113037794B (en) 2019-12-25 2019-12-25 Method, device and system for computing resource allocation scheduling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911355811.2A CN113037794B (en) 2019-12-25 2019-12-25 Method, device and system for computing resource allocation scheduling

Publications (2)

Publication Number Publication Date
CN113037794A CN113037794A (en) 2021-06-25
CN113037794B true CN113037794B (en) 2023-04-18

Family

ID=76458074

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911355811.2A Active CN113037794B (en) 2019-12-25 2019-12-25 Method, device and system for computing resource allocation scheduling

Country Status (1)

Country Link
CN (1) CN113037794B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113821306B (en) * 2021-09-24 2023-09-05 亿咖通(湖北)技术有限公司 Resource isolation method, device and terminal equipment
CN116204268A (en) * 2021-11-30 2023-06-02 华为技术有限公司 Expansion and contraction method of cloud instance and related equipment thereof
CN114666102B (en) * 2022-03-03 2024-02-06 华云数据控股集团有限公司 Resource licensing method and system
CN115145690B (en) * 2022-06-30 2023-08-04 小米汽车科技有限公司 Cluster resource expansion method and device and container cluster management system
CN116302339A (en) * 2023-03-09 2023-06-23 上海道客网络科技有限公司 Container group in-situ expansion and contraction method and system based on container cloud platform
CN117032950A (en) * 2023-07-10 2023-11-10 企迈科技有限公司 Real-time data transparent transmission method and system based on log
CN116974703B (en) * 2023-09-22 2024-01-02 山东未来网络研究院(紫金山实验室工业互联网创新应用基地) Kubernetes application resource management method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103618644A (en) * 2013-11-26 2014-03-05 曙光信息产业股份有限公司 Distributed monitoring system based on hadoop cluster and method thereof
CN105765556A (en) * 2013-11-25 2016-07-13 亚马逊科技公司 Customer-directed networking limits in distributed systems
CN106790565A (en) * 2016-12-27 2017-05-31 中国电子科技集团公司第五十二研究所 A kind of network attached storage group system
CN109739640A (en) * 2018-12-13 2019-05-10 北京计算机技术及应用研究所 A kind of container resource management system based on Shen prestige framework
CN110417870A (en) * 2019-07-05 2019-11-05 北京达佳互联信息技术有限公司 configuration file management method, device, equipment and storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7035943B2 (en) * 1998-05-29 2006-04-25 Yahoo! Inc. Web server content replication
CN106603631B (en) * 2016-11-17 2020-07-03 北京奇艺世纪科技有限公司 Distributed message system and capacity expansion method
CN110099021B (en) * 2018-01-30 2021-10-26 成都华为技术有限公司 Synchronization signal configuration method and device
CN108829509A (en) * 2018-05-03 2018-11-16 山东汇贸电子口岸有限公司 Distributed container cluster framework resources management method based on domestic CPU and operating system
US10084826B1 (en) * 2018-05-14 2018-09-25 Xage Security, Inc. Protocol agnostic security by using out-of-band health check

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105765556A (en) * 2013-11-25 2016-07-13 亚马逊科技公司 Customer-directed networking limits in distributed systems
CN103618644A (en) * 2013-11-26 2014-03-05 曙光信息产业股份有限公司 Distributed monitoring system based on hadoop cluster and method thereof
CN106790565A (en) * 2016-12-27 2017-05-31 中国电子科技集团公司第五十二研究所 A kind of network attached storage group system
CN109739640A (en) * 2018-12-13 2019-05-10 北京计算机技术及应用研究所 A kind of container resource management system based on Shen prestige framework
CN110417870A (en) * 2019-07-05 2019-11-05 北京达佳互联信息技术有限公司 configuration file management method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN113037794A (en) 2021-06-25

Similar Documents

Publication Publication Date Title
CN113037794B (en) Method, device and system for computing resource allocation scheduling
US11704144B2 (en) Creating virtual machine groups based on request
US11553034B2 (en) Server computer management system for supporting highly available virtual desktops of multiple different tenants
US20170031622A1 (en) Methods for allocating storage cluster hardware resources and devices thereof
WO2017186123A1 (en) System and method for distributed resource management
CN111880936B (en) Resource scheduling method, device, container cluster, computer equipment and storage medium
EP3442201B1 (en) Cloud platform construction method and cloud platform
CN111427675B (en) Data processing method and device and computer readable storage medium
CN109412875B (en) zookeeper cluster automatic maintenance method and device
CN113382077B (en) Micro-service scheduling method, micro-service scheduling device, computer equipment and storage medium
US20190056942A1 (en) Method and apparatus for hardware acceleration in heterogeneous distributed computing
CN113204353B (en) Big data platform assembly deployment method and device
CN110569104A (en) Management method and computer storage medium for task training in deep learning system
CN112532758B (en) Method, device and medium for establishing network edge computing system
CN109005071B (en) Decision deployment method and scheduling equipment
CN110308914B (en) Upgrade processing method, device, equipment, system and computer readable storage medium
CN107547622B (en) Resource adjusting method and device
CN115225645A (en) Service updating method, device, system and storage medium
CN114629958A (en) Resource allocation method, device, electronic equipment and storage medium
CN110968406B (en) Method, device, storage medium and processor for processing task
US11687269B2 (en) Determining data copy resources
CN114257505B (en) Server node configuration method, device, equipment and storage medium
CN113873052B (en) Domain name resolution method, device and equipment of Kubernetes cluster
CN114584625B (en) Message processing method and device, electronic equipment and storage medium
CN117435302A (en) Container capacity adjustment method, device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant