CN113268344A - Resource balancing method and system, first Pod node and resource proxy server - Google Patents

Resource balancing method and system, first Pod node and resource proxy server Download PDF

Info

Publication number
CN113268344A
CN113268344A CN202110543062.7A CN202110543062A CN113268344A CN 113268344 A CN113268344 A CN 113268344A CN 202110543062 A CN202110543062 A CN 202110543062A CN 113268344 A CN113268344 A CN 113268344A
Authority
CN
China
Prior art keywords
resource
pod node
proxy server
receiving
pod
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110543062.7A
Other languages
Chinese (zh)
Inventor
屈阳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China United Network Communications Group Co Ltd
Original Assignee
China United Network Communications Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China United Network Communications Group Co Ltd filed Critical China United Network Communications Group Co Ltd
Priority to CN202110543062.7A priority Critical patent/CN113268344A/en
Publication of CN113268344A publication Critical patent/CN113268344A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Hardware Redundancy (AREA)

Abstract

The application discloses a resource balancing method and system, a first Pod node and a resource proxy server, wherein the resource balancing method comprises the following steps: receiving a large-scale promotion service request sent by a client terminal; under the condition of the current running state, sending a resource capacity expansion request to a resource proxy server; receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion; and after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server. The resource tenna scheduling method and device improve the efficiency of resource tenna scheduling, and are low in cost, safe, fast and easy to implement.

Description

Resource balancing method and system, first Pod node and resource proxy server
Technical Field
The application relates to the technical field of internet, in particular to a resource balancing method and system, a first pod (print On demand) node and a resource proxy server.
Background
At present, with the change of consumption habits and network behaviors, the service scene of large-scale promotion is more complex, the applications related to large-scale promotion under different scenes are different, and each application needs enough resources when coping with unique peaks. The preparation of sufficient resources before large-scale promotion in different applications is naturally the period of easily passing large-scale promotion flood peaks, but the large-scale promotion flood peaks are short in time, a large amount of resource investment is in an idle state for a long time, and the method is not suitable for reasonable application of resources.
The traditional large-scale resource balancing method in a short time mainly adopts a resource vacation scheduling mode, idle resources after large-scale promotion of some applications are released, but when the resource scale is very large, the efficiency of capacity reduction and expansion is extremely low, the applications need to apply for the resources and start the applications, the time consumption is very long, and if the time interval between a plurality of large-scale promotion activities is very short, the time needed for vacation is very likely to exceed the time interval of large-scale promotion. Secondly, the largest problem of resource tenna-move scheduling is that the application needs cold start, the time needed by the cold start and the time needed by preheating after the cold start greatly influence the time of the whole resource tenna-move scheduling.
Disclosure of Invention
The application provides a resource balancing method and system, a first Pod node and a resource proxy server, which can improve the efficiency of resource tenuation scheduling.
A first aspect of the present application provides a resource balancing method, which is applied to a first Pod node, and the method includes:
receiving a large-scale promotion service request sent by a client terminal;
under the condition of the current running state, sending a resource capacity expansion request to a resource proxy server;
receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
and after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server.
In some exemplary embodiments, in the case of being currently in the keep-alive state, before the sending the resource expansion request to the resource proxy server, the method further includes: switching from the keep-alive state to the run state.
A second aspect of the present application provides a resource balancing method, applied to a resource proxy server, including:
receiving a resource capacity expansion request sent by a first Pod node;
determining a second Pod node in a keep-alive state at present, and sending a resource capacity reduction request to the second Pod node;
receiving a resource capacity expansion completion message sent by the second Pod node, and sending a resource capacity expansion response to the first Pod node;
and receiving a resource capacity expansion completion message sent by the first Pod node.
In some exemplary embodiments, the method further comprises:
updating the resource occupation condition of the second Pod node;
and updating the resource occupation condition of the first Pod node.
In some exemplary embodiments, the determining the second Pod node currently in the keep-alive state includes:
and determining a second Pod node in the keep-alive state at present according to the corresponding relation between the Pod nodes and the states.
A third aspect of the present application provides a first Pod node, including:
the first receiving module is used for receiving a large-scale promotion service request sent by a client terminal;
the first sending module is used for sending a resource capacity expansion request to the resource proxy server under the condition that the current running state is achieved;
the first receiving module is further configured to: receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
the first sending module is further configured to: and after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server.
In some exemplary embodiments, further comprising:
and the switching module is used for switching from the keep-alive state to the running state under the condition that the current keep-alive state is achieved.
A fourth aspect of the present application provides a resource proxy server, including:
the second receiving module is used for receiving a resource capacity expansion request sent by the first Pod node;
the determining module is used for determining the second Pod node in the keep-alive state;
a second sending module, configured to send a resource capacity reduction request to the second Pod node;
the second receiving module is further configured to: receiving a resource capacity reduction completion message sent by the second Pod node;
the second sending module is further configured to: sending a resource capacity expansion response to the first Pod node;
the second receiving module is further configured to: and receiving a resource capacity expansion completion message sent by the first Pod node.
In some exemplary embodiments, further comprising:
an updating module, configured to update a resource occupation situation of the second Pod node;
and updating the resource occupation condition of the first Pod node.
In some exemplary embodiments, the determining module is specifically configured to:
and determining a second Pod node in the keep-alive state at present according to the corresponding relation between the Pod nodes and the states.
A fifth aspect of the present application provides a resource balancing system, including:
a first Pod node to:
receiving a large-scale promotion service request sent by a client terminal;
under the condition of the current running state, sending a resource capacity expansion request to a resource proxy server;
receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server;
a resource proxy server to:
receiving a resource capacity expansion request sent by a first Pod node;
determining a second Pod node in a keep-alive state at present, and sending a resource capacity reduction request to the second Pod node;
receiving a resource capacity expansion completion message sent by the second Pod node, and sending a resource capacity expansion response to the first Pod node;
and receiving a resource capacity expansion completion message sent by the first Pod node.
This application has following advantage:
in the embodiment of the application, when a large-scale promotion service request sent by a client terminal is received, if the large-scale promotion service request is currently in a running state, a resource capacity expansion request is sent to the resource proxy server to directly realize resource capacity expansion, namely, the resource vacation and move scheduling can be realized without restarting application, so that the efficiency of the resource vacation and move scheduling is improved, and the method is low in cost, safe, fast and easy to implement.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description serve to explain the principles of the application and not to limit the application.
Fig. 1 is a flowchart of a resource balancing method applied to a first Pod node according to an embodiment of the present application;
fig. 2 is a schematic diagram of a Pod node state according to an embodiment of the present application;
fig. 3 is a flowchart of a resource balancing method applied to a resource proxy server according to another embodiment of the present application;
fig. 4 is a schematic structural component diagram of a first Pod node according to another embodiment of the present application;
fig. 5 is a schematic structural component diagram of a resource proxy server according to another embodiment of the present application;
fig. 6 is a schematic structural diagram of a resource balancing system according to another embodiment of the present application.
Detailed Description
The following detailed description of embodiments of the present application will be made with reference to the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating the present application, are given by way of illustration and explanation only, and are not intended to limit the present application.
As used in this disclosure, the term "and/or" includes any and all combinations of one or more of the associated listed items.
The terminology used in the present disclosure is for the purpose of describing particular embodiments only and is not intended to be limiting of the disclosure. As used in this disclosure, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise.
When the term "comprises/comprising" and/or "made of.. is used in this disclosure, the presence of the stated features, integers, steps, operations, elements, and/or components are specified, but does not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
Embodiments of the present disclosure may be described with reference to plan and/or cross-sectional views in light of idealized schematic illustrations of the present disclosure. Accordingly, the example illustrations can be modified in accordance with manufacturing techniques and/or tolerances.
Unless otherwise defined, all terms (including technical and scientific terms) used in this disclosure have the same meaning as commonly understood by one of ordinary skill in the art. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and the present disclosure, and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Fig. 1 is a flowchart of a resource balancing method applied to a first Pod node according to an embodiment of the present disclosure.
As shown in fig. 1, an embodiment of the present application provides a resource balancing method, which is applied to a first Pod node, and the method includes:
step 100, receiving a large-scale promotion service request sent by a client terminal.
Step 101, sending a resource capacity expansion request to a resource proxy server under the current running state.
In the embodiment of the application, service and basic implementation are decoupled, the Pod node is divided into the running state and the keep-alive state, the running state and the keep-alive state can be switched rapidly, and only a Java Virtual Machine (JVM) memory Swap and network flow ratio based switching are needed during switching.
In an embodiment of the present application, a method for splitting and decoupling functions of an application server based on traditional decentralized includes: redefining the boundary between the infrastructure and the application server, reasonably segmenting the service management capability and the logic capability in the application server, sinking the service management capability to the infrastructure, and setting the logic as a proxy, wherein different infrastructures are mutually called by a consumer and a provider in a proxy module. The aim of decentralization is achieved, and the scalability of the system is guaranteed; service management and service logic decoupling are realized, the service management and the service logic can be independently evolved without mutual interference, and the evolution flexibility of the whole architecture is improved; meanwhile, the service grid architecture reduces the invasion of business logic and reduces the complexity of multi-language support.
In some exemplary embodiments, the service governance capability may refer to a process of changing a relationship between different application servers into a high-cohesion, low-coupling state, for example, the simplest service governance capability is to implement mutual decoupling between different application servers, unify external interfaces, and prohibit direct access to internal data; logical capabilities may refer to calling relationships with other application servers.
In the embodiment of the application, when the Pod node is in the running state, the application server in the Pod node is in the full-speed running state, and can run at full speed by using sufficient resources to carry 100% of traffic; and when the Pod node is in the keep-alive state, the application server in the Pod node is in the low-speed running state, only the limited resource can be used for running at low speed, 1% of traffic is borne, and the remaining 99% of traffic is forwarded to the Pod node in the running state.
For example, as shown in fig. 2, the first resource control module and the first APP server access the first Pod node, share the same Network Namespace (Network Namespace), Central Processing Unit (CPU), and Memory (Memory), and the APP Remote Procedure Call (RPC) traffic and the message traffic of the first APP server are not directly external, but directly interact with the first resource control module, and the first resource control module directly interfaces with the request of the first APP client terminal and then forwards the request to the first APP server.
The second resource control module and the second APP server are accessed to the second Pod node, share the same Network Namespace, CPU and Memory, the App RPC flow and the message flow of the second APP server are not directly outward any more, but directly interact with the second resource control module, the second resource control module directly interfaces the request of the second APP client terminal, and then forwards the request to the second APP server.
The first Pod node is in a running state, and the first APPlication program (APP) server is in a full-speed running state, and can run at full speed by using sufficient resources, for example, the resources can be used to 4C8G, and carry 100% of traffic; the second Pod node is in a keep-alive state, the second APP server is in a low-speed running state, and only limited resources can be used for low-speed running, for example, the resources can use 1C2G to carry 1% of traffic, and the remaining 99% of traffic is forwarded to the first APP server for processing.
In the embodiment of the present application, the first APP server may refer to an APP server to be promoted on a large scale, and the second APP server may refer to an APP server waiting for promotion on a large scale.
In the embodiment of the application, a first resource control module and a second resource control module are both registered on a resource proxy server, and the resource proxy server sends heartbeat requests to the first resource control module and the second resource control module at regular time; the first resource control module returns the current resource use condition of the first Pod node or the current occupied resource amount to the resource proxy server; and the second resource control module returns the current resource use condition of the second Pod node or the current occupied resource amount to the resource proxy server.
And 102, receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion.
In some exemplary embodiments, the resource expansion response comprises: the resource amount of the second Pod node for capacity reduction, namely the difference between the resource amount occupied by the second Pod node before capacity reduction and the resource amount occupied by the second Pod node after capacity reduction.
Step 103, after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server.
In some exemplary embodiments, the resource expansion completion message includes: the resource amount currently occupied by the first Pod node and the current state of the first Pod node.
In some exemplary embodiments, in the case of being currently in the keep-alive state, before the sending the resource expansion request to the resource proxy server, the method further includes: switching from the keep-alive state to the run state.
In the embodiment of the application, when a large-scale promotion service request sent by a client terminal is received, if the large-scale promotion service request is currently in a running state, a resource capacity expansion request is sent to the resource proxy server to directly realize resource capacity expansion, namely, the resource vacation and move scheduling can be realized without restarting application, so that the efficiency of the resource vacation and move scheduling is improved, and the method is low in cost, safe, fast and easy to implement.
Fig. 3 is a flowchart of a resource balancing method applied to a resource proxy server according to another embodiment of the present application.
As shown in fig. 3, another embodiment of the present application provides a resource balancing method applied to a resource proxy server, where the method includes:
step 300, receiving a resource capacity expansion request sent by the first Pod node.
Step 301, determining a second Pod node currently in a keep-alive state, and sending a resource capacity reduction request to the second Pod node.
In some exemplary embodiments, the determining the second Pod node currently in the keep-alive state includes:
and determining a second Pod node in the keep-alive state at present according to the corresponding relation between the Pod nodes and the states.
Step 302, receiving a resource capacity expansion completion message sent by the second Pod node, and sending a resource capacity expansion response to the first Pod node.
In some exemplary embodiments, the resource capacity reduction completion message includes: the resource amount of the second Pod node for capacity reduction, namely the difference between the resource amount occupied by the second Pod node before capacity reduction and the resource amount occupied by the second Pod node after capacity reduction.
In some exemplary embodiments, the resource capacity reduction completion message further includes: the resource amount currently occupied by the second Pod node and the current state of the second Pod node.
In some exemplary embodiments, the resource expansion response comprises: the resource amount of the second Pod node for capacity reduction, namely the difference between the resource amount occupied by the second Pod node before capacity reduction and the resource amount occupied by the second Pod node after capacity reduction.
And step 303, receiving a resource expansion completion message sent by the first Pod node.
In some exemplary embodiments, the resource expansion completion message includes: the resource amount currently occupied by the first Pod node and the current state of the first Pod node.
In some exemplary embodiments, the method further comprises:
updating the resource occupation condition of the second Pod node;
and updating the resource occupation condition of the first Pod node.
In some exemplary embodiments, the resource occupancy of the second Pod node includes: the amount of resources currently occupied by the second Pod node; the resource occupation situation of the first Pod node comprises the following steps: the amount of resources currently occupied by the first Pod node.
The steps of the above methods are divided for clarity, and the implementation may be combined into one step or split some steps, and the steps are divided into multiple steps, so long as the same logical relationship is included, which are all within the protection scope of the present patent; it is within the scope of the patent to add insignificant modifications to the algorithms or processes or to introduce insignificant design changes to the core design without changing the algorithms or processes.
Fig. 4 is a schematic structural component diagram of a first Pod node according to another embodiment of the present application.
As shown in fig. 4, another embodiment of the present application provides a first Pod node, including:
a first receiving module 401, configured to receive a large-scale promotion service request sent by a client terminal;
a first sending module 402, configured to send a resource capacity expansion request to the resource proxy server when the current state is in an operating state;
the first receiving module 401 is further configured to: receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
the first sending module 402 is further configured to: and after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server.
In some exemplary embodiments, further comprising:
a switching module 403, configured to switch from the keep-alive state to the running state when the current state is in the keep-alive state.
The specific implementation process of the first Pod node is the same as the specific implementation process of the resource balancing method at the first Pod node side in the foregoing embodiment, and details are not repeated here.
Fig. 5 is a schematic structural diagram of a resource proxy server according to another embodiment of the present application.
As shown in fig. 5, another embodiment of the present application provides a resource proxy server, including:
a second receiving module 501, configured to receive a resource expansion request sent by a first Pod node;
a determining module 502, configured to determine a second Pod node currently in a keep-alive state;
a second sending module 503, configured to send a resource capacity reduction request to the second Pod node;
the second receiving module 501 is further configured to: receiving a resource capacity reduction completion message sent by the second Pod node;
the second sending module 503 is further configured to: sending a resource capacity expansion response to the first Pod node;
the second receiving module 501 is further configured to: and receiving a resource capacity expansion completion message sent by the first Pod node.
In some exemplary embodiments, further comprising:
an updating module 504, configured to update a resource occupation situation of the second Pod node;
and updating the resource occupation condition of the first Pod node.
In some exemplary embodiments, the determining module 502 is specifically configured to:
and determining a second Pod node in the keep-alive state at present according to the corresponding relation between the Pod nodes and the states.
The specific implementation process of the resource proxy server in the embodiment of the present application is the same as the specific implementation process of the resource balancing method on the resource proxy server side in the foregoing embodiment, and details are not repeated here.
Fig. 6 is a schematic structural diagram of a resource balancing system according to another embodiment of the present application.
As shown in fig. 6, another embodiment of the present application provides a resource balancing system, including:
a first Pod node 601 configured to:
receiving a large-scale promotion service request sent by a client terminal;
under the condition of the current running state, sending a resource capacity expansion request to a resource proxy server;
receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server;
a resource proxy server 602 for:
receiving a resource capacity expansion request sent by a first Pod node;
determining a second Pod node in a keep-alive state at present, and sending a resource capacity reduction request to the second Pod node;
receiving a resource capacity expansion completion message sent by the second Pod node, and sending a resource capacity expansion response to the first Pod node;
and receiving a resource capacity expansion completion message sent by the first Pod node.
In some exemplary embodiments, further comprising:
a second Pod node 603 configured to:
receiving a resource capacity reduction request sent by a resource proxy server;
carrying out resource capacity reduction;
and after the resource capacity reduction is completed, sending a resource capacity reduction completion message to the resource proxy server.
The specific implementation process of the resource balancing system is the same as that of the resource balancing method in the foregoing embodiment, and is not described here again.
Each module in the present embodiment is a logical module, and in practical applications, one logical unit may be one physical unit, may be a part of one physical unit, or may be implemented by a combination of a plurality of physical units. In addition, in order to highlight the innovative part of the present application, a unit that is not so closely related to solving the technical problem proposed by the present application is not introduced in the present embodiment, but it does not indicate that no other unit exists in the present embodiment.
The present embodiments also provide an electronic device, comprising one or more processors; the storage device stores one or more programs, and when the one or more programs are executed by the one or more processors, the one or more processors implement the resource balancing method provided in this embodiment, and in order to avoid repeated descriptions, detailed steps of the resource balancing method are not described herein again.
The present embodiment further provides a computer readable medium, on which a computer program is stored, where the computer program, when executed by a processor, implements the resource balancing method provided in the present embodiment, and in order to avoid repeated descriptions, specific steps of the resource balancing method are not described herein again.
It will be understood by those of ordinary skill in the art that all or some of the steps of the methods, systems, functional modules/units in the devices disclosed above may be implemented as software, firmware, hardware, and suitable combinations thereof. In a hardware implementation, the division between functional modules/units mentioned in the above description does not necessarily correspond to the division of physical components; for example, one physical component may have multiple functions, or one function or step may be performed by several physical components in cooperation. Some or all of the physical components may be implemented as software executed by a processor, such as a central processing unit, digital signal processor, or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit. Such software may be distributed on computer readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). The term computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data, as is well known to those of ordinary skill in the art. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by a computer. In addition, communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media as known to those skilled in the art.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. The term "comprising" is used to specify the presence of stated features, integers, steps, operations, elements, components, operations, components, or the components, and/components.
Those skilled in the art will appreciate that although some embodiments described herein include some features included in other embodiments instead of others, combinations of features of different embodiments are meant to be within the scope of the embodiments and form different embodiments.
It is to be understood that the above embodiments are merely exemplary embodiments that are employed to illustrate the principles of the present application, and that the present application is not limited thereto. It will be apparent to those skilled in the art that various changes and modifications can be made therein without departing from the spirit and scope of the application, and these changes and modifications are to be considered as the scope of the application.

Claims (11)

1. A resource balancing method is applied to a first Pod node, and comprises the following steps:
receiving a large-scale promotion service request sent by a client terminal;
under the condition of the current running state, sending a resource capacity expansion request to a resource proxy server;
receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
and after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server.
2. The resource balancing method according to claim 1, wherein before the sending the resource expansion request to the resource proxy server in the keep-alive state, the method further comprises: switching from the keep-alive state to the run state.
3. A resource balancing method is applied to a resource proxy server, and comprises the following steps:
receiving a resource capacity expansion request sent by a first Pod node;
determining a second Pod node in a keep-alive state at present, and sending a resource capacity reduction request to the second Pod node;
receiving a resource capacity expansion completion message sent by the second Pod node, and sending a resource capacity expansion response to the first Pod node;
and receiving a resource capacity expansion completion message sent by the first Pod node.
4. The method of resource balancing according to claim 3, further comprising:
updating the resource occupation condition of the second Pod node;
and updating the resource occupation condition of the first Pod node.
5. The resource balancing method of claim 3, wherein the determining the second Pod node currently in the keep-alive state comprises:
and determining a second Pod node in the keep-alive state at present according to the corresponding relation between the Pod nodes and the states.
6. A first Pod node, comprising:
the first receiving module is used for receiving a large-scale promotion service request sent by a client terminal;
the first sending module is used for sending a resource capacity expansion request to the resource proxy server under the condition that the current running state is achieved;
the first receiving module is further configured to: receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
the first sending module is further configured to: and after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server.
7. The first Pod node of claim 6, further comprising:
and the switching module is used for switching from the keep-alive state to the running state under the condition that the current keep-alive state is achieved.
8. A resource proxy server, comprising:
the second receiving module is used for receiving a resource capacity expansion request sent by the first Pod node;
the determining module is used for determining the second Pod node in the keep-alive state;
a second sending module, configured to send a resource capacity reduction request to the second Pod node;
the second receiving module is further configured to: receiving a resource capacity reduction completion message sent by the second Pod node;
the second sending module is further configured to: sending a resource capacity expansion response to the first Pod node;
the second receiving module is further configured to: and receiving a resource capacity expansion completion message sent by the first Pod node.
9. The resource proxy server of claim 8, further comprising:
an updating module, configured to update a resource occupation situation of the second Pod node;
and updating the resource occupation condition of the first Pod node.
10. The resource proxy server according to claim 8, wherein the determining module is specifically configured to:
and determining a second Pod node in the keep-alive state at present according to the corresponding relation between the Pod nodes and the states.
11. A resource balancing system, comprising:
a first Pod node to:
receiving a large-scale promotion service request sent by a client terminal;
under the condition of the current running state, sending a resource capacity expansion request to a resource proxy server;
receiving a resource capacity expansion response sent by the resource proxy server, and performing resource capacity expansion;
after the resource expansion is completed, sending a resource expansion completion message to the resource proxy server;
a resource proxy server to:
receiving a resource capacity expansion request sent by a first Pod node;
determining a second Pod node in a keep-alive state at present, and sending a resource capacity reduction request to the second Pod node;
receiving a resource capacity expansion completion message sent by the second Pod node, and sending a resource capacity expansion response to the first Pod node;
and receiving a resource capacity expansion completion message sent by the first Pod node.
CN202110543062.7A 2021-05-18 2021-05-18 Resource balancing method and system, first Pod node and resource proxy server Pending CN113268344A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110543062.7A CN113268344A (en) 2021-05-18 2021-05-18 Resource balancing method and system, first Pod node and resource proxy server

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110543062.7A CN113268344A (en) 2021-05-18 2021-05-18 Resource balancing method and system, first Pod node and resource proxy server

Publications (1)

Publication Number Publication Date
CN113268344A true CN113268344A (en) 2021-08-17

Family

ID=77231617

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110543062.7A Pending CN113268344A (en) 2021-05-18 2021-05-18 Resource balancing method and system, first Pod node and resource proxy server

Country Status (1)

Country Link
CN (1) CN113268344A (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012221273A (en) * 2011-04-11 2012-11-12 Internatl Business Mach Corp <Ibm> Method, system and program for dynamically assigning resource
CN104023058A (en) * 2014-06-03 2014-09-03 浙江宇视科技有限公司 Method and system for managing and allocating storage resources
CN108768877A (en) * 2018-07-20 2018-11-06 网宿科技股份有限公司 A kind of distribution method of burst flow, device and proxy server
CN110493891A (en) * 2018-05-14 2019-11-22 杭州海康威视系统技术有限公司 Resource allocation methods, apparatus and system
CN110535961A (en) * 2019-09-03 2019-12-03 中国联合网络通信集团有限公司 Resource acquiring method, device, electronic equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012221273A (en) * 2011-04-11 2012-11-12 Internatl Business Mach Corp <Ibm> Method, system and program for dynamically assigning resource
CN104023058A (en) * 2014-06-03 2014-09-03 浙江宇视科技有限公司 Method and system for managing and allocating storage resources
CN110493891A (en) * 2018-05-14 2019-11-22 杭州海康威视系统技术有限公司 Resource allocation methods, apparatus and system
CN108768877A (en) * 2018-07-20 2018-11-06 网宿科技股份有限公司 A kind of distribution method of burst flow, device and proxy server
CN110535961A (en) * 2019-09-03 2019-12-03 中国联合网络通信集团有限公司 Resource acquiring method, device, electronic equipment and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
尚小东;张煜;郭成昊;: "基于作战任务优先级的容器云弹性伸缩系统", 指挥信息系统与技术, no. 03, pages 40 - 47 *
金融级分布式架构: "Service Mesh落地负责人亲述:蚂蚁金服双十一四大考题", pages 1 - 11, Retrieved from the Internet <URL:《https://developer.aliyun.com/article/728131》> *

Similar Documents

Publication Publication Date Title
JP7130142B2 (en) Communication method, apparatus, entity and computer readable storage medium
CN111857873A (en) Method for realizing cloud native container network
CN113067850B (en) Cluster arrangement system under multi-cloud scene
CN110069346B (en) Method and device for sharing resources among multiple processes and electronic equipment
CN105786603B (en) Distributed high-concurrency service processing system and method
CN110968603B (en) Data access method and device
CN111064626B (en) Configuration updating method, device, server and readable storage medium
CN116489179A (en) Service processing system and service processing method
CN110290009B (en) Data scheduling method and device and computer readable storage medium
CN108111630B (en) Zookeeper cluster system and connection method and system thereof
CN112769671A (en) Message processing method, device and system
CN109388501B (en) Communication matching method, device, equipment and medium based on face recognition request
CN113268310B (en) Pod resource quota adjustment method and device, electronic equipment and storage medium
CN111913792B (en) Service processing method and device
CN113867915A (en) Task scheduling method, electronic device and storage medium
CN113268344A (en) Resource balancing method and system, first Pod node and resource proxy server
CN115134373A (en) Data synchronization method and device, storage medium and electronic equipment
CN114900449B (en) Resource information management method, system and device
CN115098254A (en) Method and system for triggering execution of subtask sequence and electronic equipment
CN113515458A (en) Method and system for reducing resource consumption of test environment based on Envoy plug-in
CN110413427B (en) Subscription data pulling method, device, equipment and storage medium
US10200301B1 (en) Logical control groups for distributed system resources
CN114979153B (en) Load balancing method, computer equipment and storage device
CN113098914B (en) Message bus system, message transmission method and device, and electronic equipment
CN117041980B (en) Network element management method and device, storage medium and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination