CN112929408A - Dynamic load balancing method and device - Google Patents

Dynamic load balancing method and device Download PDF

Info

Publication number
CN112929408A
CN112929408A CN202110071520.1A CN202110071520A CN112929408A CN 112929408 A CN112929408 A CN 112929408A CN 202110071520 A CN202110071520 A CN 202110071520A CN 112929408 A CN112929408 A CN 112929408A
Authority
CN
China
Prior art keywords
service node
load
node
parameter information
service request
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110071520.1A
Other languages
Chinese (zh)
Inventor
王波
李涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou Apas Digital Cloud Information Technology Co ltd
Original Assignee
Zhengzhou Apas Digital Cloud Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou Apas Digital Cloud Information Technology Co ltd filed Critical Zhengzhou Apas Digital Cloud Information Technology Co ltd
Priority to CN202110071520.1A priority Critical patent/CN112929408A/en
Publication of CN112929408A publication Critical patent/CN112929408A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1004Server selection for load balancing
    • H04L67/1008Server selection for load balancing based on parameters of servers, e.g. available memory or workload
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1029Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers using data related to the state of servers by a load balancer

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer And Data Communications (AREA)

Abstract

The embodiment of the application discloses a dynamic load balancing method and a dynamic load balancing device, which are used for solving the problem that the load balancing of a back-end service node is difficult to accurately and timely carry out in the prior art. The method comprises the following steps: responding to the received service request, and acquiring pre-configured configuration information corresponding to the service request; determining a first target service node corresponding to the service request according to the configuration information, and routing the service request to the first target service node; after the first target service node successfully receives the service request, determining the current load state corresponding to each back-end service node; and determining second load parameter information respectively corresponding to each back-end service node according to the current load state and a preset load balancing strategy, and dynamically adjusting first load parameter information respectively corresponding to each back-end service node in the configuration information into the second load parameter information. The technical scheme can accurately and timely carry out load balancing on the back-end service node.

Description

Dynamic load balancing method and device
Technical Field
The present application relates to the field of network communication technologies, and in particular, to a dynamic load balancing method and apparatus.
Background
In the prior art, when the loads of the backend service nodes are unbalanced, configuration information of each backend service node is generally adjusted in a manual intervention manner to achieve load balancing. The load imbalance of the rear-end service nodes can be determined only when the loads of part of the service nodes are high and the part of the service nodes are idle in the rear-end service nodes, so that manual adjustment is performed, the manual adjustment is easily affected by subjective factors, and the accuracy of the adjustment is difficult to ensure. Therefore, it is difficult to accurately and timely load balance the backend service nodes according to the prior art.
Disclosure of Invention
An object of the embodiments of the present application is to provide a dynamic load balancing method and apparatus, so as to solve the problem in the prior art that it is difficult to accurately and timely perform load balancing on a backend service node.
In order to solve the above technical problem, the embodiment of the present application is implemented as follows:
in one aspect, an embodiment of the present application provides a dynamic load balancing method, including:
responding to a received service request, and acquiring pre-configured configuration information corresponding to the service request; the configuration information comprises a plurality of back-end service nodes for processing the service request and first load parameter information corresponding to each back-end service node; the first load parameter information is used for representing the priority of the back-end service node for receiving the service request;
determining a first target service node corresponding to the service request according to the configuration information, and routing the service request to the first target service node;
after the first target service node successfully receives the service request, determining the current load state corresponding to each back-end service node;
and determining second load parameter information corresponding to each back-end service node according to the current load state and a preset load balancing strategy, and dynamically adjusting the first load parameter information corresponding to each back-end service node into the second load parameter information, so that the priority corresponding to each back-end service node is determined again based on the second load parameter information.
On the other hand, an embodiment of the present application provides a dynamic load balancing apparatus, including:
the acquisition module is used for responding to the received service request and acquiring pre-configured configuration information corresponding to the service request; the configuration information comprises a plurality of back-end service nodes for processing the service request and first load parameter information corresponding to each back-end service node; the first load parameter information is used for representing the priority of the back-end service node for receiving the service request;
the determining and executing module is used for determining a first target service node corresponding to the service request according to the configuration information and routing the service request to the first target service node;
a determining module, configured to determine, after the first target service node successfully receives the service request, current load states corresponding to the backend service nodes, respectively;
and the determining and adjusting module is used for determining second load parameter information corresponding to each rear-end service node according to the current load state and a preset load balancing strategy, and dynamically adjusting the first load parameter information corresponding to each rear-end service node into the second load parameter information, so that the priority corresponding to each rear-end service node is re-determined based on the second load parameter information.
In another aspect, an embodiment of the present application provides a dynamic load balancing apparatus, which includes a processor and a memory electrically connected to the processor, where the memory stores a computer program, and the processor is configured to call and execute the computer program from the memory to implement the above dynamic load balancing method.
In yet another aspect, an embodiment of the present application provides a storage medium for storing a computer program, where the computer program is executable by a processor to implement the above dynamic load balancing method.
By adopting the technical scheme of the embodiment of the application, the received service request is responded, the service request is routed to the first target service node according to the pre-configured configuration information corresponding to the service request, and after the first target service node successfully receives the service request, the first load parameter information corresponding to each rear-end service node is dynamically adjusted to the second load parameter information according to the current load state corresponding to each rear-end service node and the preset load balancing strategy, so that the priority corresponding to each rear-end service node is determined again based on the second load parameter information. Therefore, the technical scheme can dynamically adjust the load parameter information respectively corresponding to each back-end service node according to the current load state respectively corresponding to each back-end service node and by combining the preset load balancing strategy, so as to realize load balancing, and therefore, the load parameter information respectively corresponding to each back-end service node can be accurately and timely adjusted, namely, the priority of each back-end service node for receiving the service request is adjusted, so that the utilization rate of each back-end service node is improved, the received service request can be timely processed, and the processing efficiency of the service request is improved.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly introduced below, it is obvious that the drawings in the following description are only some embodiments described in the present application, and for those skilled in the art, other drawings can be obtained according to the drawings without any creative effort.
FIG. 1 is a schematic flow chart diagram of a method of dynamic load balancing according to an embodiment of the present application;
FIG. 2 is a schematic flow chart diagram of a method of dynamic load balancing according to another embodiment of the present application;
FIG. 3 is a schematic block diagram of a dynamic load balancing apparatus according to an embodiment of the present application;
fig. 4 is a schematic hardware structure diagram of a dynamic load balancing apparatus according to an embodiment of the present application.
Detailed Description
An object of the embodiments of the present application is to provide a dynamic load balancing method and apparatus, so as to solve the problem in the prior art that it is difficult to accurately and timely perform load balancing on a backend service node.
In order to make those skilled in the art better understand the technical solutions in the present application, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The dynamic load balancing method provided by the embodiment of the application can be applied to a load balancing system formed by a gateway and a back-end service node, wherein the gateway is used for receiving a service request and distributing the service request to the back-end service node according to configuration information. The dynamic load balancing method provided in the embodiments of the present application is described in detail below by taking the application to the load balancing system as an example.
Fig. 1 is a schematic flow chart of a dynamic load balancing method according to an embodiment of the present application, as shown in fig. 1, the method includes:
s102, responding to the received service request, and acquiring pre-configured configuration information corresponding to the service request. The configuration information includes a plurality of backend service nodes for processing the service request and first load parameter information corresponding to each backend service node.
The first load parameter information is used for representing the priority of the back-end service node for receiving the service request.
S104, determining a first target service node corresponding to the service request according to the configuration information, and routing the service request to the first target service node.
And S106, after the first target service node successfully receives the service request, determining the current load state corresponding to each back-end service node.
And S108, determining second load parameter information corresponding to each back-end service node according to the current load state and a preset load balancing strategy, and dynamically adjusting the first load parameter information corresponding to each back-end service node into the second load parameter information, so that the priority corresponding to each back-end service node is re-determined based on the second load parameter information.
By adopting the technical scheme of the embodiment of the application, the received service request is responded, the service request is routed to the first target service node according to the pre-configured configuration information corresponding to the service request, and after the first target service node successfully receives the service request, the first load parameter information corresponding to each rear-end service node is dynamically adjusted to the second load parameter information according to the current load state corresponding to each rear-end service node and the preset load balancing strategy, so that the priority corresponding to each rear-end service node is determined again based on the second load parameter information. Therefore, the technical scheme can dynamically adjust the load parameter information respectively corresponding to each back-end service node according to the current load state respectively corresponding to each back-end service node and by combining the preset load balancing strategy, so as to realize load balancing, and therefore, the load parameter information respectively corresponding to each back-end service node can be accurately and timely adjusted, namely, the priority of each back-end service node for receiving the service request is adjusted, so that the utilization rate of each back-end service node is improved, the received service request can be timely processed, and the processing efficiency of the service request is improved.
In an embodiment, before the pre-configured configuration information corresponding to the service request is obtained in response to the received service request, the configuration information corresponding to the service request may be obtained through an application Programming interface (api), and the obtained configuration information is configured in the resource library and the memory, and then the configuration information in the resource library may be updated according to a preset configuration information update policy, and the updated configuration information in the resource library is synchronized to the memory.
The configuration information updating policy may include any one or more of the following policies: and updating the strategy of the configuration information at regular time according to a preset time interval and updating the strategy of the configuration information after the first target service node is determined to successfully receive the service request so as to enable the configuration information to be matched with the current load state respectively corresponding to each back-end service node.
In this embodiment, the policy for updating the configuration information at regular time according to the preset time interval, and when the configuration information is updated, a first time interval for updating the configuration information in the resource library and a second time interval for updating the configuration information in the memory may be the same or different. When the first time interval is the same as the second time interval, the configuration information in the resource library and the configuration information in the memory can be updated simultaneously, so that the timeliness of updating the configuration information in the memory can be ensured.
In one embodiment, the gateway may be an API gateway based on OpenResty (a high-performance Web platform based on Nginx and Lua), and the repository may be consul (a service discovery, configuration management center service developed in go language from google). In consul, configuration information is stored in the form of Key-Value. In an init _ worker _ by _ lua phase (a loading configuration phase) in the OpenResty service starting process, each worker (work) process creates a timer, initiates an http request to a consul in the timer, and requests to retrieve and configure configuration information into a shared dictionary (an OpenResty memory). Thereafter, each timer may initiate a request periodically (e.g., every 1 second) to periodically update the configuration information in the memory. In the balance _ by _ lua stage (load balancing stage) of OpenResty, configuration information in the shared dictionary can be read, and a received service request is routed. Examples of configuration information are as follows:
Figure BDA0002905968750000051
Figure BDA0002905968750000061
in the configuration information, the health check is the configuration information corresponding to the health check policy. active indicates active health check, which can be performed according to tcp (transmission control protocol), interval indicates 5 seconds, and if the number of successful times is 3, the service node is a healthy node. The server is configuration information of the back-end service node, including an IP (Internet Protocol) address (e.g. 192.168.0.12), a port number (e.g. 8080) and a weight of the back-end service node. And backing up represents a standby node, and when the back-end service nodes in the primary are all unhealthy nodes, the service request is distributed to the standby service nodes in the backing up so as to ensure that the service request can be correctly processed and is not discarded finally.
In one embodiment, a Dashboard service may be included in the load balancing system, which provides a visual web interface to view various information about the current system. In response to the received service request, before obtaining the pre-configured configuration information corresponding to the service request, in addition to the listed configuration information corresponding to the service request obtained through the API, the configuration information edited by the user may also be obtained through the Dashboard service. The configuration information can be acquired without directly editing the configuration file by a user, so that the error rate of operation can be reduced, and the accuracy and efficiency of the acquired configuration information are improved. In addition, according to the actual situation, the configuration file can be manually changed through an API or a Dashboard service.
In the above embodiment, the configuration information corresponding to the service request may be acquired through the API, and then the acquired configuration information may be configured in the resource library and the memory, and the configuration information in the resource library may be updated according to a preset configuration information update policy, and the updated configuration information in the resource library may be synchronized into the memory, so that a service (e.g., OpenResty) does not need to be restarted, thereby implementing hot loading of the configuration information.
In one embodiment, the configuration information further includes a health check policy corresponding to each backend service node. When determining the first target service node corresponding to the service request according to the configuration information, the method may first determine, according to first load parameter information corresponding to each back-end service node, a priority level of receiving the service request corresponding to each back-end service node, determine that the back-end service node with the highest priority level is the first service node corresponding to the service request, perform health check on the first service node according to a health check policy to determine whether the first service node is a healthy node, and determine that the first service node is the first target service node corresponding to the service request when the first service node is a healthy node.
Following the configuration information example listed in the above embodiment, it can be known that the configuration information includes two backend service nodes, and assuming that the first load parameter information is the weight corresponding to each backend service node, it can be determined that the weights of the two backend service nodes are equal, so that the priorities of receiving the service requests corresponding to the two backend service nodes are equal. When a service request is received, a polling mechanism can be adopted to determine a first service node corresponding to the service request according to the configuration information, and then health check is carried out on the first service node according to a health check strategy so as to judge whether the first service node is a healthy node or not.
In this embodiment, a first service node corresponding to a service request is determined according to first load parameter information corresponding to each back-end service node, and health check is performed on the first service node, and when the first service node is a healthy node, the first service node is determined to be a first target service node corresponding to the service request, which provides a basis for accurately routing the service request, and is beneficial to improving the utilization rate of each back-end service node, and the received service request can be processed in time, thereby improving the processing efficiency of the service request.
In one embodiment, after performing health check on the first service node according to the health check policy, if it is determined that the first service node is an unhealthy node, the backend service node with the highest priority may be selected from other backend service nodes except the unhealthy node as the first target service node.
After selecting the backend service node with the highest priority from other backend service nodes except the unhealthy node, performing health check on the backend service node according to a health check strategy to judge whether the backend service node is a healthy node, and when the backend service node is determined to be a healthy node, determining that the backend service node is a first target service node corresponding to the service request.
In an embodiment, after performing health check on the first service node according to the health check policy, if it is determined that the first service node is a non-healthy node, it may also be determined that the standby node corresponding to the standby node information is the first target service node. Wherein the configuration information comprises standby node information.
In the embodiment, the service request can be ensured to be correctly processed and not discarded finally, and the reliability of the load balancing system is improved.
In an embodiment, according to the health check policy, health check is performed on the first service node to determine whether the first service node is a healthy node, and when the first service node is determined to be a healthy node, connection operation may be performed on a port corresponding to the first service node for a preset number of times, a connection success number in the connection operation is determined, and whether the connection success number is greater than or equal to a preset threshold value is determined. If the connection success times are larger than or equal to a preset threshold value, determining that the first service node is a healthy node; and if the connection success times are smaller than a preset threshold value, determining that the first service node is an unhealthy node.
In this embodiment, health check may be performed on the port corresponding to the first service node according to the tcp protocol, the connection operation performed on the port corresponding to the first service node for the preset number of times may be three-way handshake of the tcp protocol, and the first service node may be considered as a healthy node after the three-way handshake is completed, that is, the preset threshold is 3 times, and when the connection success number is equal to 3 times, the first service node may be determined as a healthy node. If the connection success times are less than 3, the first service node is determined to be an unhealthy node.
For the unhealthy node, the gateway may remove the service node from the back-end service node corresponding to the service request. When the service node is judged to be a healthy node by subsequent health examination, the gateway can automatically add the service node into a back-end service node corresponding to the service request.
When the gateway performs health check on each back-end service node, two health check modes of active health check and passive health check can be adopted. When the health check mode is active health check, the gateway can actively initiate a health check request at regular time and initiate a health check request when receiving a service request. When the health check mode is passive health check, the gateway only initiates the health check request when receiving the service request. In addition, in addition to the above-mentioned way of performing health check on the port corresponding to the backend service node according to the tcp protocol, it may also be determined whether the backend service node is a healthy node according to data such as an address endpoint or a returned status code that is configured in advance in the backend service node and returned in response to the health check request.
In this embodiment, the gateway integrates a function of performing health check on the back-end service node, and by determining whether the back-end service node is a healthy node, the back-end service node corresponding to the service request is dynamically reduced or increased, so that the service request is prevented from being distributed to the unhealthy node, and the distribution accuracy and efficiency of the service request are improved.
In one embodiment, when determining the current load state corresponding to each backend service node, and determining the second load parameter information corresponding to each backend service node according to the current load state and a preset load balancing policy, the current load state corresponding to each backend service node may be collected by a designated service node in the backend service nodes, and the second load parameter information corresponding to each backend service node may be determined according to each collected current load state by the designated service node.
In an embodiment, when determining the current load state corresponding to each backend service node, and determining the second load parameter information corresponding to each backend service node according to the current load state and a preset load balancing policy, each backend service node may further determine its own current load state, and determine its own second load parameter information according to its own current load state.
In the above embodiment, a service of Gossip protocol (protocol for information exchange between nodes or processes based on epidemic propagation manner) may be deployed in each backend service node, so that the backend service nodes may be aware of each other. After the services based on the Gossip protocol are deployed, each back-end service node forms a back-end service cluster. In the back-end service cluster, each back-end service node judges the load state of the back-end service node according to the current network state of the back-end service node (such as whether the network is connected or not, whether the network is congested or not), the running condition of a CPU (central processing unit), the occupied size of a memory and other indexes, and then the current load state of each back-end service node in the back-end service cluster can be determined based on the service of the Gossip protocol, so that second load parameter information corresponding to each back-end service node is determined according to the current load state of each back-end service node.
Through applying the Gossip protocol, the back-end service nodes can be mutually sensed, so that the current load state of each back-end service node in the back-end service cluster can be determined based on the load state of each back-end service node, and the configuration information of the gateway is dynamically modified based on the current load state of each back-end service node. Therefore, the load balancing system can dynamically and intelligently adjust the configuration information of the gateway, reduce human intervention, enable the load balancing to be more accurate and timely, and improve the utilization rate of each back-end service node.
In this embodiment, the second load parameter information corresponding to each back-end service node is determined according to the current load state of each back-end service node, so that the determined second load parameter information is more accurate, and the accuracy of subsequent load balancing can be improved.
In one embodiment, the current load status of each back-end service node may include a current load amount, the first load parameter information may include a first load weight, and the second load parameter information may include a second load weight.
The current load amount may be a load size of each back-end service node, a ratio of the load size of each back-end service node in a total load size corresponding to each back-end service node, and the like. The priority of receiving the service request is positively correlated with the load weight.
When the first load parameter information corresponding to each back-end service node is dynamically adjusted to the second load parameter information according to the current load state of each back-end service node and a preset load balancing strategy, if the current load amount corresponding to the back-end service node meets a first preset condition, the first load weight corresponding to the back-end service node is reduced to the second load weight.
When the current load amount is the load size of each backend service node, the first preset condition may include: the current load is greater than or equal to a first preset threshold, the ratio of the current load in the total load corresponding to each rear-end service node is greater than or equal to a second preset threshold, and the like.
In addition, if the current load amount corresponding to the back-end service node meets a second preset condition, the first load weight corresponding to the back-end service node is increased to a second load weight. When the current load amount is the load size of each backend service node, the second preset condition may include: the current load is smaller than a third preset threshold, the ratio of the current load in the total load corresponding to each rear-end service node is smaller than a fourth preset threshold, and the like.
When the first load parameter information corresponding to each backend service node is dynamically adjusted to the second load parameter information, the second load parameter information corresponding to each backend service node can be reported by the designated service node through an API provided by the gateway, so that the gateway dynamically adjusts the first load parameter information corresponding to each backend service node in the resource library based on the received second load parameter information.
Or, each back-end service node reports the second load parameter information corresponding to the back-end service node through an API provided by the gateway, so that the gateway dynamically adjusts the first load parameter information corresponding to each back-end service node in the resource library based on the received second load parameter information. Or, the first load parameter information corresponding to each back-end service node in the resource library is dynamically adjusted by the designated service node or each back-end service node according to the second load parameter information.
In this embodiment, when the current load amount corresponding to the back-end service node satisfies the first preset condition, the first load weight corresponding to the back-end service node is reduced to the second load weight, so as to reduce the service requests that need to be processed by the back-end service node. On the contrary, when the current load amount corresponding to the back-end service node meets the second preset condition, the first load weight corresponding to the back-end service node is increased to the second load weight, so that the utilization rate of the back-end service node is improved. The system resources of the back-end service nodes are fully utilized, and the water level of the system resources of each back-end service node is effectively ensured to be kept in a relatively balanced state.
In an embodiment, after the first load parameter information corresponding to each backend service node is dynamically adjusted to the second load parameter information, when a new service request is received, a second target service node corresponding to the new service request may be determined according to the second load parameter information, and the new service request is switched to the second target service node.
When a second target service node corresponding to a new service request is determined according to second load parameter information corresponding to each back-end service node, priorities of receiving the new service request corresponding to each back-end service node can be determined according to the second load parameter information, the back-end service node with the highest priority is determined to be the second service node corresponding to the new service request, then health check is carried out on the second service node according to a health check strategy to judge whether the second service node is a healthy node, and if the second service node is the healthy node, the second service node is determined to be the second target service node corresponding to the new service request. If the second service node is an unhealthy node, the second target service node corresponding to the new service request may be determined again according to the method for determining the first target service node recited in the above embodiment.
When the new service request is switched to the second target service node, the new service request can be directly routed to the second target service node. In addition, for a new service request of a preset main thread (that is, a default backend service node corresponding to the new service request is set), when the new service request is switched to a second target service node, the new service request can be switched from the backend service node corresponding to the main thread to the second target service node.
In this embodiment, after receiving the new service request, the second target service node corresponding to the new service request may be determined according to the second load parameter information corresponding to each backend service node, and the new service request is switched to the second target service node, so that the new service request can be processed in time, and the processing efficiency of the service request is improved.
Fig. 2 is a schematic flow chart of a dynamic load balancing method according to another embodiment of the present application, as shown in fig. 2, the method includes:
s201, responding to the received service request, and acquiring pre-configured configuration information corresponding to the service request. The configuration information includes a plurality of backend service nodes for processing the service request and first load parameter information corresponding to each backend service node. The first load parameter information is used for representing the priority of the back-end service node for receiving the service request.
In one embodiment, before the received service request is responded, the pre-configured configuration information corresponding to the service request is acquired through an application program interface API, the acquired configuration information is configured in the resource library and the memory, then the configuration information in the resource library is updated according to a preset configuration information updating strategy, and the updated configuration information in the resource library is synchronized in the memory. The configuration information updating policy may include any one or more of the following policies: and updating the strategy of the configuration information at regular time according to a preset time interval and updating the strategy of the configuration information after the target service node is determined to successfully receive the service request so as to enable the configuration information to be matched with the current load state respectively corresponding to each back-end service node.
S202, determining the priority of receiving the service request corresponding to each back-end service node according to the first load parameter information corresponding to each back-end service node.
S203, the back-end service node with the highest priority is determined as the first service node corresponding to the service request.
S204, performing health check on the first service node according to the health check strategy.
The configuration information includes health check strategies corresponding to the backend service nodes respectively.
S205, judging whether the first service node is a healthy node; if yes, go to S206; if not, go to S207.
S206, the first service node is determined to be a first target service node corresponding to the service request. After that, S208 is executed.
S207, selecting a back-end service node with the highest priority as a first target service node from other back-end service nodes except the unhealthy nodes; or, determining that the standby node corresponding to the standby node information in the configuration information is the first target service node.
S208, the service request is routed to the first target service node.
S209, after the first target service node successfully receives the service request, determining the current load status corresponding to each back-end service node.
S210, according to the current load state and a preset load balancing strategy, second load parameter information corresponding to each back-end service node is determined, and first load parameter information corresponding to each back-end service node is dynamically adjusted to be the second load parameter information, so that the priority corresponding to each back-end service node is re-determined based on the second load parameter information.
In one embodiment, the current load state corresponding to each backend service node can be collected through a designated service node in the backend service nodes, and the second load parameter information corresponding to each backend service node is determined through the designated service node according to each collected current load state. Or, each back-end service node determines its own current load state, and each back-end service node determines its own second load parameter information according to its own current load state.
In one embodiment, the current load state includes a current load amount, the first load parameter information includes a first load weight, and the second load parameter information includes a second load weight. And if the current load amount corresponding to the rear-end service node meets a first preset condition, reducing the first load weight corresponding to the rear-end service node to a second load weight. Wherein, the first preset condition may include: the current load is greater than or equal to a first preset threshold, the ratio of the current load in the total load corresponding to each rear-end service node is greater than or equal to a second preset threshold, and the like.
And S211, when a new service request is received, determining a second target service node corresponding to the new service request according to the second load parameter information respectively corresponding to each back-end service node, and switching the new service request to the second target service node.
The specific processes of S201 to S211 are described in detail in the above embodiments, and are not described herein again.
By adopting the technical scheme of the embodiment of the application, the received service request is responded, the service request is routed to the first target service node according to the pre-configured configuration information corresponding to the service request, and after the first target service node successfully receives the service request, the first load parameter information corresponding to each rear-end service node is dynamically adjusted to the second load parameter information according to the current load state corresponding to each rear-end service node and the preset load balancing strategy, so that the priority corresponding to each rear-end service node is determined again based on the second load parameter information. Therefore, the technical scheme can dynamically adjust the load parameter information respectively corresponding to each back-end service node according to the current load state respectively corresponding to each back-end service node and by combining the preset load balancing strategy, so as to realize load balancing, and therefore, the load parameter information respectively corresponding to each back-end service node can be accurately and timely adjusted, namely, the priority of each back-end service node for receiving the service request is adjusted, so that the utilization rate of each back-end service node is improved, the received service request can be timely processed, and the processing efficiency of the service request is improved.
In summary, particular embodiments of the present subject matter have been described. Other embodiments are within the scope of the following claims. In some cases, the actions recited in the claims can be performed in a different order and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing may be advantageous.
Based on the same idea, the dynamic load balancing method provided in the embodiments of the present application further provides a dynamic load balancing device.
Fig. 3 is a schematic block diagram of a dynamic load balancing apparatus according to an embodiment of the present application, and as shown in fig. 3, the dynamic load balancing apparatus includes:
a first obtaining module 310, configured to, in response to a received service request, obtain pre-configured configuration information corresponding to the service request; the configuration information comprises a plurality of back-end service nodes used for processing the service request and first load parameter information corresponding to each back-end service node; the first load parameter information is used for representing the priority of the back-end service node for receiving the service request;
the determining and executing module 320 is configured to determine a first target service node corresponding to the service request according to the configuration information, and route the service request to the first target service node;
a first determining module 330, configured to determine, after the first target service node successfully receives the service request, current load states corresponding to the backend service nodes, respectively;
the determining and adjusting module 340 is configured to determine second load parameter information corresponding to each backend service node according to the current load state and a preset load balancing policy, and dynamically adjust first load parameter information corresponding to each backend service node to the second load parameter information, so that the priority corresponding to each backend service node is determined again based on the second load parameter information.
In one embodiment, the dynamic load balancing apparatus further comprises:
the second determining module is used for determining a second target service node corresponding to the new service request according to the second load parameter information when the new service request is received;
and the switching module is used for switching the new service request to the second target service node.
In one embodiment, the current load state includes a current load amount; the first load parameter information includes a first load weight; the second load parameter information includes a second load weight;
the determining and adjusting module 340 includes:
the execution unit is used for reducing the first load weight corresponding to the back-end service node to a second load weight if the current load amount corresponding to the back-end service node meets a first preset condition;
wherein the first preset condition comprises at least one of the following: the current load is greater than or equal to a first preset threshold, and the ratio of the current load to the total load corresponding to each rear-end service node is greater than or equal to a second preset threshold.
In one embodiment, the determining module 330, determining and adjusting module 340 includes:
the collection and determination unit is used for collecting the current load state corresponding to each back-end service node through a designated service node in the back-end service nodes; determining second load parameter information corresponding to each rear-end service node respectively through the designated service node and according to each collected current load state; and/or the presence of a gas in the gas,
the first determining unit is used for determining the current load state of each back-end service node by each back-end service node, and determining the second load parameter information of each back-end service node according to the current load state of each back-end service node.
In one embodiment, the dynamic load balancing apparatus further comprises:
the second acquisition module is used for acquiring the configuration information corresponding to the service request through an Application Program Interface (API) before the pre-configured configuration information corresponding to the service request is acquired in response to the received service request;
the configuration module is used for configuring the acquired configuration information into a resource library and a memory;
and the updating module is used for updating the configuration information in the resource library according to a preset configuration information updating strategy and synchronizing the updated configuration information in the resource library into the memory.
In one embodiment, the configuration information update policy includes:
updating the strategy of the configuration information at regular time according to a preset time interval;
and/or the presence of a gas in the gas,
and updating the strategy of the configuration information after the first target service node is determined to successfully receive the service request so as to enable the configuration information to be matched with the current load state respectively corresponding to each back-end service node.
In one embodiment, the configuration information further includes a health check policy corresponding to each backend service node;
the determining and executing module 320 includes:
a second determining unit, configured to determine, according to the first load parameter information corresponding to each backend service node, a priority for receiving the service request corresponding to each backend service node;
a third determining unit, configured to determine a back-end service node with a highest priority as a first service node corresponding to the service request;
the execution and judgment unit is used for carrying out health check on the first service node according to the health check strategy so as to judge whether the first service node is a healthy node;
and if so, determining that the first service node is a first target service node corresponding to the service request.
In one embodiment, the dynamic load balancing apparatus further comprises:
the execution module is used for selecting a back-end service node with the highest priority as a first target service node from other back-end service nodes except the unhealthy node if the first service node is determined to be the unhealthy node; and/or the presence of a gas in the gas,
a third determining module, configured to determine, if it is determined that the first service node is an unhealthy node, that a standby node corresponding to the standby node information is a first target service node; wherein the configuration information comprises standby node information.
By adopting the device of the embodiment of the application, the received service request is responded, the service request is routed to the first target service node according to the pre-configured configuration information corresponding to the service request, and after the first target service node successfully receives the service request, the first load parameter information corresponding to each rear-end service node is dynamically adjusted to the second load parameter information according to the current load state corresponding to each rear-end service node and the preset load balancing strategy, so that the priority corresponding to each rear-end service node is determined again based on the second load parameter information. Therefore, the device can dynamically adjust the load parameter information corresponding to each back-end service node according to the current load state corresponding to each back-end service node and by combining the preset load balancing strategy, so as to realize load balancing, and therefore, the load parameter information corresponding to each back-end service node can be accurately and timely adjusted, namely, the priority of each back-end service node for receiving the service request is adjusted, so that the utilization rate of each back-end service node is improved, the received service request can be timely processed, and the processing efficiency of the service request is improved.
It should be understood by those skilled in the art that the dynamic load balancing apparatus in fig. 3 can be used to implement the aforementioned dynamic load balancing method, and the detailed description thereof should be similar to that of the foregoing method, and is not repeated herein to avoid complexity.
Based on the same idea, the embodiment of the present application further provides a dynamic load balancing device, as shown in fig. 4. The dynamic load balancing apparatus may have a large difference due to different configurations or performances, and may include one or more processors 401 and a memory 402, where the memory 402 may store one or more stored applications or data. Wherein memory 402 may be transient or persistent. The application program stored in memory 402 may include one or more modules (not shown), each of which may include a series of computer-executable instructions for a dynamic load balancing device. Still further, the processor 401 may be configured to communicate with the memory 402 to execute a series of computer-executable instructions in the memory 402 on a dynamic load balancing device. The dynamic load balancing apparatus may also include one or more power supplies 403, one or more wired or wireless network interfaces 404, one or more input-output interfaces 405, one or more keyboards 406.
In particular, in this embodiment, the dynamic load balancing apparatus includes a memory, and one or more programs, wherein the one or more programs are stored in the memory, and the one or more programs may include one or more modules, and each module may include a series of computer-executable instructions for the dynamic load balancing apparatus, and the one or more programs configured to be executed by the one or more processors include computer-executable instructions for:
responding to the received service request, and acquiring pre-configured configuration information corresponding to the service request; the configuration information comprises a plurality of back-end service nodes used for processing the service request and first load parameter information corresponding to each back-end service node; the first load parameter information is used for representing the priority of the back-end service node for receiving the service request;
determining a first target service node corresponding to the service request according to the configuration information, and routing the service request to the first target service node;
after the first target service node successfully receives the service request, determining the current load state corresponding to each back-end service node;
and according to the current load state and a preset load balancing strategy, determining second load parameter information corresponding to each back-end service node, and dynamically adjusting the first load parameter information corresponding to each back-end service node into the second load parameter information, so that the priority corresponding to each back-end service node is re-determined based on the second load parameter information.
By adopting the device of the embodiment of the application, the received service request is responded, the service request is routed to the first target service node according to the pre-configured configuration information corresponding to the service request, and after the first target service node successfully receives the service request, the first load parameter information corresponding to each rear-end service node is dynamically adjusted to the second load parameter information according to the current load state corresponding to each rear-end service node and the preset load balancing strategy, so that the priority corresponding to each rear-end service node is determined again based on the second load parameter information. Therefore, the equipment can dynamically adjust the load parameter information corresponding to each back-end service node according to the current load state corresponding to each back-end service node and by combining the preset load balancing strategy, so as to realize load balancing, and therefore, the load parameter information corresponding to each back-end service node can be accurately and timely adjusted, namely, the priority of each back-end service node for receiving the service request is adjusted, so that the utilization rate of each back-end service node is improved, the received service request can be timely processed, and the processing efficiency of the service request is improved.
The embodiment of the present application further provides a storage medium, where the storage medium stores one or more computer programs, where the one or more computer programs include an instruction, and when the instruction is executed by a dynamic load balancing device including multiple application programs, the dynamic load balancing device can execute each process of the dynamic load balancing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not described here again.
The systems, devices, modules or units illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product with certain functions. One typical implementation device is a computer. In particular, the computer may be, for example, a personal computer, a laptop computer, a cellular telephone, a camera phone, a smartphone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.
For convenience of description, the above devices are described as being divided into various units by function, and are described separately. Of course, the functionality of the units may be implemented in one or more software and/or hardware when implementing the present application.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include forms of volatile memory in a computer readable medium, Random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
Computer-readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, a computer readable medium does not include a transitory computer readable medium such as a modulated data signal and a carrier wave.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The application may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. The application may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
The embodiments in the present specification are described in a progressive manner, and the same and similar parts among the embodiments are referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the system embodiment, since it is substantially similar to the method embodiment, the description is simple, and for the relevant points, reference may be made to the partial description of the method embodiment.
The above description is only an example of the present application and is not intended to limit the present application. Various modifications and changes may occur to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the scope of the claims of the present application.

Claims (11)

1. A method for dynamic load balancing, comprising:
responding to a received service request, and acquiring pre-configured configuration information corresponding to the service request; the configuration information comprises a plurality of back-end service nodes for processing the service request and first load parameter information corresponding to each back-end service node; the first load parameter information is used for representing the priority of the back-end service node for receiving the service request;
determining a first target service node corresponding to the service request according to the configuration information, and routing the service request to the first target service node;
after the first target service node successfully receives the service request, determining the current load state corresponding to each back-end service node;
and determining second load parameter information corresponding to each back-end service node according to the current load state and a preset load balancing strategy, and dynamically adjusting the first load parameter information corresponding to each back-end service node into the second load parameter information, so that the priority corresponding to each back-end service node is determined again based on the second load parameter information.
2. The method according to claim 1, wherein after the first load parameter information corresponding to each of the backend service nodes is dynamically adjusted to the second load parameter information, the method further comprises:
when a new service request is received, determining a second target service node corresponding to the new service request according to the second load parameter information;
and switching the new service request to the second target service node.
3. The method of claim 1, wherein the current load state comprises a current load amount; the first load parameter information comprises a first load weight; the second load parameter information comprises a second load weight;
the determining, according to the current load state and a preset load balancing policy, second load parameter information corresponding to each back-end service node, and dynamically adjusting the first load parameter information corresponding to each back-end service node to the second load parameter information includes:
if the current load amount corresponding to the back-end service node meets a first preset condition, reducing the first load weight corresponding to the back-end service node to the second load weight;
wherein the first preset condition comprises at least one of: the current load is greater than or equal to a first preset threshold, and the ratio of the current load in the total load corresponding to each backend service node is greater than or equal to a second preset threshold.
4. The method according to claim 1, wherein the determining a current load status corresponding to each of the back-end service nodes; determining second load parameter information respectively corresponding to each back-end service node according to the current load state and a preset load balancing strategy, wherein the second load parameter information comprises:
collecting the current load state corresponding to each back-end service node through a designated service node in the back-end service nodes; determining, by the designated service node and according to the collected current load states, second load parameter information corresponding to each of the backend service nodes, respectively; and/or the presence of a gas in the gas,
and determining the current load state of each back-end service node by each back-end service node, and determining the second load parameter information of each back-end service node according to the current load state of each back-end service node.
5. The method of claim 1, further comprising:
the method comprises the steps that before a received service request is responded, pre-configured configuration information corresponding to the service request is obtained, and the configuration information corresponding to the service request is obtained through an Application Program Interface (API);
configuring the acquired configuration information into a resource library and a memory;
updating the configuration information in the resource library according to a preset configuration information updating strategy, and synchronizing the updated configuration information in the resource library to the memory.
6. The method of claim 5, wherein the configuration information update policy comprises:
updating the strategy of the configuration information at regular time according to a preset time interval;
and/or the presence of a gas in the gas,
and updating the policy of the configuration information after the first target service node is determined to successfully receive the service request, so that the configuration information is matched with the current load state respectively corresponding to each back-end service node.
7. The method of claim 1, wherein the configuration information further includes a health check policy corresponding to each backend service node;
the determining, according to the configuration information, a first target service node corresponding to the service request includes:
determining the priority of each back-end service node for receiving the service request according to the first load parameter information corresponding to each back-end service node;
determining the back-end service node with the highest priority as a first service node corresponding to the service request;
according to the health check strategy, performing health check on the first service node to judge whether the first service node is a healthy node;
if so, determining that the first service node is the first target service node corresponding to the service request.
8. The method of claim 7, wherein after performing the health check on the first service node according to the health check policy to determine whether the first service node is a healthy node, the method further comprises:
if the first service node is determined to be an unhealthy node, selecting the back-end service node with the highest priority from other back-end service nodes except the unhealthy node as the first target service node; and/or the presence of a gas in the gas,
if the first service node is determined to be an unhealthy node, determining a standby node corresponding to standby node information to be the first target service node; wherein the configuration information includes the standby node information.
9. A dynamic load balancing apparatus, comprising:
the first acquisition module is used for responding to the received service request and acquiring pre-configured configuration information corresponding to the service request; the configuration information comprises a plurality of back-end service nodes for processing the service request and first load parameter information corresponding to each back-end service node; the first load parameter information is used for representing the priority of the back-end service node for receiving the service request;
the determining and executing module is used for determining a first target service node corresponding to the service request according to the configuration information and routing the service request to the first target service node;
a first determining module, configured to determine, after the first target service node successfully receives the service request, current load states corresponding to the backend service nodes, respectively;
and the determining and adjusting module is used for determining second load parameter information corresponding to each rear-end service node according to the current load state and a preset load balancing strategy, and dynamically adjusting the first load parameter information corresponding to each rear-end service node into the second load parameter information, so that the priority corresponding to each rear-end service node is re-determined based on the second load parameter information.
10. A dynamic load balancing device comprising a processor and a memory electrically connected to the processor, the memory storing a computer program, the processor being configured to invoke and execute the computer program from the memory to implement the dynamic load balancing method of any one of claims 1 to 8.
11. A storage medium for storing a computer program for execution by a processor to implement the dynamic load balancing method of any one of claims 1 to 8.
CN202110071520.1A 2021-01-19 2021-01-19 Dynamic load balancing method and device Pending CN112929408A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110071520.1A CN112929408A (en) 2021-01-19 2021-01-19 Dynamic load balancing method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110071520.1A CN112929408A (en) 2021-01-19 2021-01-19 Dynamic load balancing method and device

Publications (1)

Publication Number Publication Date
CN112929408A true CN112929408A (en) 2021-06-08

Family

ID=76163758

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110071520.1A Pending CN112929408A (en) 2021-01-19 2021-01-19 Dynamic load balancing method and device

Country Status (1)

Country Link
CN (1) CN112929408A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113467910A (en) * 2021-06-30 2021-10-01 乐视云计算有限公司 Overload protection scheduling method based on service grade
CN113746933A (en) * 2021-09-14 2021-12-03 京东科技信息技术有限公司 Method and device for displaying information
CN114003371A (en) * 2021-09-28 2022-02-01 中汽创智科技有限公司 Automatic driving load balancing method and device, electronic equipment and storage medium
CN114205361A (en) * 2021-12-08 2022-03-18 聚好看科技股份有限公司 Load balancing method and server
CN114866457A (en) * 2022-04-27 2022-08-05 猪八戒股份有限公司 High-performance dynamic route forwarding method, system and equipment based on Nginx and Lua
WO2023279555A1 (en) * 2021-07-08 2023-01-12 苏州浪潮智能科技有限公司 Topology-aware load balancing method and apparatus, and computer device
CN116450054A (en) * 2023-06-16 2023-07-18 成都泛联智存科技有限公司 IO request processing method, device, host and computer readable storage medium
WO2023155550A1 (en) * 2022-02-18 2023-08-24 中兴通讯股份有限公司 Message sending methods, message sending apparatus and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013050682A1 (en) * 2011-10-06 2013-04-11 Bull Sas Pseudo-dynamic adaptive routing method in a cluster including static communication links, and computer program implementing said method
CN104394224A (en) * 2014-11-28 2015-03-04 无锡华云数据技术服务有限公司 Load balance system
CN105007312A (en) * 2015-07-03 2015-10-28 叶秀兰 Method and system for controlling adaptive load-balancing of cloud computing server
CN107911470A (en) * 2017-11-30 2018-04-13 掌阅科技股份有限公司 Distributed dynamic load-balancing method, computing device and computer-readable storage medium
CN108965381A (en) * 2018-05-31 2018-12-07 康键信息技术(深圳)有限公司 Implementation of load balancing, device, computer equipment and medium based on Nginx
CN109104500A (en) * 2018-09-29 2018-12-28 广东省信息工程有限公司 A kind of server load balancing method and device of dynamic adjustment
CN109274707A (en) * 2017-07-18 2019-01-25 阿里巴巴集团控股有限公司 A kind of load dispatching method and device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013050682A1 (en) * 2011-10-06 2013-04-11 Bull Sas Pseudo-dynamic adaptive routing method in a cluster including static communication links, and computer program implementing said method
CN104394224A (en) * 2014-11-28 2015-03-04 无锡华云数据技术服务有限公司 Load balance system
CN105007312A (en) * 2015-07-03 2015-10-28 叶秀兰 Method and system for controlling adaptive load-balancing of cloud computing server
CN109274707A (en) * 2017-07-18 2019-01-25 阿里巴巴集团控股有限公司 A kind of load dispatching method and device
CN107911470A (en) * 2017-11-30 2018-04-13 掌阅科技股份有限公司 Distributed dynamic load-balancing method, computing device and computer-readable storage medium
CN108965381A (en) * 2018-05-31 2018-12-07 康键信息技术(深圳)有限公司 Implementation of load balancing, device, computer equipment and medium based on Nginx
CN109104500A (en) * 2018-09-29 2018-12-28 广东省信息工程有限公司 A kind of server load balancing method and device of dynamic adjustment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
汪佳文等: "基于权重轮询负载均衡算法的优化", 《计算机系统应用》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113467910A (en) * 2021-06-30 2021-10-01 乐视云计算有限公司 Overload protection scheduling method based on service grade
CN113467910B (en) * 2021-06-30 2024-05-28 乐视云网络技术(北京)有限公司 Overload protection scheduling method based on service level
US11902362B2 (en) 2021-07-08 2024-02-13 Inspur Suzhou Intelligent Technology Co., Ltd. Topology-aware load balancing method and apparatus, and computer device
WO2023279555A1 (en) * 2021-07-08 2023-01-12 苏州浪潮智能科技有限公司 Topology-aware load balancing method and apparatus, and computer device
CN113746933A (en) * 2021-09-14 2021-12-03 京东科技信息技术有限公司 Method and device for displaying information
CN114003371A (en) * 2021-09-28 2022-02-01 中汽创智科技有限公司 Automatic driving load balancing method and device, electronic equipment and storage medium
CN114003371B (en) * 2021-09-28 2023-03-07 中汽创智科技有限公司 Automatic driving load balancing method and device, electronic equipment and storage medium
CN114205361A (en) * 2021-12-08 2022-03-18 聚好看科技股份有限公司 Load balancing method and server
CN114205361B (en) * 2021-12-08 2023-10-27 聚好看科技股份有限公司 Load balancing method and server
WO2023155550A1 (en) * 2022-02-18 2023-08-24 中兴通讯股份有限公司 Message sending methods, message sending apparatus and storage medium
CN114866457B (en) * 2022-04-27 2024-01-16 猪八戒股份有限公司 High-performance dynamic route forwarding method, system and equipment based on Nginx and Lua
CN114866457A (en) * 2022-04-27 2022-08-05 猪八戒股份有限公司 High-performance dynamic route forwarding method, system and equipment based on Nginx and Lua
CN116450054B (en) * 2023-06-16 2023-09-26 成都泛联智存科技有限公司 IO request processing method, device, host and computer readable storage medium
CN116450054A (en) * 2023-06-16 2023-07-18 成都泛联智存科技有限公司 IO request processing method, device, host and computer readable storage medium

Similar Documents

Publication Publication Date Title
CN112929408A (en) Dynamic load balancing method and device
EP3637733B1 (en) Load balancing engine, client, distributed computing system, and load balancing method
US10009284B2 (en) Policy-based session establishment and transfer in a virtualized/cloud environment
WO2022093319A1 (en) Methods, systems, and computer readable media for rank processing for network function selection
KR20210119504A (en) Communication method and apparatus, entity and computer readable storage medium
JP2009244945A (en) Load distribution program, load distribution method, load distribution device and system including the same
CN111737003A (en) Pod balanced scheduling method and device, main node and storage medium
CN110650209B (en) Method and device for realizing load balancing
US20220329651A1 (en) Apparatus for container orchestration in geographically distributed multi-cloud environment and method using the same
CN112738220A (en) Management method, load balancing method and load balancing device of server cluster
CN111090507A (en) Task scheduling method based on cloud edge fusion server network architecture and application
CN115242798A (en) Task scheduling method based on edge cloud, electronic equipment and storage medium
CN110515728B (en) Server scheduling method and device, electronic equipment and machine-readable storage medium
JP2005092862A (en) Load distribution method and client-server system
O'Sullivan et al. The cloud personal assistant for providing services to mobile clients
CN114296953A (en) Multi-cloud heterogeneous system and task processing method
CN111338803B (en) Thread processing method and device
CN108124021B (en) Method, device and system for obtaining Internet Protocol (IP) address and accessing website
CN111913792A (en) Service processing method and device
CN109302302A (en) Service scalable method, system and the computer readable storage medium of network element
CN111522664A (en) Service resource management and control method and device based on distributed service
CN113630327B (en) Flow control method and device, electronic equipment and computer readable medium
CN114296869A (en) Server node service method and device based on TCP long connection
Shiraz et al. A distributed and elastic application processing model for mobile cloud computing
CN108520025B (en) Service node determination method, device, equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210608

RJ01 Rejection of invention patent application after publication