WO2013116664A1 - Dynamic load balancing in a scalable environment - Google Patents

Dynamic load balancing in a scalable environment Download PDF

Info

Publication number
WO2013116664A1
WO2013116664A1 PCT/US2013/024366 US2013024366W WO2013116664A1 WO 2013116664 A1 WO2013116664 A1 WO 2013116664A1 US 2013024366 W US2013024366 W US 2013024366W WO 2013116664 A1 WO2013116664 A1 WO 2013116664A1
Authority
WO
WIPO (PCT)
Prior art keywords
partition
load balancing
server
partitions
master
Prior art date
Application number
PCT/US2013/024366
Other languages
French (fr)
Inventor
Arild E. SKJOLSVOLD
Bradley Gene Calder
Ju Wang
Original Assignee
Microsoft Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Corporation filed Critical Microsoft Corporation
Priority to CN201380007789.7A priority Critical patent/CN104081353B/en
Priority to JP2014555763A priority patent/JP6185486B2/en
Priority to KR1020147021737A priority patent/KR102013004B1/en
Priority to EP13742889.2A priority patent/EP2810163A4/en
Publication of WO2013116664A1 publication Critical patent/WO2013116664A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5077Logical partitioning of resources; Management or configuration of virtualized resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system

Definitions

  • Cloud computing environments provide a potential platform for allowing users to have access to large amounts of computing resources without having to invest in corresponding infrastructure. Instead, the computing environment can be provided as a service by a cloud computing provider. This can allow a user to tune the requested computing resources to match the size and importance of a computing job. However, taking full advantage of the additional computing resources available as a service may require dividing computing tasks into smaller portions, so that multiple physical or virtual processors can be used to perform a task.
  • systems and methods are provided that enable a general framework for partitioning application-defined jobs in a scalable environment.
  • the general framework decouples partitioning of a job from the other aspects of the job. As a result, the effort required to define the application-defined job is reduced or minimized, as the user is not required to provide a partitioning algorithm.
  • the general framework further allows a user to provide load balancing conditions to allow for modification of how partitions are assigned.
  • FIGS. 1 to 3 schematically show various aspects of an example of a distributed computing environment, in accordance with aspects of the present invention
  • FIG. 4 schematically shows the interaction between various interfaces for performing a job in a generic partitioning environment, in accordance with aspects of the present invention
  • FIG. 5 shows an example of a generic partitioning environment, in accordance with aspects of the present invention
  • FIG. 6 shows an example of providing backup machines for active master roles among, in accordance with aspects of the present invention
  • FIG. 7 is a block diagram of an exemplary computing environment suitable for use in implementing embodiments of the present invention.
  • FIG. 8 shows another example of providing backup machines for master roles, in accordance with aspects of the present invention.
  • FIGS. 9 to 11 show examples of methods according to various embodiments of the invention.
  • FIG 12 shows an exemplary aspect of providing backup machines for master roles relative to fault domains and upgrade domains, in accordance with aspects of the present invention.
  • a user in a first location can submit a job or computing task to a computing service and have the task performed on a group of computers that the user has no direct knowledge of.
  • the computing resources for performing the user's task may be distributed over multiple locations.
  • a first group of computing resources located in one or more locations can store the data and other information for performing the user's computing task, while a second group of computing resources, in the same locations or possibly in a different set of one or more locations, can be used to perform the computing task.
  • Access to a variety of distributed computing resources allows a user to perform job tasks without concern for where the computing resources are located.
  • the distributed resources also provide an opportunity for a user to scale up (or scale down) the amount of resources used in order to meet goals for a computing task, such as completing the computing task by a specified time.
  • using distributed computing resources poses a number of challenges for a user. Conventionally, a user or application designer would need to divert resources away from designing an application or job in order to determine how to take advantage of a specific distributed computing environment.
  • systems and methods are provided that enable a generic framework for building scalable applications in a distributed environment, such as a cloud computing environment.
  • the generic framework can allow the partitioning of jobs or work items to be decoupled from performing computations associated with the jobs or work items. This allows the owner of a distributed environment to provide resources to a user in an accessible or simplified manner while still providing a high availability for the distributed resources.
  • the generic framework reduces the complexity of designing applications for a distributed environment by providing a framework for handling features such as scalability, fault tolerance, and/or availability while reducing or minimizing the amount of effort required to address these features. As a result, the framework can allow an application designer to focus on the application instead of spending additional time on requirements of the distributed environment.
  • a "key” is one of the basic concepts used in a generic partitioning framework.
  • a key is a value from a namespace or domain.
  • An example of a namespace is an identifier corresponding to all storage accounts in a cloud computing environment. In such an example, a key can correspond to an account name, account number, or another identifier that allows a specific account to be referenced.
  • Another example of a namespace is the range of possible alphanumeric values for an input parameter for a job.
  • the keys that span a namespace will also have some method of organization so that keys can be described as a range of serial values. For example, the keys can be organized numerically, alphabetically, based on hashed values, or in any other convenient serial manner that allows a range of keys to be defined based on specifying two keys as a range beginning and a range end.
  • a "partition” is a range defined by a low (inclusive) and high (exclusive) key.
  • a partition can never be an empty range. If desired, a single partition may contain the entire range of the domain. Partitions are defined to be mutually exclusive, so that there is not an overlap between the ranges of two different partitions. The union of all partitions will span the entire domain or namespace.
  • a "partition server” is a virtual machine within a cloud computing environment that corresponds to a role instance for serving zero or more partitions.
  • a partition is not served by more than one partition server at the same time. However, a given partition server may not have any partitions that are currently being served by the partition server.
  • a partition server can include both application-defined interfaces and fixed interfaces (defined by the partitioning system) for performing various actions.
  • a "partition master” is a role that manages partition servers for a given type of role, such as by assigning and unassigning partitions to partition servers.
  • the partition master role also monitors partition servers, such as by verifying that the partition servers are still serving assigned partitions. Typically, this role is redundant for fault tolerance.
  • a partition master can include both application-defined interfaces and fixed interfaces (defined by the partitioning system) for performing various actions.
  • An "application-defined interface” refers to a computation, operation, or other function defined by a client for performance by a role instance.
  • An application-defined interface can be created for either a partition master or partition server role.
  • An application-defined interface is in contrast to a "fixed interface" for a partitioning system.
  • a fixed interface refers to an application interface that is provided as part of the partitioning system.
  • a client cannot modify the action of a fixed interface.
  • an application-defined interface can be used in conjunction with a fixed interface.
  • a fixed interface for a master role can be to send a heartbeat message every time period to verify the status of any servers being managed by the master role.
  • a corresponding fixed interface for a server role can be an interface to respond to the heartbeat message in an appropriate manner.
  • An example of an application-defined interface can be an interface that adds additional information to a heartbeat message, while another example can be an interface to extract such additional information from a heartbeat message.
  • the sending of the heartbeat message itself is a fixed interface.
  • a user or application cannot modify the protocol for sending the message.
  • the interface for defining the message content can be modified by a user.
  • the basic structure of the interface for supplementing the information in a heartbeat message may be provided by the system.
  • a client can modify the content provided by this interface, such an interface is defined herein as an application-defined interface.
  • application-defined interfaces can be used in conjunction with fixed interfaces to provide the overall feature of load balancing.
  • the interfaces for assigning partitions to a partition server or splitting a partition into two partitions are fixed interfaces.
  • application-defined interfaces can provide expressions for when load balancing actions should occur, like changing partition assignments between machines or deciding when to split a partition.
  • Each active partition master role or partition server role has a corresponding storage object that controls the data corresponding to the role.
  • An example of a storage object is a binary large object or blob.
  • the storage object includes the identity of the partitions being served. Note that the storage object for a partition server will typically not contain the underlying data corresponding to the partition. By only storing partition identities while leaving the underlying data in a separate data storage location, partitions can be moved from one server to another with a minimum of data transfer.
  • the storage object includes information for establishing that a particular instance of the partition master is the active instance.
  • the storage object can optionally also include information regarding all storage objects for servers being managed by the master.
  • master and server roles can maintain a lease on the corresponding storage object. When a lease on a storage object for a role is broken, the corresponding role can be terminated.
  • partition masters for a given type of role are preferably redundant, so that at least one additional partition master is available if a failure occurs.
  • dictator is defined as the partition master that current performs the partition master functions for a given type of role.
  • the dictator is the partition master that maintains the lease for the storage object associated with the given partition master role.
  • the active partition master can communicate with the partition servers via heartbeats.
  • the basic heartbeat is of type 'keepalive' and is always used.
  • an application-defined interface can be used to add optional commands and/or information to this heartbeat message.
  • a generic partitioning framework is provided so that a user or application can define a computation that is automatically partitioned for processing in a distributed computing environment.
  • a user can define a computation based on a namespace.
  • a namespace corresponds to one or more computations or calculations that are optionally performed on data associated with the namespace.
  • the user-defined namespace can have certain properties as described below.
  • the generic partitioning framework will be able to automatically partition the namespace based on the application-defined interfaces. This partitioning is accomplished using only the user-defined namespace and fixed interfaces provided with the partitioning system. By partitioning a namespace using only fixed interfaces of the generic partitioning system, the partitioning is decoupled from any application-defined interfaces provided by the user as well as being decoupled from the nature of the computation(s) being performed.
  • a namespace or domain for a job can correspond to a set of identifiers for the full range of a type of data that will be operated on and/or generated by a computation. Additionally or alternately, a namespace or domain can correspond to a listing or range of possible states that could be operated on and/or generated by the computation. Still another option is that the namespace can correspond to a set of identifiers for multiple instances of a calculation being performed in a distributed environment. It is not necessary for all potential namespaces or domains for a computation to be identified. Instead, the user can identify the namespaces or states that will be used for partitioning. The identified namespaces or domains should span the entirety of the computations that a user will desire to perform using a given role.
  • the entirety of the computations can be spanned by spanning the entirety of a data set operated on by a computation, by spanning the entirety of instances of calculations within the computation, or in any other convenient manner.
  • Keys can be used to specify discrete values or states within a namespace. Keys can also be used to specify a range of values. Since keys can be used to specify a range, it is understood that the namespace should include some type of serial ordering, so that specifying an upper key and a lower key will uniquely identify a range. This ordering can be based on a conventional ordering, such as ordering based on an alphabet. Alternatively, the ordering can be arbitrary, such as a serial order based on a file containing a listing of the values within a namespace.
  • a user can also provide one or more application-defined interfaces associated with a server role.
  • An application-defined interface represents one or more tasks or functions that a partition server can perform, the tasks or functions optionally involving one or more namespaces.
  • the application-defined interfaces for a server role can include all tasks or functions that will be performed on a given namespace, so that different server roles do not attempt to access the same data or states.
  • a server role may also include sub-roles, so that some application-defined interfaces within a role operate on different namespaces than other application-defined interfaces within the same role.
  • a client may desire to perform a computation that involves performing at least one type of calculation on a data set.
  • a client may define a single role of a server that performs at least one type of calculation on requested element(s) from the data set.
  • This single role can represent one or more servers corresponding to role instances that are configured to perform a scientific calculation, one or more related data mining functions on a data set, or any other convenient computation.
  • the computation and any related data can be defined as part of the definition of a namespace for performing the computation.
  • one or more application-defined interfaces can be provided for the server role, such as an application-defined interface for providing metrics regarding the computation to a master role instance.
  • a server role instance can determine the element or elements of the data set to work on based on key values passed to the server.
  • At least one additional master role can manage the servers, including assigning data set partitions to the one or more servers.
  • a computation can be partitioned so that multiple partition servers handle or perform different portions of processing for the namespace.
  • Each partition corresponds to a range of key values.
  • the server performs the desired computation for any requests that contain a key value within the range corresponding to an assigned partition.
  • the partitions assigned to a server do not need to be consecutive relative to the serial ordering of the namespace.
  • a partition table can be used to track the current assignments of partitions to partition servers.
  • the partition table can be updated first to reflect the assignment.
  • the partition table can then be used to determine the partition server that will handle a client request based on the key specified in the client request.
  • one partition table can be used to track the partition assignments for multiple roles, as opposed to having a different partition table for each namespace that is partitioned.
  • an entry in a partition table can include the low key for a range, the high key for the range, and the role instance or server instance that will perform a requested task on the data or state corresponding to requested key.
  • a partition table can also include other data, such as an epoch number or version number, as will be discussed in greater detail below.
  • One option is to use a separate dictatorship library to implement dictatorship via storage object lease. Having a separate dictatorship library can offer several advantages, such as allowing a role outside of the partitioning system to implement redundancy. This allows roles that are not involved in partitioning to use the same method for selecting an active master role.
  • Each server maintains a lease on its own storage object.
  • Server storage object names are generated by the master each time a server receives its first partition assignment.
  • Each server also stores its current partition assignment (list of partitions including epoch numbers) in the storage object.
  • a third party can forcefully terminate the lease between a server and its storage object. This functionality can be used by the partition master dictator to break server leases in the partition assignment protocols.
  • a stale update refers to a situation where a message or instruction from a prior dictator is received by a server or data store after a new dictator has been selected.
  • the problem of stale updates may affect any role or piece of code which maintains state in an external store where messages to the external store may be delayed or re-ordered (e.g. over the internet), and where the state is read from that store upon failover of the role.
  • partition master (dictator) that fails over in the middle of updating the partition table for the role.
  • the old master initiates an update of row 'X' of the partition table.
  • the old master stops working.
  • a redundant instance of the master is selected as the new dictator.
  • This new active master updates, reads, or otherwise performs an action using row 'X' of the partition table.
  • the update from the old master then goes through after the action on row 'X' by the new active master. If the update from the old master is incorporated into the partition table, the update will cause the partition table to change without the new (current) master being aware of the change. This can lead to an inconsistent state for the partition table.
  • One solution to the problem in the example above is to somehow block the stale update from the old master.
  • One option is to also allow partition masters to obtain leases on portions of the partition table.
  • a lease for a master can correspond to all of the partitions in the table corresponding to a given namespace. Any convenient method for specifying a lease range can be used. For example, if it is desired, a lease can span only a portion of a partition and/or can span multiple partitions.
  • the masters will still compete for dictatorship via a lease, where an epoch number is provided so the servers can block stale updates, such as stale heartbeat messages.
  • the master dictator can also acquire a lease on the (relevant part of the) partition table before reading the table when building its state.
  • both masters and servers can use the lease mechanism to avoid stale updates by acquiring leases to blobs, tables, and/or other data structures when the master or server is assigned to handle a data structure.
  • the lease is released.
  • Stale updates can also be a concern for communication between a master and a server. Processing of messages from an old master can be avoided, for example, by having a master obtain an epoch number when it becomes the dictator. The epoch number is increased for every dictatorship failover. This epoch number can be transmitted on every server heartbeat and/or other message from a master to a server. A server can ignore any heartbeat with a lower epoch number than the highest epoch number the server has seen. This high- watermark epoch number can be stored in the servers' soft state. If a master and a server are also communicating status via storage object leases, the above method can be sufficient to avoid stale message issues. Alternatively, other convenient solutions for avoiding stale messages between a master and a server can also be used.
  • a similar epoch number method can be used for each range partition to avoid stale updates from a previously assigned server.
  • each partition can have a current epoch number that is updated by the master when a change occurs for the partition.
  • changes for a partition include assignment of a partition to a new server, splitting of a partition, and merging of two partitions. Assignments of a partition to a new server can cause the epoch number to increase by one. Splitting of a partition into two or more new partitions can cause each child partition to receive the parent's epoch number incremented by one.
  • the epoch number for the merged partition can be the maximum epoch number for any of the partitions prior to merge incremented by one.
  • any other convenient method for tracking an incrementing an epoch number for a partition can also be used.
  • the third server X can be, for example, a data server that contains the actual data set for the namespace operated on by the partition master and servers SI and S2. Servers SI and S2 issue commands (or other messages) to X.
  • a partition P with epoch 3 can initially be served by server SI .
  • SI can send a message Ml to X.
  • Message Ml contains a key corresponding to partition P as well as the epoch number 3.
  • the partition master then moves partition P from SI to S2.
  • S2 sends a message M2 to server X that includes the new epoch number.
  • X receives the message M2 prior to receiving message Ml . This renders Ml stale.
  • X then receives stale message Ml .
  • server X can recognize that the stale message Ml is from a server that no longer is responsible for partition P.
  • an epoch validation library can be used.
  • An epoch validation library validates that a received message is not stale. For example, when a server receives a message from a new dictator or receives a request related to a new partition range, the server can check an epoch validation library to verify that the message contains the current epoch number for the master or the partition.
  • FIG. 4 shows a schematic example of the relationship between roles, partitions, and keys.
  • a user 405 submits a request to perform an action on a desired piece of data, such as accessing information in the user's account.
  • the data is identified by a key.
  • the client request and the key are passed to role 420 which handles the type of request made by the client or user 405.
  • Role 420 includes a client library 430 that defines how the role handles a given type of request.
  • the role 420 consults the partition table 455 to find the current server that handles the key range corresponding to the key in the request.
  • the information in partition table 455 is populated based on the partition decisions made by partition master 460.
  • Partition master 460 is shown in FIG. 4 as being one of a plurality of potential partition masters.
  • the additional potential partition masters are for redundancy, and are not active until needed.
  • a plurality of partition servers 465 are available as role instances for performing tasks requested from the role 420.
  • partition server N from plurality of partition servers 465 is handling the range of keys corresponding to the key in the client request.
  • FIG. 5 shows an example of a group of modules and/or interfaces for providing a generic partitioning framework according to the invention.
  • FIG. 5 also displays examples of application-defined interfaces for providing computational jobs that can take advantage of the generic partitioning environment.
  • the interfaces for partitioning a namespace are fixed interfaces provided by the partitioning system. This decouples partitioning of the namespace from performing the job or work item that operates on the data corresponding to the namespace.
  • FIG. 5 there are at least two types of components or interfaces that are provided or designated by the user.
  • the user provides a definition for a key (and corresponding namespace) 510 and a plurality of server interfaces 520 for the application.
  • the key describes the namespace for the application. This allows the partitioning system to know the range of variables, states, and/or computations for potential partitioning.
  • the user also provides at least three interfaces related to the key.
  • the interfaces related to the key 510 provide functions to serialize the keys in a namespace, to deserialize keys, and for comparing two keys. Since the user selects the namespace and key values, the operation of these interfaces is not offered as a fixed interface by the partitioning system.
  • the interfaces related to keys and namespaces are instead provided by the user as application-defined interfaces.
  • the user also provides application-defined interfaces 520 that are used by server role instances.
  • the application-defined interfaces for server role instances 520 include functions for what operations a server will perform when a server receives a request to operate on a key.
  • an interface for StartServeKeyRange can define one or more operations a server can perform when the server that starts serving a given key range.
  • a corresponding interface for Stop ServeKeyRange can allow a server to end serving of a key range in an orderly manner. Additionally, it may be desirable to collect information from the server.
  • An interface such as an OnReceivedHeartbeatRequest interface can define how a server will extract additional information from a heartbeat message received from a current partition master.
  • BeforeSendingHeartbeatResponse can define what a server will include as additional information in a response to a heartbeat message. This can allow, for example, a server to communicate load information to a partition master for use in load balancing.
  • the user can also provide a user defined interfaces for the master role 530.
  • the user defined interfaces for the master role 530 are a complement to the fixed interfaces for the master role, schematically shown as GPMaster 535.
  • the user defined interfaces for the master role 530 are not required.
  • the general partitioning environment is enabled by five modules and/or groups of fixed partitioning system interfaces. These include the fixed interfaces shown as being part of GPClient module 515, the GPServer module 525, the GPMaster module 535, the GPDictator module 536, and the GPLease module 545. Of course, other ways of arranging the modules, interfaces, and/or functions shown in FIG. 5 can also be used.
  • the GPClient module 515 shown in FIG. 5 has interfaces that allow for translation from a key specified by a client application or user 599 to the address for the server that will handle a user request.
  • the interfaces for GPClient module 515 perform this lookup by consulting a partition map or partition table 559.
  • the GPClient module interfaces can then, for example, return the server address corresponding to the key to the client application 599 so that the client application can direct the request to the correct server.
  • the GPServer module 525 has interfaces that work in conjunction with the application-defined interfaces 520 to perform desired user tasks.
  • the GPServer module 525 includes interfaces for communication between master role instances and server role instances.
  • the GPServer module interfaces also communicate with the interfaces for GPLease module 545 to manage the lease objects and lease contents associated with each server role instance.
  • the GPMaster module 535 provides interfaces for the core functions for managing master role instances.
  • the interfaces for GPMaster module 535 handle election of a dictator from the master role instances, communication between a dictator and server role instances (such as via heartbeat messages), and partition management.
  • the partition management can include providing updates to partition table 559.
  • the interfaces for GPMaster module 535 can perform load balancing using an internal algorithm, or the GPMaster module can receive alternative load balancing expressions 560 that are provided by the user as an application-defined interface.
  • the messaging functions handled by the GPMaster module 535 can be performed by one or more separate messaging interfaces.
  • the GPDictator module 536 provides interfaces that handle functions related to a master role instance that will serve as dictator.
  • the GPDictator interfaces can handle a master role instance claiming dictatorship (such as after winning the dictatorship in an election conducted by GPMaster module 535), releasing dictatorship (such as after a failover), and handling changes to sequence or epoch numbers in order to avoid stale updates.
  • the GPLease module 545 provides interfaces that manage leases within the generic partitioning environment. This can include allowing a master or server to acquire a lease on an associated storage object, on a partition, or on any other type of data structure where leases can be obtained.
  • a partition master can communicate with partition servers via a regularly sent keepalive ('ping') heartbeat.
  • This heartbeat can be sent to all servers, including servers not serving any partitions.
  • a server can respond to this keepalive heartbeat (or other types of heartbeat messages) with a list of the partitions that the server is currently serving.
  • the heartbeat responses can be used in conjunction with server partition leases to allow a master to verify that a server is serving the correct partitions.
  • the server When a server is serving one or more partitions, the server also maintains a lease on its own private storage object, such as a binary object or blob. When a server starts up it does not hold any lease - it will only do so upon the initial partition assignment.
  • the storage object lease should also contain a listing of the partitions being served by a server. If either the heartbeat response or the information in a storage object lease differs from the partition information expected by the master, a conflict in partition information exists.
  • a conflict occurs between a master and a server regarding the partitions being served, and if the server is attempting to serve one or more partitions, the conflict is considered a fatal error.
  • the master may think the server is serving PI and P2 whereas the server reports P2, P4 and P5. In this situation, the master will break the server's lease on the corresponding partition server blob. The master will then issue an alarm and rebuild the proper state for the partition table. This may involve terminating the master so that the proper state is rebuilt upon restart of the master.
  • Another potential failure situation is when a server is holding a lease but the server does not respond to a heartbeat (such as a 'keepalive') within some timeout period.
  • a heartbeat such as a 'keepalive'
  • the master can attempt to delete the server's storage object to resolve the conflict. If the storage object is successfully deleted, then the server is no longer functioning. Any partitions the server was serving can be reassigned to other servers. If the storage object cannot be deleted, the master can read the storage object content to check for conflicts between the master and server regarding partition assignments. If there is no conflict the server may continue serving, and the master can try to 'ping' the server again shortly via the normal heartbeat processing. Preferably, there is a limit to how long the master cannot communicate with a server via heartbeats before it breaks the server's lease on the storage object. If a conflict is detected, the conflict can be handled as described above.
  • partition assignment For partition assignment the master piggy-backs a partition assignment request on the heartbeat to the relevant partition server (the heartbeat is accelerated in this case).
  • a partition assignment request consists of the server storage object name plus the complete new list of partitions to be served.
  • the partition assignment consist of both PI and P2. This makes unassignment orthogonal to assignment: to remove PI from a server serving PI and P2, simply send an assignment consisting of P2 only.
  • the master has already updated the partition table when the partition assignment takes place. The partition table is only updated when (before) a partition is assigned to a server; when a partition is removed from a server no update is needed.
  • a partition server maintains a storage object lease that starts when the first partition is assigned to the server.
  • the storage object lease is maintained until the server dies or the master forces the lease to be broken, such as due to a conflict between the assignment information for the server in the partition table and the assignment information reported by the server to the master.
  • the storage object name is passed along with the partition assignment. All subsequent partition assignments will contain the same storage object name.
  • a server receives a partition assignment, a server without an existing storage object name will acquire the storage object name provided in the partition assignment. If the server already has a storage object with a name, the server can compare the name provided in the request with the existing name. If the names are different, the server can issue an alarm and terminate, as the different names indicate a conflict in partition assignment information.
  • the server can write any associated information into the storage object for the server.
  • the associated information can include, for example, key ranges to serve, the epoch number for the dictator, and/or the epoch number for the partitions.
  • the server can then respond to the master that provided the partition assignment, start serving any new partitions, and stop serving any removed partitions.
  • a master role instance After a partition assignment, a master role instance will expect a response from the server that confirms the assignment. If the response does not match the assignment, or if the response is delayed, the master role instance can terminate the lease for the server. Alternatively, the master role instance can investigate the blob for the server to determine the server status. For example, if the server's storage object indicates that the assignment has succeeded, and if the response is merely delayed or lost instead of inaccurate, the master role instance can wait and see if the server responds correctly to a subsequent heartbeat or other message. If errors are found and the master is not able to break the storage object lease for the server, the master can terminate to force a reconstruction of the partition map by a new master role instance.
  • a server role instance can provide statistics for each range
  • the statistics are opaque to the general partitioning system, and can be represented as a property bag of name/value pairs. These optional statistics can be incorporated into load balancing formulas, including load balancing formulas provided by a user.
  • the master When the master becomes the active master or dictator, it first obtains a new epoch number from the storage object on which it is keeping a lease. It then collects three pieces of information in order to build its view of the system, and to correct any information that is inconsistent.
  • the master reads the partition table.
  • the partition table contains the truth about which partitions exist.
  • the master obtains a lease on at least the relevant part of the partition table prior to reading the partition table in order to prevent stale writes by a previous master.
  • the master obtains a list of all existing server storage objects. This can be done by maintaining a list of server storage objects, by requiring that all server storage objects are located in a specified location, or by another convenient method.
  • a heartbeat or another type of message is also used to query each server regarding the server's current partition assignments. This query can include a query for the name of the storage object for a server. Note that the above tasks can proceed in parallel.
  • the master can identify any inconsistencies between the assignments in the partition table and the assignments reported by each server. If a mismatch is present, one or more corrective actions can be taken, such as breaking the lease of a server on the server object and restarting the master. Additionally, if any server storage objects are identified that are not mentioned in the partition table, these storage objects can be deleted. Finally, any assignments in the partition table that are not confirmed by a server can be queued for assignment to a new partition server. After resolving the conflicts, the master can start normal heartbeat processing, assigning of partitions, and any other master functions.
  • Load balancing can roughly be divided into three activities.
  • Load balancing can include moving a partition from one server to another, splitting a partition into multiple partitions, or merging multiple partitions into a single partition.
  • changing a partition assignment from a first server to a second server will be in response to the first server having a sufficiently high load based on one or more metrics.
  • splitting of the partition can be used to allow a large load to be divided among multiple servers.
  • Merging of partitions allows partitions that have lower amounts of activity to be combined. This reduces the overhead required to track and maintain the various partitions for a data set.
  • a user can define an upper limit on the number of partitions for a namespace.
  • the thresholds for initiating a merge of partitions can be reduced as the number of partitions approaches the upper limit.
  • the upper limit for number of partitions can be dynamically configured.
  • all partitions for a namespace can be sorted based on load.
  • the load can refer to one or more metrics related to performing calculations for a partition.
  • the load can refer to overall CPU usage for a server or an individual partition; storage used for a server or partition; a number of requests received by a server overall or for a single partition; or any other convenient value that indicates the amount of work being performed by a server and/or for a given partition.
  • the top N partitions which have a higher load than a configurable multiple of the average partition load can be split. N is dynamically configurable.
  • the load of each server is calculated by adding together the servers' partition loads.
  • the servers can then be sorted by load, and the top N servers with a load greater than some configurable multiple of the average are chosen for movement of partitions between servers.
  • a plurality of servers with loads below the average load are selected to receive partitions from higher load servers.
  • moving a partition from a first server to a second server is performed to reduce the load of the first server to a value closer to the average load without causing the load of the first server to be below the average load. This is easier to do if all partitions have reasonably similar load. Splitting partitions as described above can be used to reduce the disparity in loads between partitions.
  • the load for a partition can be derived from statistics collected by the server role instances. This information is passed to the master role instance, such as via heartbeat messages at regular intervals. Preferably, the load statistics are defined so that a load metric is additive. This allows the load for a server to be determined based on summing the loads for the individual partitions on the server.
  • One or more formulas for determining the load for a partition and/or a server can be stored in a separate blob or other storage area.
  • the rules or expressions for partitioning can be default rules provided by the generic partitioning environment, or a user can provide rules and/or expressions. Further, it is contemplated that an expression may also be comprise d of a piece of code submitted by an application using the GP system.
  • a user can first identify one or more metrics that are desired as load balancing metrics. Examples of suitable metrics include CPU usage, network bandwidth usage, number of requests processed per time period, or any other convenient metric. Some metrics may be specific to a partition while other metrics may correspond to a value for all partitions on a partition server. Based on the desired metrics, a user then provides one or more interfaces for collecting the desired metric on each server. Optionally, common metrics such as CPU usage or number of requests per time period can be provided as standard interfaces that are simply accessed by a user. The collected metrics are then passed from partition servers to the corresponding master using messages, such as a heartbeat message used by a server to verify the current status of a server with the master.
  • suitable metrics include CPU usage, network bandwidth usage, number of requests processed per time period, or any other convenient metric. Some metrics may be specific to a partition while other metrics may correspond to a value for all partitions on a partition server. Based on the desired metrics, a user then provides one or more interfaces for collecting the desired metric on
  • a series of values recognized by the general partitioning system can be calculated.
  • the general partitioning system can recognize "dimensions" as variables that are defined by a user.
  • a dimension in the general partitioning system can have an expected format.
  • the dimension can include an expression for a PartitionMetric corresponding to the value of the metric over a partition.
  • Another expression can be for a ServerMetric, which corresponds to the value of the metric over all partitions on the server.
  • Still another expression can be for a Condition value for the metric, which defines a situation where an action is taken.
  • CPU usage can be defined as a dimension by a user.
  • the CPU usage dimension is used to determine when a server is sufficiently busy that a partition should be moved to another server.
  • the percentage of CPU usage on a virtual machine dedicated for processing requests for a given partition is defined as the PartitionMetric.
  • the sum of CPU usage percentages over all partitions on the virtual machine can be defined as the ServerMetric.
  • the Condition can be defined as ServerMetric usage being greater than 80% of total CPU usage. When this Condition occurs, a partition can be moved to another server. The partition for movement is selected based on the PartitionMetric. It is noted that the PartitionMetric and ServerMetric are both defined by the user. Thus, a user is not required to have a PartitionMetric that is analogous to the ServerMetric.
  • the ServerMetric could be a combination CPU usage and network bandwidth usage, while the PartitionMetric is only related to request rate.
  • a user can also define dimensions for triggering a partition split.
  • the definition of the dimension for triggering a partition split can be similar to the dimension for a reassigning a partition, or a different format of dimension can be used.
  • a ServerMetric expression may not be necessary for a dimension for triggering a partition split, since a PartitionMetric expression will more likely be useful for determining when to split a partition.
  • a dimension for triggering a partition split could include a dimension for how to split a partition when a Condition for splitting the partition is satisfied. It is noted that the dimensions for triggering a partition split may also be useful for identifying when to merge two partitions. Alternatively, a user can define separate dimensions for partition merging.
  • any convenient number of conditions can be specified for determining when a load balancing action should occur.
  • the conditions can incorporate metrics corresponding to one or more dimensions, such as metrics corresponding to a plurality of dimensions.
  • the conditions can be ordered, so that the conditions are evaluated in a specific order for performing load balancing. For example, a condition related to splitting a partition can be placed earlier in the order than a condition for moving partitions to different partition servers. In such an example, if one partition on a partition server is responsible for a large percentage of the load, moving other partitions might not be effective for balancing load among a plurality of servers. By checking for whether to split the partition first, the partition causing the problem can be split into portions with (presumably) lower loads.
  • partition server with a large number of low load partitions might appear to be unavailable due to an excessive number of partitions, even though the overall load on the server is below average. Merging partitions before moving assignments allows such a server to become available for receiving an additional partition assignment.
  • any convenient type of metric may be included in a condition.
  • metrics for load on a single partition, load on a plurality of partitions, load on a server, or load on a plurality of servers may be used together or separately as desired.
  • metrics related to a plurality of partitions or plurality of servers a simple example is to determine the load across all servers in order to define an average load.
  • a condition for performing load balancing could then be related to a difference between load on a server versus average server load, such as a difference in the absolute value of the server load versus average load, or a comparison of the server load with a standard deviation from the average server load.
  • the desired load balancing action may be different for a partition server with multiple partitions that have a similar load as opposed to a partition server with only one high load partition.
  • a user can also define one or more filters for constraining actions on partitions based on the dimensions. For example, it may be desirable to prevent a server from a receiving a new partition assignment when the server has either a CPU usage ServerMetric of more than 70% or the number of partitions is greater than 10.
  • An AssignmentFilter value can be defined that prevents such assignments.
  • Other examples of potential filters include filters for preventing the movement of a partition from an existing partition server, preventing a split of a partition, or preventing a merge of a partition.
  • a filter may prevent a load balancing action from occurring.
  • a filter can modify the order of consideration of conditions, or the filter can cause a condition to be skipped entirely during a load balancing calculation cycle.
  • each server role instance collects an average of the request rate over a longer time period (RR SlowMA) and an average of the request rate over a shorter time period (RR FastMA). These request rate averages are transmitted back to the master role instance as name/value pairs in a property bag.
  • the dimension "load” would be identified in advance by the user.
  • the right hand side can be any arbitrary expression that generates a load value that is assigned to the PartitionMetric.
  • the load corresponds to the number of requests based on one of a plurality of moving averages.
  • one or more Conditions can be defined regarding whether partitions should be split, merged, or moved between servers.
  • a user can define any combination of metrics and expressions (such as conditions) for making load balancing decisions.
  • the expressions defined by a user for making a load balancing decisions can also include conditional logic and/or support for multi-dimensional constraints/optimization goals.
  • a user can define a decisions tree or use other conditional logic to provide an ordering for how and when expressions are evaluated. For example, a user could have a first expression evaluated, and then select from a plurality of potential additional expressions to evaluate based on the value of the first expression. This could be based on an "if-then-else" type of conditional logic, a lookup table for the next expression based on a determined value, or any other convenient type of conditional logic.
  • a user has flexibility to specify the types of metrics and expressions for use in load balancing, including whether a given expression is evaluated as well as providing an order for evaluating such expressions. Note that the order of evaluation for expressions can also be determined dynamically based on the value of previously evaluated expressions.
  • Another example of expressions defined by a user for making load balancing decisions may include multi-dimensional constraints/optimization goals.
  • a user may defined multiple dimensions (e.g., two), and for each dimensions the user may define a separate optimization goal or constraint.
  • CPU utilization and request latency are two exemplary dimensions.
  • the user may specify rules such that a partition servers' CPU usage is below a first threshold (e.g., 90%) and at the same time minimize an average request latency across a defined set of partition servers (e.g., all partition servers).
  • This approach may differ from an if-then-else type logic where users specify exactly what to do.
  • users define limits and optimization goals, which allow the system to find a solution automatically.
  • a user can provide a plurality of conditions or expressions for evaluation.
  • the expressions are based on various CPU usage metrics for the partition servers that a serving a given namespace.
  • a first expression evaluates whether any partition server has a CPU usage related to the namespace of greater than 60%. For this user, if no CPU has a usage greater than 60%>, the user does not desire load balancing. Thus, if the result of the first expression is false (i.e., no partition server has a CPU usage greater than 60%), then no further expressions are evaluated, as load balancing is not desired. If at least one partition server has a CPU usage greater than 60%, a series of expressions can then be evaluated to determine a load balancing action to perform.
  • a partition can be moved from a first server to a second server by having the master role instance issue two assignment requests.
  • An assignment request to the first server does not include the partition, which results in the first server stopping service for the partition.
  • a second assignment request to the second server includes the partition.
  • a master role instance can initiate a split by determining a split key, which corresponds to a key value that will form the end of the inclusive range for one of the new partitions.
  • the split key can be selected in any convenient manner.
  • a master or server role instance can select a split key based on the partition, such as by selecting a key value at or near the middle of the range of the partition.
  • a server can select a split key based on additional statistics regarding the partition. For example, a sampling-based bucket mechanism could be used to track the load for various portions of the partition in a manner similar to determining the load for the full partition. The split key can then be selected so that the load is similar for the buckets assigned to the new partitions.
  • the (active) master is responsible for spreading the load across the partition servers.
  • the master will prevent one or more servers from becoming overloaded and thus unable to process requests.
  • adjustments of load could be done by keeping one partition/range per server and adjusting these ranges. By instead moving partitions, an adjustment to the load can be performed while impacting a smaller number of servers.
  • partitions per server It is often desirable to have a minimum number of partitions per server so that load can be moved around smoothly via partition reassignments. When the number of partitions falls to the minimum level, further mergers are not performed. Similarly, it is often also desirable to avoid having too many partitions. As the maximum number of partitions is approached for a server, the likelihood of merging partitions can increase. As an example, it may be desirable to maintain between 5 and 8 partitions per server. Of course, the various embodiments of the invention can work with any number of partitions per server, such as from as few as 1 partition per server to hundreds or more per server.
  • both the split and merge protocols are stateless. Either the master or the involved server(s) may fail over at any time without causing an error for the partitioning system. In other words, if either the master or a server fails during the split or merge process, the next master or server will be able to construct a valid list of partition assignments regardless of when the failure occurred.
  • the participating server is not required to perform any of the split actions.
  • a partition table can include a partition on a server S 1 that ranges from a low key value of D to a high key value of H. In this example, the epoch number for the partition is 2.
  • the partition should be split, so that part of the partition can be assigned to another server.
  • the master role instance asks server SI for a split key.
  • the server SI returns a key of G as the split key.
  • the master modifies the partition table.
  • the table now contains two partitions. One has a low key value of D and a high key value of G, while the second partition has a low key value of G and a high key value of H.
  • the partition range definitions based on a low key value and high key value are inclusive of the low key value and exclusive of the high key value.
  • the change in the partition table can occur by modifying the existing entry and adding a new entry, by removing the existing entry and adding two new entries, or by any other convenient method.
  • the master detects a conflict between the partitions served by server SI and the information in the partition table. Because a split has just occurred the master does not terminate the blob lease of the server S 1. Instead, the master sends an assignment to server S 1 with a partition range of D to G and an epoch of 3. This modifies the assignment of the partition at SI to match one of the split partitions in the partition table. After receiving an acknowledgment of the new assignment from server SI, the master can assign the second split partition to another server. The second split partition would also have an epoch number of 3. Alternatively, both of the split partitions can be assigned to server S 1 initially, with one or both partitions being moved at a later time to perform load balancing.
  • partitions for merger are unassigned from the current server.
  • a first partition on server S2 can have a low key value of K and a high key value of M.
  • the epoch number for the first partition is 7.
  • a second partition on server S4 can have a low key value of M and a high key value of N.
  • the epoch value for the second partition is 9 in this example.
  • the partitions can be unassigned from their respective servers, so that the partition table shows a non-assigned value for the server.
  • the two partition entries are then replaced with a single entry having a low key of K and a high key of N.
  • the epoch number assigned to this partition is one greater than the highest value of the merged partitions, which corresponds to 10 in this example.
  • the new partition can then be assigned to a server. Additional Examples
  • a "client” is defined as a role that issues one or more requests for action by a application-defined interface against a namespace or domain.
  • a client can correspond to a user or to a process initiated on behalf of a user. For example, a request to a lookup a particular account corresponds to a request directed to an application for account lookup that is made against the domain of all accounts with a key corresponding to the desired account.
  • a "work item” is a static representation of a job to be run in the cloud computing environment.
  • a work item can specify various aspects of a job, including job binaries, pointers to the data to be processed, and optionally the command line to launch tasks for performing the job.
  • a work item may specify the reoccurrence schedule, priority and constraints. For example, a work item can specify to be launched every day at 5 PM.
  • a "job” is a running instance of a work item.
  • a job contains a collection of tasks that work together to perform a distributed computation.
  • the tasks can run on one or more virtual machines in the cloud computing environment.
  • a "task” is the fundamental execution unit of a job. Each task runs on a virtual machine. Users can specify additional input to the command line and pointers to input data for each task.
  • a task may create a hierarchy of files under its working directory on the virtual machine performing the task during the course of execution of the task.
  • FIG. 1 schematically shows an example of a system suitable for performing tasks within a cloud computing environment.
  • the system in FIG. 1 includes a task runtime layer 110, a third party task runtime layer 120, a resource management layer 130, and a scheduling and execution layer 140.
  • the task runtime layer 110 is responsible for setting up the execution environment and security context for tasks from a user 105.
  • the task runtime layer 110 can also launch tasks and monitor the status of the tasks.
  • the task runtime layer 110 can take the form of a system agent running on each virtual machine.
  • the task runtime layer may also include a runtime library that can be linked into a users' task executables. Having runtime libraries as part of the task runtime layer 110 can potentially provide richer capability to tasks executed by the system agent. Examples of runtime libraries include one or more efficient communication libraries to allow fast communication among tasks; an efficient remote file access library support to read files from other virtual machines and/or other tasks; a checkpoint library to allow tasks to checkpoint (e.g. into binary large objects) and resume; a logging library; and a library for providing a distributed file system to be used across virtual machines performing a given task within a pool of virtual machines.
  • the third party task runtime layer 120 allows additional runtimes to be built and run on top of task runtime layer 110.
  • the third party task runtime layer 120 also can provide additional capabilities for coordinating the running of tasks for a job. Examples may include a MapReduce runtime to a library for providing a distributed file system to be used across virtual machines performing a given task within a pool of virtual machines. This allows a user to organize the cloud computing environment in a manner tailored for the user's jobs or tasks.
  • a job manager task can facilitate allowing a user to use a third party runtime layer to run and/or control cloud computing resources.
  • Resource management layer 130 deals with managing the computing resources available in the cloud computing environment.
  • One option is to have the resource management layer 130 manage the resources at three different levels.
  • the resource management layer 130 manages the allocation and deallocation of virtual machines associated with a job (i.e., execution of a work item) as well as the files stored on each virtual machine associated with a task.
  • the virtual machines associated with a job can be grouped into pools of machines.
  • a pool can contain virtual machines associated with one or more jobs and/or work items.
  • a single pool can span across multiple virtual machine clusters, such as all virtual machine clusters in a data center, a plurality of virtual machine clusters across a plurality of data centers within a geographic region, or a plurality of clusters across data centers in a plurality of geographic regions.
  • a single pool can contain a large number of virtual machines, such as millions.
  • the virtual machines can be contained in a large number of pools, such as up to billions.
  • the resource management layer manages the amount of virtual machines available for association with jobs or work items in a given group of pools. This allows for dynamic adjustment of the amount of compute resources used based on the current load of the system. Additionally, virtual machines that are not being used by a current group of pools may be released back to the cloud computing environment for incorporation into other groups of pools.
  • scheduling and execution layer 140 manages work items, jobs, and tasks that are being performed by a user.
  • the scheduling and execution layer 140 makes scheduling decisions and is responsible for launching jobs and tasks as well as retries on failures.
  • Such a scheduling and execution layer 140 can include components for managing jobs and/or tasks at various levels.
  • FIG. 2 schematically shows an example of how processors at different locations can be integrated within a single cloud computing architecture.
  • one or more task tenants 215 can be used to manage pools of virtual machines.
  • a task tenant 215 can maintain a set of virtual machines.
  • the jobs of one or more users can run on the virtual machines within a task tenant 215 as part of one or more pools of virtual machines.
  • One or more task tenants 215 can be used in a given geographic region.
  • the responsibilities of a task tenant 215 can include maintaining the set of virtual machines and dynamically growing or shrink the task tenant based on the resource utilization within the task tenant. This allows a task tenant 215 to increase the number of virtual machines within the task tenant to accommodate increased customer demand. This also allows a task tenant 215 to release unused virtual machines so that the virtual machines can be allocated to other hosted services in the data center handling service for other customers.
  • Another responsibility of a task tenant 215 can be implementing part of the pool allocation/deallocation/management logic. This allows the task tenant 215 to participate in determining how virtual machines are assigned to pools associated with a task for a customer. The task tenant 215 can also be responsible for scheduling and execution of tasks on the virtual machines within the task tenant.
  • one or more task location services 225 are provided that control a plurality of task tenants 215.
  • the plurality of task tenants can correspond to all task tenants in a given geographic region, various task tenants from around the world, or any other convenient grouping of task tenants.
  • task location services 225 are shown that serve regions labeled "US North" and US South".
  • the responsibilities of a task location service 225 can include management of task accounts for the given geographic region.
  • the task location services 225 can also provide application programming interfaces (APIs) for allowing users to interact with the cloud computing environment.
  • APIs can include handling APIs associated with pools of virtual machines, pool management logic, and coordination of pool management logic across task tenants within a given geographic region.
  • the APIs can also include APIs for handling tasks submitted by a user, as well as maintaining, scheduling, and terminating work items or jobs associated with the user tasks.
  • the APIs can further include APIs for statistics collection, aggregation, and reporting for all work items, jobs, tasks, and pools in a geographic region.
  • the APIs can include APIs for allowing auction of available virtual machines as preemptible virtual machines to users on a short term basis based on a spot market for virtual machines.
  • the APIs can also include APIs for metering usage and providing billing support.
  • the task location services 225 can be linked together by a global location service 235.
  • the global location service 235 can be responsible for account creation and management of accounts, including managing task accounts in conjunction with the task location service tenants 225. This includes being responsible for disaster recovery and being responsible for availability of work items and jobs if there is a major data center disaster. This may include running a work item or job in a different location due to a data center not being available for any reason. This can also include allowing customers to migrate their work items, jobs, and pools from one data center to another data center. Typically there will be only one active global location service 235. This active global location service 235 is in communication with the various task location services 225 as well as service components for managing data storage (not shown). The global location service can maintain a global account namespace 237.
  • FIG. 3 shows a potential configuration for a task location service.
  • a task location service can include one or more account servers 321.
  • the account servers handle account management for accounts in a given geographic region, including creation, deletion, or property updates.
  • Account front ends 322 serve as the front end nodes for account service.
  • the account front ends 322 are behind an account virtual IP address 324 as shown in the figure.
  • the account front ends 322 process the account API requests coming from global location service, such as API requests to create accounts or delete accounts.
  • the configuration in FIG. 3 also includes one or more pool servers 331.
  • a pool server 331 handles pool management and pool transactions for pools of virtual machines in a given geographic region.
  • a pool server 331 handles pool creation, deletion and property updates.
  • a pool server 331 also manages the high level virtual machine allocation algorithm across multiple task tenants. Virtual machine allocation can take into consideration the connectivity of a virtual machine with storage for a given user. The pool server may also perform other tasks related to allocation of virtual machines.
  • the configuration in FIG. 3 also includes one or more work item or job servers (WIJ) 336.
  • WIJ servers 336 handle creation, deletion, and updates of work items and jobs.
  • the WIJ servers 336 may initiate the creation and deletion of pools associated with the work items or jobs.
  • the WIJ servers 336 also use generic partitioning mechanisms for scaling. In an embodiment, there are multiple WIJ servers 336 in each task location service, and each of the WIJ servers handles a range of work items.
  • the pool servers 331 and WIJ servers 336 receive requests from users via task location service front ends 338.
  • the task location service front ends 338 are also responsible for calling corresponding components to process requests from users.
  • the task location service front ends 338 are behind an account virtual IP address 334 as shown in the figure.
  • the configuration in FIG. 3 further includes a task location service master
  • the task location service master 342 has two main responsibilities.
  • the task location service master 325 serves as a master system for implementing partitioning logic for the corresponding servers in a task location service 225.
  • the task location service master 342 can be responsible for computing the new market price for preemptible virtual machines at the beginning of each spot period for the entire geographic region of the task location service. It can collect current bids and resource availability information from the pool servers and task tenants, and computes the new market price accordingly.
  • the task location service master can send the bid and resource availability information to a spot price market service. It also makes high level allocation guidance to pool servers about preemptible virtual machines across all task tenants in a geographic region.
  • a task location service master 342 can communicate with one or more statistics aggregation servers 355.
  • the statistics aggregation servers are responsible for collecting and aggregating detailed statistics for tasks, jobs, work items and pools.
  • the other components in the system emit fine-grained statistics for tasks and virtual machines.
  • the statistics aggregation servers aggregate these fine-grained statistics from task level or virtual machine level statistics into work item, account level, and/or pool level statistics.
  • the statistics can be exposed for use via an API.
  • the statistics aggregation servers can be responsible for generating hourly metering records for each account for use in billing.
  • FIG. 3 The top level role demonstrated in FIG. 3 is a task location service or tenant. If multiple instances of the task location service are present, one of the instances will correspond to task location service master (or dictator) 342. Within the tenant are an account server role 321, a pool server role 331, and a work items-jobs server role 336. Each of these roles also represent instances of the task location service, but these role instances handle a set of functions within the overall tenant. For example, a request for account information is handled by the account server role within the tenant. If multiple instances of the task location service or tenant are present, the master for each of the roles within the tenant can correspond to a different instance.
  • FIG. 6 shows a conventional example of how redundancy can be provided for multiple master roles.
  • each master role needs to have two additional instances in order to improve availability.
  • a fault domain includes nodes that have common failure patterns and they may fail together. For example, nodes on the same rack sharing the same power supply may be in a common fault domain as they may fail as the result of a common problem.
  • An upgrade domain corresponds to a set of nodes that may be taken offline at the same time during a system upgrade.
  • FIG. 8 shows an example of how virtual machines in a distributed computing environment using generic partitioning can be organized to provide the various instances and masters for a role.
  • separate GP Masters 821, 831, and 841 are shown for account servers, pool servers, and WIJ servers respectively. Since the GP Master module and any fixed interfaces are the same regardless of the role being managed, the backup server needs for the GP Masters 821, 831, and 841 can be combined on a single machine. Thus, a single backup 852 can be provided as the backup for the three GP Masters. If one of the GP Masters 821, 831, or 841 experiences a failure, the same GP Master module and fixed interfaces may be used.
  • GP Master backup The only additional information types, in this example, needed by the failover backup to take over the GP Master role that experienced a failure are the key for the corresponding namespace and any application- defined interfaces.
  • a single second backup 853 can be used for all three of GP Masters 821, 831, and 841. Consequently, in this example, only two GP Master backup servers (852 and 853) are used for the at least three GP Master roles. Although three GP Master roles are shown as being backed up by common machines, any convenient number of GP Master roles belonging to the same user or account could be backed up using a common machine.
  • FIG. 12 shows an exemplary aspect of providing backup machines for master roles relative to fault domains and upgrade domains, in accordance with aspects of the present invention. Similar to those concepts discussed hereinabove with respect to FIG. 8, multiple GP Master roles may be backed up on a fewer number of servers.
  • FIG. 12 depicts an Account GP Master 1202 in a first fault domain and a first upgrade domain, a Pool GP Master 1204 in a second fault domain and in a second upgrade domain, a WIH GP Master 1210 in a third fault domain and in a third upgrade domain, a first GP Backup 1206 and a second GP Backup 1208.
  • the first GP Backup 1206 and the second GP Backup 1208 are each in different fault domains and upgrade domains from the GP Master roles.
  • a single Generic Partitioning system which in this example only requires five servers (or potentially four servers with a single backup) to host all of the masters for the three roles.
  • nine different servers may be required for those same three master roles.
  • the approach illustrated in FIG. 8 may be accomplished through the utilization of two additional servers that can be used for any type of role being hosted by the system. Consequently, a backup server (e.g., GP Backup 1206) may be used if one or more of the master roles are unavailable do to a failure of a fault domain or an unavailable upgrade domain.
  • computing device 700 is but one example of a suitable computing environment and is not intended to suggest any limitation as to the scope of use or functionality of the invention. Neither should the computing device 700 be interpreted as having any dependency or requirement relating to any one or combination of components illustrated.
  • Embodiments of the invention may be described in the general context of computer code or machine-useable instructions, including computer-executable instructions such as program modules, being executed by a computer or other machine, such as a personal data assistant or other handheld device.
  • program modules including routines, programs, objects, components, data structures, etc., refer to code that perform particular tasks or implement particular abstract data types.
  • the invention may be practiced in a variety of system configurations, including hand-held devices, consumer electronics, general-purpose computers, more specialty computing devices, and the like.
  • the invention may also be practiced in distributed computing environments where tasks are performed by remote-processing devices that are linked through a communications network.
  • computing device 700 includes a bus
  • Bus 710 that directly or indirectly couples the following devices: memory 712, one or more processors 714, one or more presentation components 716, input/output (I/O) ports 718, I/O components 720, and an illustrative power supply 722.
  • Bus 710 represents what may be one or more busses (such as an address bus, data bus, or combination thereof).
  • busses such as an address bus, data bus, or combination thereof.
  • FIG. 7 is merely illustrative of an exemplary computing device that can be used in connection with one or more embodiments of the present invention. Distinction is not made between such categories as “workstation,” “server,” “laptop,” “hand-held device,” etc., as all are contemplated within the scope of FIG. 7 and reference to “computing device.”
  • the computing device 700 typically includes a variety of computer- readable media.
  • Computer-readable media can be any available media that can be accessed by computing device 700 and includes both volatile and nonvolatile media, removable and non-removable media.
  • Computer- readable media may comprise computer storage media and communication media.
  • Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data.
  • Computer storage media includes, but is not limited to, Random Access Memory (RAM), Read Only Memory (ROM), Electronically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other holographic memory, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to encode desired information and which can be accessed by the computing device 700.
  • RAM Random Access Memory
  • ROM Read Only Memory
  • EEPROM Electronically Erasable Programmable Read Only Memory
  • flash memory or other memory technology
  • CD-ROM compact discs
  • DVD digital versatile disks
  • magnetic cassettes magnetic tape
  • magnetic disk storage magnetic disk storage devices
  • Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and includes any information delivery media.
  • modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
  • communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer-readable media.
  • the memory 712 can include computer-storage media in the form of volatile and/or nonvolatile memory.
  • the memory may be removable, non-removable, or a combination thereof.
  • Exemplary hardware devices include solid-state memory, hard drives, optical-disc drives, etc.
  • the computing device 700 includes one or more processors that read data from various entities such as the memory 712 or the I/O components 720.
  • the presentation component(s) 716 present data indications to a user or other device.
  • Exemplary presentation components include a display device, speaker, printing component, vibrating component, and the like.
  • the I/O ports 718 can allow the computing device 700 to be logically coupled to other devices including the I/O components 720, some of which may be built in.
  • Illustrative components can include a microphone, joystick, game pad, satellite dish, scanner, printer, wireless device, etc.
  • FIG. 9 shows an example of a method according to the invention.
  • a namespace is received 910 from a client application.
  • the client application can correspond to any convenient type of computation that a client desires to perform in a distributed computing environment.
  • the namespace comprises a plurality of key values that span the namespace.
  • a plurality of partitions based on the namespace are assigned 920 to one or more partition servers.
  • the partition servers can implement one or more application-defined partitioning system interfaces.
  • One or more metrics are received 930 from the partition servers based on activity related to the assigned partitions.
  • One or more application-defined load balancing conditions are also received 940.
  • a load balancing action is subsequently performed 950 based on a corresponding load balancing conditions being satisfied.
  • the load balancing action can include a change in a partition assignment for at least one partition server, a split of a partition, a merger of two partitions, or a combination thereof.
  • FIG. 10 shows another example of a method according to the invention.
  • a namespace is received 1010 from a client application.
  • the client application can correspond to any convenient type of computation that a client desires to perform in a distributed computing environment.
  • the namespace comprises a plurality of key values that span the namespace.
  • An application-defined interface is also received 1020.
  • the application-defined interface comprises a plurality of expressions corresponding to load balancing conditions and one or more conditional logic expressions.
  • a value is determined 1030 for at least one conditional logic expression.
  • One or more of the plurality of expressions corresponding to load balancing conditions are then selected 1040 based on the determined value of the at least one conditional logic expression. After evaluation, it is determined that at least one load balancing condition corresponding to a selected expression is satisfied 1050.
  • a load balancing action corresponding to the satisfied condition is then performed 1060.
  • the method may further be comprised of an evaluation of a plurality of load balancing conditions or expressions or procedures without a predetermined order to reach an optimization target while still obtaining the constraints specified in the load balancing conditions.
  • FIG. 11 shows yet another example of a method according to the invention.
  • a namespace is received 1110 from a client application.
  • the client application can correspond to any convenient type of computation that a client desires to perform in a distributed computing environment.
  • the namespace comprises a plurality of key values that span the namespace.
  • a plurality of partitions based on the namespace are assigned 1120 to one or more partition servers.
  • the partition servers can implement one or more application-defined partitioning system interfaces.
  • One or more metrics are received 1130 from the partition servers based on activity related to the assigned partitions.
  • a plurality of application-defined load balancing conditions are also received 1140.
  • the plurality of application-defined load balancing conditions include an ordering for evaluation of the conditions.
  • the plurality of application-defined load balancing conditions are evaluated 1150 according to the received ordering.
  • a load balancing action is performed 1160 that corresponds to a received load balancing condition that is satisfied.
  • a method for performing load balancing in a distributed computing environment.
  • the method includes receiving a namespace from a client application comprising a plurality of key values that span the namespace; assigning, by a fixed partitioning system interface, a plurality of partitions based on the namespace to one or more partition servers, the plurality of partitions corresponding to key ranges that span the namespace, the partition servers implementing one or more application-defined partitioning system interfaces; receiving one or more metrics from the partition servers based on activity related to assigned partitions; receiving one or more application-defined load balancing conditions, a load balancing condition comprising an expression that can be evaluated based on at least one received metric; and performing, by a fixed partitioning system interface, a load balancing action corresponding to a received load balancing condition, the load balancing action being performed based on the corresponding load balancing condition being satisfied, the load balancing action including a change in a partition assignment for at least one partition server.
  • a method for performing load balancing in a distributed computing environment. The method includes receiving a namespace from a client comprising a plurality of key values that span the namespace; receiving an application-defined interface comprising a plurality of expressions corresponding to load balancing conditions based on one or more of the application- defined metrics and one or more conditional logic expressions; determining a value for at least one conditional logic expression; selecting one or more of the plurality of expressions corresponding to load balancing conditions for evaluation based on the determined value for the at least one conditional logic expression; determining that at least one load balancing condition corresponding to the selected one or more expressions is satisfied; and performing, by a fixed partitioning system interface, a load balancing action corresponding to the satisfied condition.
  • a method for performing load balancing in a distributed computing environment includes receiving a namespace from a client application comprising a plurality of key values that span the namespace; assigning, by a fixed partitioning system interface, a plurality of partitions based on the namespace to one or more partition servers, the plurality of partitions corresponding to key ranges that span the namespace, the partition servers implementing one or more application-defined partitioning system interfaces; receiving one or more metrics from the partition servers based activity related to assigned partitions; receiving a plurality of application-defined load balancing conditions, a load balancing condition comprising an expression that can be evaluated based on at least one received metric, the plurality of load balancing conditions further comprising an ordering for the conditions; evaluating the plurality of application-defined load balancing conditions according to the received ordering; and performing, by a fixed partitioning system interface, a load balancing action corresponding to a received load balancing condition, the load balancing action

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Computer And Data Communications (AREA)

Abstract

Systems and methods are provided that enable a general framework for partitioning application-defined jobs (e.g., computation) in a scalable environment. The general framework decouples partitioning of a job from the other aspects of the job. As a result, the effort required to define the application-defined computation in a scalable environment is reduced or minimized, as the user is not required to provide a partitioning algorithm. The general framework further allows a user to provide load balancing conditions to allow for modification of how partitions are assigned.

Description

DYNAMIC LOAD BALANCING IN A SCALABLE ENVIRONMENT
BACKGROUND
[0001] Cloud computing environments provide a potential platform for allowing users to have access to large amounts of computing resources without having to invest in corresponding infrastructure. Instead, the computing environment can be provided as a service by a cloud computing provider. This can allow a user to tune the requested computing resources to match the size and importance of a computing job. However, taking full advantage of the additional computing resources available as a service may require dividing computing tasks into smaller portions, so that multiple physical or virtual processors can be used to perform a task.
SUMMARY
[0002] In various embodiments, systems and methods are provided that enable a general framework for partitioning application-defined jobs in a scalable environment. The general framework decouples partitioning of a job from the other aspects of the job. As a result, the effort required to define the application-defined job is reduced or minimized, as the user is not required to provide a partitioning algorithm. The general framework further allows a user to provide load balancing conditions to allow for modification of how partitions are assigned.
[0003] This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This
Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid, in isolation, in determining the scope of the claimed subject matter.
BRIEF DESCRIPTION OF THE DRAWINGS
[0004] The invention is described in detail below with reference to the attached drawing figures, wherein:
[0005] FIGS. 1 to 3 schematically show various aspects of an example of a distributed computing environment, in accordance with aspects of the present invention;
[0006] FIG. 4 schematically shows the interaction between various interfaces for performing a job in a generic partitioning environment, in accordance with aspects of the present invention;
[0007] FIG. 5 shows an example of a generic partitioning environment, in accordance with aspects of the present invention; [0008] FIG. 6 shows an example of providing backup machines for active master roles among, in accordance with aspects of the present invention;
[0009] FIG. 7 is a block diagram of an exemplary computing environment suitable for use in implementing embodiments of the present invention;
[0010] FIG. 8 shows another example of providing backup machines for master roles, in accordance with aspects of the present invention;
[0011] FIGS. 9 to 11 show examples of methods according to various embodiments of the invention; and
[0012] FIG 12 shows an exemplary aspect of providing backup machines for master roles relative to fault domains and upgrade domains, in accordance with aspects of the present invention.
DETAILED DESCRIPTION
Overview
[0013] Due to increases in the speed of data transmission over networks and improvements in other network features, it is increasingly possible to perform large scale computing tasks in an environment where computing resources are distributed over a large network. A user in a first location can submit a job or computing task to a computing service and have the task performed on a group of computers that the user has no direct knowledge of. The computing resources for performing the user's task may be distributed over multiple locations. A first group of computing resources located in one or more locations can store the data and other information for performing the user's computing task, while a second group of computing resources, in the same locations or possibly in a different set of one or more locations, can be used to perform the computing task.
[0014] Access to a variety of distributed computing resources allows a user to perform job tasks without concern for where the computing resources are located. The distributed resources also provide an opportunity for a user to scale up (or scale down) the amount of resources used in order to meet goals for a computing task, such as completing the computing task by a specified time. However, using distributed computing resources poses a number of challenges for a user. Conventionally, a user or application designer would need to divert resources away from designing an application or job in order to determine how to take advantage of a specific distributed computing environment.
[0015] In various embodiments, systems and methods are provided that enable a generic framework for building scalable applications in a distributed environment, such as a cloud computing environment. The generic framework can allow the partitioning of jobs or work items to be decoupled from performing computations associated with the jobs or work items. This allows the owner of a distributed environment to provide resources to a user in an accessible or simplified manner while still providing a high availability for the distributed resources. The generic framework reduces the complexity of designing applications for a distributed environment by providing a framework for handling features such as scalability, fault tolerance, and/or availability while reducing or minimizing the amount of effort required to address these features. As a result, the framework can allow an application designer to focus on the application instead of spending additional time on requirements of the distributed environment.
Definitions
[0016] A "key" is one of the basic concepts used in a generic partitioning framework. A key is a value from a namespace or domain. An example of a namespace is an identifier corresponding to all storage accounts in a cloud computing environment. In such an example, a key can correspond to an account name, account number, or another identifier that allows a specific account to be referenced. Another example of a namespace is the range of possible alphanumeric values for an input parameter for a job. Preferably, the keys that span a namespace will also have some method of organization so that keys can be described as a range of serial values. For example, the keys can be organized numerically, alphabetically, based on hashed values, or in any other convenient serial manner that allows a range of keys to be defined based on specifying two keys as a range beginning and a range end.
[0017] A "partition" is a range defined by a low (inclusive) and high (exclusive) key. A partition can never be an empty range. If desired, a single partition may contain the entire range of the domain. Partitions are defined to be mutually exclusive, so that there is not an overlap between the ranges of two different partitions. The union of all partitions will span the entire domain or namespace.
[0018] A "partition server" is a virtual machine within a cloud computing environment that corresponds to a role instance for serving zero or more partitions. A partition is not served by more than one partition server at the same time. However, a given partition server may not have any partitions that are currently being served by the partition server. A partition server can include both application-defined interfaces and fixed interfaces (defined by the partitioning system) for performing various actions.
[0019] A "partition master" is a role that manages partition servers for a given type of role, such as by assigning and unassigning partitions to partition servers. The partition master role also monitors partition servers, such as by verifying that the partition servers are still serving assigned partitions. Typically, this role is redundant for fault tolerance. A partition master can include both application-defined interfaces and fixed interfaces (defined by the partitioning system) for performing various actions.
[0020] An "application-defined interface" refers to a computation, operation, or other function defined by a client for performance by a role instance. An application- defined interface can be created for either a partition master or partition server role. An application-defined interface is in contrast to a "fixed interface" for a partitioning system. A fixed interface refers to an application interface that is provided as part of the partitioning system. A client cannot modify the action of a fixed interface. However, an application-defined interface can be used in conjunction with a fixed interface. For example, a fixed interface for a master role can be to send a heartbeat message every time period to verify the status of any servers being managed by the master role. A corresponding fixed interface for a server role can be an interface to respond to the heartbeat message in an appropriate manner. An example of an application-defined interface can be an interface that adds additional information to a heartbeat message, while another example can be an interface to extract such additional information from a heartbeat message. In this situation, the sending of the heartbeat message itself is a fixed interface. A user or application cannot modify the protocol for sending the message. However, the interface for defining the message content can be modified by a user. The basic structure of the interface for supplementing the information in a heartbeat message may be provided by the system. However, since a client can modify the content provided by this interface, such an interface is defined herein as an application-defined interface.
[0021] As another example, application-defined interfaces can be used in conjunction with fixed interfaces to provide the overall feature of load balancing. The interfaces for assigning partitions to a partition server or splitting a partition into two partitions are fixed interfaces. However, application-defined interfaces can provide expressions for when load balancing actions should occur, like changing partition assignments between machines or deciding when to split a partition.
[0022] Each active partition master role or partition server role has a corresponding storage object that controls the data corresponding to the role. An example of a storage object is a binary large object or blob. For a partition server, the storage object includes the identity of the partitions being served. Note that the storage object for a partition server will typically not contain the underlying data corresponding to the partition. By only storing partition identities while leaving the underlying data in a separate data storage location, partitions can be moved from one server to another with a minimum of data transfer. For a partition master, the storage object includes information for establishing that a particular instance of the partition master is the active instance. The storage object can optionally also include information regarding all storage objects for servers being managed by the master. During operation, master and server roles can maintain a lease on the corresponding storage object. When a lease on a storage object for a role is broken, the corresponding role can be terminated.
[0023] As noted above, partition masters for a given type of role are preferably redundant, so that at least one additional partition master is available if a failure occurs. A
"dictator" is defined as the partition master that current performs the partition master functions for a given type of role. The dictator is the partition master that maintains the lease for the storage object associated with the given partition master role.
[0024] The active partition master (dictator) can communicate with the partition servers via heartbeats. The basic heartbeat is of type 'keepalive' and is always used. As noted above, an application-defined interface can be used to add optional commands and/or information to this heartbeat message.
Namespaces, Keys, and Partitions
[0025] In various embodiments, a generic partitioning framework is provided so that a user or application can define a computation that is automatically partitioned for processing in a distributed computing environment. To take advantage of the generic partitioning framework, a user can define a computation based on a namespace. A namespace corresponds to one or more computations or calculations that are optionally performed on data associated with the namespace. The user-defined namespace can have certain properties as described below. As long as the namespace has the necessary properties, the generic partitioning framework will be able to automatically partition the namespace based on the application-defined interfaces. This partitioning is accomplished using only the user-defined namespace and fixed interfaces provided with the partitioning system. By partitioning a namespace using only fixed interfaces of the generic partitioning system, the partitioning is decoupled from any application-defined interfaces provided by the user as well as being decoupled from the nature of the computation(s) being performed.
[0026] A namespace or domain for a job can correspond to a set of identifiers for the full range of a type of data that will be operated on and/or generated by a computation. Additionally or alternately, a namespace or domain can correspond to a listing or range of possible states that could be operated on and/or generated by the computation. Still another option is that the namespace can correspond to a set of identifiers for multiple instances of a calculation being performed in a distributed environment. It is not necessary for all potential namespaces or domains for a computation to be identified. Instead, the user can identify the namespaces or states that will be used for partitioning. The identified namespaces or domains should span the entirety of the computations that a user will desire to perform using a given role. The entirety of the computations can be spanned by spanning the entirety of a data set operated on by a computation, by spanning the entirety of instances of calculations within the computation, or in any other convenient manner. Keys can be used to specify discrete values or states within a namespace. Keys can also be used to specify a range of values. Since keys can be used to specify a range, it is understood that the namespace should include some type of serial ordering, so that specifying an upper key and a lower key will uniquely identify a range. This ordering can be based on a conventional ordering, such as ordering based on an alphabet. Alternatively, the ordering can be arbitrary, such as a serial order based on a file containing a listing of the values within a namespace.
[0027] In addition to defining a namespace (including any computations performed using that namespace), a user can also provide one or more application-defined interfaces associated with a server role. An application-defined interface represents one or more tasks or functions that a partition server can perform, the tasks or functions optionally involving one or more namespaces. The application-defined interfaces for a server role can include all tasks or functions that will be performed on a given namespace, so that different server roles do not attempt to access the same data or states. Optionally, it is contemplated that in an exemplary aspect a server role may also include sub-roles, so that some application-defined interfaces within a role operate on different namespaces than other application-defined interfaces within the same role. In a simple example, a client may desire to perform a computation that involves performing at least one type of calculation on a data set. In this situation, a client may define a single role of a server that performs at least one type of calculation on requested element(s) from the data set. This single role can represent one or more servers corresponding to role instances that are configured to perform a scientific calculation, one or more related data mining functions on a data set, or any other convenient computation. The computation and any related data can be defined as part of the definition of a namespace for performing the computation. Additionally, one or more application-defined interfaces can be provided for the server role, such as an application-defined interface for providing metrics regarding the computation to a master role instance. A server role instance can determine the element or elements of the data set to work on based on key values passed to the server. At least one additional master role can manage the servers, including assigning data set partitions to the one or more servers.
[0028] Based on a namespace, a computation can be partitioned so that multiple partition servers handle or perform different portions of processing for the namespace. Each partition corresponds to a range of key values. When a partition is assigned to a partition server, the server performs the desired computation for any requests that contain a key value within the range corresponding to an assigned partition. The partitions assigned to a server do not need to be consecutive relative to the serial ordering of the namespace.
[0029] In a generic partitioning environment, a partition table can be used to track the current assignments of partitions to partition servers. When an active master or dictator assigns a partition to a server, the partition table can be updated first to reflect the assignment. The partition table can then be used to determine the partition server that will handle a client request based on the key specified in the client request. Optionally, one partition table can be used to track the partition assignments for multiple roles, as opposed to having a different partition table for each namespace that is partitioned. As an example, an entry in a partition table can include the low key for a range, the high key for the range, and the role instance or server instance that will perform a requested task on the data or state corresponding to requested key. A partition table can also include other data, such as an epoch number or version number, as will be discussed in greater detail below.
Managing Master and Server Role Instances
[0030] When performing a job, it will often be desirable to have several master role instances to provide redundancy. However, to avoid conflicts, only one master role instance can be the active master at a given time. When a plurality of master role instances are present, the master role instances compete for a lease on the storage object corresponding to the full namespace. The master role instance that is granted the lease becomes the active master or dictator. In addition a master epoch number is stored in the storage object for the master role instance. When a master becomes dictator it increments this number, writes it back to the master role storage object, and subsequently uses the epoch number to communicate with the corresponding partition servers. The active master instance or dictator can also obtain a lease on the partition table or at least a portion of the partition table. Partition servers will ignore heartbeats with a master epoch lower than the highest epoch already seen, thus avoiding stale heartbeats from a master role that is no longer the dictator.
[0031] One option is to use a separate dictatorship library to implement dictatorship via storage object lease. Having a separate dictatorship library can offer several advantages, such as allowing a role outside of the partitioning system to implement redundancy. This allows roles that are not involved in partitioning to use the same method for selecting an active master role.
[0032] Each server maintains a lease on its own storage object. Server storage object names are generated by the master each time a server receives its first partition assignment. Each server also stores its current partition assignment (list of partitions including epoch numbers) in the storage object. A third party can forcefully terminate the lease between a server and its storage object. This functionality can be used by the partition master dictator to break server leases in the partition assignment protocols.
[0033] While having redundant instances of a master role allows for improved performance when a failure event occurs, the redundant instances can also potentially cause a problem due to stale updates. A stale update refers to a situation where a message or instruction from a prior dictator is received by a server or data store after a new dictator has been selected. The problem of stale updates may affect any role or piece of code which maintains state in an external store where messages to the external store may be delayed or re-ordered (e.g. over the internet), and where the state is read from that store upon failover of the role.
[0034] As an example consider a partition master (dictator) that fails over in the middle of updating the partition table for the role. First, the old master initiates an update of row 'X' of the partition table. The old master then stops working. A redundant instance of the master is selected as the new dictator. This new active master updates, reads, or otherwise performs an action using row 'X' of the partition table. The update from the old master then goes through after the action on row 'X' by the new active master. If the update from the old master is incorporated into the partition table, the update will cause the partition table to change without the new (current) master being aware of the change. This can lead to an inconsistent state for the partition table.
[0035] One solution to the problem in the example above is to somehow block the stale update from the old master. One option is to also allow partition masters to obtain leases on portions of the partition table. A lease for a master can correspond to all of the partitions in the table corresponding to a given namespace. Any convenient method for specifying a lease range can be used. For example, if it is desired, a lease can span only a portion of a partition and/or can span multiple partitions. When a new dictator is selected, the masters will still compete for dictatorship via a lease, where an epoch number is provided so the servers can block stale updates, such as stale heartbeat messages. In addition, the master dictator can also acquire a lease on the (relevant part of the) partition table before reading the table when building its state.
[0036] More generally, both masters and servers can use the lease mechanism to avoid stale updates by acquiring leases to blobs, tables, and/or other data structures when the master or server is assigned to handle a data structure. When the master or server is no longer assigned to the data structure, such as due to a failover or an explicit request by a master to change an assignment, the lease is released.
[0037] Stale updates can also be a concern for communication between a master and a server. Processing of messages from an old master can be avoided, for example, by having a master obtain an epoch number when it becomes the dictator. The epoch number is increased for every dictatorship failover. This epoch number can be transmitted on every server heartbeat and/or other message from a master to a server. A server can ignore any heartbeat with a lower epoch number than the highest epoch number the server has seen. This high- watermark epoch number can be stored in the servers' soft state. If a master and a server are also communicating status via storage object leases, the above method can be sufficient to avoid stale message issues. Alternatively, other convenient solutions for avoiding stale messages between a master and a server can also be used.
[0038] A similar epoch number method can be used for each range partition to avoid stale updates from a previously assigned server. For example, each partition can have a current epoch number that is updated by the master when a change occurs for the partition. Examples of changes for a partition include assignment of a partition to a new server, splitting of a partition, and merging of two partitions. Assignments of a partition to a new server can cause the epoch number to increase by one. Splitting of a partition into two or more new partitions can cause each child partition to receive the parent's epoch number incremented by one. When two partitions are merged, the epoch number for the merged partition can be the maximum epoch number for any of the partitions prior to merge incremented by one. Alternatively, any other convenient method for tracking an incrementing an epoch number for a partition can also be used. [0039] As an example of how the partition epoch number may be used, consider a system with a partition master, two partition servers SI and S2, and a third server X. The third server X can be, for example, a data server that contains the actual data set for the namespace operated on by the partition master and servers SI and S2. Servers SI and S2 issue commands (or other messages) to X. Assume messages can be delayed on the way from SI or S2 to X, such as due to processing by a front-end for X. Unless X keeps track of the highest epoch seen for a given range partition it is easy to see how a stale message may be accepted by X. For example, a partition P with epoch 3 can initially be served by server SI . SI can send a message Ml to X. Message Ml contains a key corresponding to partition P as well as the epoch number 3. The partition master then moves partition P from SI to S2. After assignment, S2 sends a message M2 to server X that includes the new epoch number. X receives the message M2 prior to receiving message Ml . This renders Ml stale. X then receives stale message Ml . By tracking the epoch number, server X can recognize that the stale message Ml is from a server that no longer is responsible for partition P.
[0040] To further avoid the potential for stale updates, an epoch validation library can be used. An epoch validation library validates that a received message is not stale. For example, when a server receives a message from a new dictator or receives a request related to a new partition range, the server can check an epoch validation library to verify that the message contains the current epoch number for the master or the partition.
[0041] FIG. 4 shows a schematic example of the relationship between roles, partitions, and keys. In FIG. 4, a user 405 submits a request to perform an action on a desired piece of data, such as accessing information in the user's account. The data is identified by a key. The client request and the key are passed to role 420 which handles the type of request made by the client or user 405. Role 420 includes a client library 430 that defines how the role handles a given type of request. Based on the type of request and the key, the role 420 consults the partition table 455 to find the current server that handles the key range corresponding to the key in the request. The information in partition table 455 is populated based on the partition decisions made by partition master 460. Partition master 460 is shown in FIG. 4 as being one of a plurality of potential partition masters. The additional potential partition masters are for redundancy, and are not active until needed. In the example in FIG. 4, a plurality of partition servers 465 are available as role instances for performing tasks requested from the role 420. Based on partition table 455, partition server N from plurality of partition servers 465 is handling the range of keys corresponding to the key in the client request.
Example of Generic Partitioning Framework
[0042] FIG. 5 shows an example of a group of modules and/or interfaces for providing a generic partitioning framework according to the invention. FIG. 5 also displays examples of application-defined interfaces for providing computational jobs that can take advantage of the generic partitioning environment. In FIG. 5, although the key or namespace is provided by the user or application, the interfaces for partitioning a namespace are fixed interfaces provided by the partitioning system. This decouples partitioning of the namespace from performing the job or work item that operates on the data corresponding to the namespace.
[0043] In FIG. 5, there are at least two types of components or interfaces that are provided or designated by the user. The user provides a definition for a key (and corresponding namespace) 510 and a plurality of server interfaces 520 for the application. As described above, the key describes the namespace for the application. This allows the partitioning system to know the range of variables, states, and/or computations for potential partitioning. In addition to defining the key, the user also provides at least three interfaces related to the key. The interfaces related to the key 510 provide functions to serialize the keys in a namespace, to deserialize keys, and for comparing two keys. Since the user selects the namespace and key values, the operation of these interfaces is not offered as a fixed interface by the partitioning system. The interfaces related to keys and namespaces are instead provided by the user as application-defined interfaces.
[0044] The user also provides application-defined interfaces 520 that are used by server role instances. The application-defined interfaces for server role instances 520 include functions for what operations a server will perform when a server receives a request to operate on a key. For example, an interface for StartServeKeyRange can define one or more operations a server can perform when the server that starts serving a given key range. A corresponding interface for Stop ServeKeyRange can allow a server to end serving of a key range in an orderly manner. Additionally, it may be desirable to collect information from the server. An interface such as an OnReceivedHeartbeatRequest interface can define how a server will extract additional information from a heartbeat message received from a current partition master. Another interface such as BeforeSendingHeartbeatResponse can define what a server will include as additional information in a response to a heartbeat message. This can allow, for example, a server to communicate load information to a partition master for use in load balancing.
[0045] If a user includes additional functions or information as part of heartbeat
(or other message) exchanges, the user can also provide a user defined interfaces for the master role 530. The user defined interfaces for the master role 530 are a complement to the fixed interfaces for the master role, schematically shown as GPMaster 535. The user defined interfaces for the master role 530 are not required.
[0046] In the embodiment shown in FIG. 5, the general partitioning environment is enabled by five modules and/or groups of fixed partitioning system interfaces. These include the fixed interfaces shown as being part of GPClient module 515, the GPServer module 525, the GPMaster module 535, the GPDictator module 536, and the GPLease module 545. Of course, other ways of arranging the modules, interfaces, and/or functions shown in FIG. 5 can also be used.
[0047] The GPClient module 515 shown in FIG. 5 has interfaces that allow for translation from a key specified by a client application or user 599 to the address for the server that will handle a user request. The interfaces for GPClient module 515 perform this lookup by consulting a partition map or partition table 559. The GPClient module interfaces can then, for example, return the server address corresponding to the key to the client application 599 so that the client application can direct the request to the correct server.
[0048] The GPServer module 525 has interfaces that work in conjunction with the application-defined interfaces 520 to perform desired user tasks. The GPServer module 525 includes interfaces for communication between master role instances and server role instances. The GPServer module interfaces also communicate with the interfaces for GPLease module 545 to manage the lease objects and lease contents associated with each server role instance.
[0049] The GPMaster module 535 provides interfaces for the core functions for managing master role instances. The interfaces for GPMaster module 535 handle election of a dictator from the master role instances, communication between a dictator and server role instances (such as via heartbeat messages), and partition management. The partition management can include providing updates to partition table 559. For load balancing, the interfaces for GPMaster module 535 can perform load balancing using an internal algorithm, or the GPMaster module can receive alternative load balancing expressions 560 that are provided by the user as an application-defined interface. Optionally, the messaging functions handled by the GPMaster module 535 can be performed by one or more separate messaging interfaces.
[0050] The GPDictator module 536 provides interfaces that handle functions related to a master role instance that will serve as dictator. The GPDictator interfaces can handle a master role instance claiming dictatorship (such as after winning the dictatorship in an election conducted by GPMaster module 535), releasing dictatorship (such as after a failover), and handling changes to sequence or epoch numbers in order to avoid stale updates.
[0051] The GPLease module 545 provides interfaces that manage leases within the generic partitioning environment. This can include allowing a master or server to acquire a lease on an associated storage object, on a partition, or on any other type of data structure where leases can be obtained.
Basic Master/Server Management Protocols
[0052] One way a partition master can communicate with partition servers is via a regularly sent keepalive ('ping') heartbeat. This heartbeat can be sent to all servers, including servers not serving any partitions. A server can respond to this keepalive heartbeat (or other types of heartbeat messages) with a list of the partitions that the server is currently serving. The heartbeat responses can be used in conjunction with server partition leases to allow a master to verify that a server is serving the correct partitions. When a server is serving one or more partitions, the server also maintains a lease on its own private storage object, such as a binary object or blob. When a server starts up it does not hold any lease - it will only do so upon the initial partition assignment. The storage object lease should also contain a listing of the partitions being served by a server. If either the heartbeat response or the information in a storage object lease differs from the partition information expected by the master, a conflict in partition information exists.
[0053] If a conflict occurs between a master and a server regarding the partitions being served, and if the server is attempting to serve one or more partitions, the conflict is considered a fatal error. As an example, the master may think the server is serving PI and P2 whereas the server reports P2, P4 and P5. In this situation, the master will break the server's lease on the corresponding partition server blob. The master will then issue an alarm and rebuild the proper state for the partition table. This may involve terminating the master so that the proper state is rebuilt upon restart of the master.
[0054] It is also a conflict when the master and server differ and the server reports that no partitions are being served. However, this is not considered a fatal error. This can occur, for example, if the server fails over between two heartbeats, or is down during one or more heartbeats while still 'holding' the lease and then it restarts and responds to the next heartbeat. When a conflict occurs with the server reporting no partitions being served, the partition master can try to delete the server's storage object. If successful, any partitions can be reassigned to other available servers. However, deletion of the lease on the server's storage object may fail if the lease of the previous instance of the server has not yet expired. Thus, it may be necessary to retry the deletion for up to a lease period. Failure to delete the storage object after one lease period is a fatal error that can be handled as described above. When the master restarts it will ensure all unknown storage objects are deleted before proceeding with partition assignments.
[0055] Another potential failure situation is when a server is holding a lease but the server does not respond to a heartbeat (such as a 'keepalive') within some timeout period. Once again, the master can attempt to delete the server's storage object to resolve the conflict. If the storage object is successfully deleted, then the server is no longer functioning. Any partitions the server was serving can be reassigned to other servers. If the storage object cannot be deleted, the master can read the storage object content to check for conflicts between the master and server regarding partition assignments. If there is no conflict the server may continue serving, and the master can try to 'ping' the server again shortly via the normal heartbeat processing. Preferably, there is a limit to how long the master cannot communicate with a server via heartbeats before it breaks the server's lease on the storage object. If a conflict is detected, the conflict can be handled as described above.
[0056] For partition assignment the master piggy-backs a partition assignment request on the heartbeat to the relevant partition server (the heartbeat is accelerated in this case). A partition assignment request consists of the server storage object name plus the complete new list of partitions to be served. For example, to assign partition P2 to a server currently serving PI, the partition assignment consist of both PI and P2. This makes unassignment orthogonal to assignment: to remove PI from a server serving PI and P2, simply send an assignment consisting of P2 only. The master has already updated the partition table when the partition assignment takes place. The partition table is only updated when (before) a partition is assigned to a server; when a partition is removed from a server no update is needed.
[0057] A partition server maintains a storage object lease that starts when the first partition is assigned to the server. The storage object lease is maintained until the server dies or the master forces the lease to be broken, such as due to a conflict between the assignment information for the server in the partition table and the assignment information reported by the server to the master. The storage object name is passed along with the partition assignment. All subsequent partition assignments will contain the same storage object name. When a server receives a partition assignment, a server without an existing storage object name will acquire the storage object name provided in the partition assignment. If the server already has a storage object with a name, the server can compare the name provided in the request with the existing name. If the names are different, the server can issue an alarm and terminate, as the different names indicate a conflict in partition assignment information. After receiving the partition assignment, the server can write any associated information into the storage object for the server. The associated information can include, for example, key ranges to serve, the epoch number for the dictator, and/or the epoch number for the partitions. The server can then respond to the master that provided the partition assignment, start serving any new partitions, and stop serving any removed partitions.
[0058] After a partition assignment, a master role instance will expect a response from the server that confirms the assignment. If the response does not match the assignment, or if the response is delayed, the master role instance can terminate the lease for the server. Alternatively, the master role instance can investigate the blob for the server to determine the server status. For example, if the server's storage object indicates that the assignment has succeeded, and if the response is merely delayed or lost instead of inaccurate, the master role instance can wait and see if the server responds correctly to a subsequent heartbeat or other message. If errors are found and the master is not able to break the storage object lease for the server, the master can terminate to force a reconstruction of the partition map by a new master role instance.
[0059] Additionally, a server role instance can provide statistics for each range
(i.e. partition) it is serving. The statistics are opaque to the general partitioning system, and can be represented as a property bag of name/value pairs. These optional statistics can be incorporated into load balancing formulas, including load balancing formulas provided by a user.
[0060] When the master becomes the active master or dictator, it first obtains a new epoch number from the storage object on which it is keeping a lease. It then collects three pieces of information in order to build its view of the system, and to correct any information that is inconsistent. First, the master reads the partition table. The partition table contains the truth about which partitions exist. Preferably, the master obtains a lease on at least the relevant part of the partition table prior to reading the partition table in order to prevent stale writes by a previous master. Next, the master obtains a list of all existing server storage objects. This can be done by maintaining a list of server storage objects, by requiring that all server storage objects are located in a specified location, or by another convenient method. A heartbeat or another type of message is also used to query each server regarding the server's current partition assignments. This query can include a query for the name of the storage object for a server. Note that the above tasks can proceed in parallel.
[0061] Based on the collected information, the master can identify any inconsistencies between the assignments in the partition table and the assignments reported by each server. If a mismatch is present, one or more corrective actions can be taken, such as breaking the lease of a server on the server object and restarting the master. Additionally, if any server storage objects are identified that are not mentioned in the partition table, these storage objects can be deleted. Finally, any assignments in the partition table that are not confirmed by a server can be queued for assignment to a new partition server. After resolving the conflicts, the master can start normal heartbeat processing, assigning of partitions, and any other master functions.
Load Balancing
[0062] Load balancing can roughly be divided into three activities. Load balancing can include moving a partition from one server to another, splitting a partition into multiple partitions, or merging multiple partitions into a single partition. Typically, changing a partition assignment from a first server to a second server will be in response to the first server having a sufficiently high load based on one or more metrics. If a single partition accounts for a large amount of load, splitting of the partition can be used to allow a large load to be divided among multiple servers. Merging of partitions allows partitions that have lower amounts of activity to be combined. This reduces the overhead required to track and maintain the various partitions for a data set. Optionally, a user can define an upper limit on the number of partitions for a namespace. The thresholds for initiating a merge of partitions can be reduced as the number of partitions approaches the upper limit. The upper limit for number of partitions can be dynamically configured.
[0063] As an example of determining when to split or move a partition, all partitions for a namespace can be sorted based on load. The load can refer to one or more metrics related to performing calculations for a partition. Thus, the load can refer to overall CPU usage for a server or an individual partition; storage used for a server or partition; a number of requests received by a server overall or for a single partition; or any other convenient value that indicates the amount of work being performed by a server and/or for a given partition. Based on the load, the top N partitions which have a higher load than a configurable multiple of the average partition load can be split. N is dynamically configurable. For example, it can be a function of the current number of partitions in the system, such as based on a logarithm of the current number of partitions, or it can be a function of the maximum allowable number of partitions in the system. Additionally or alternately, the load of each server is calculated by adding together the servers' partition loads. The servers can then be sorted by load, and the top N servers with a load greater than some configurable multiple of the average are chosen for movement of partitions between servers. Similarly, a plurality of servers with loads below the average load are selected to receive partitions from higher load servers. Preferably, moving a partition from a first server to a second server is performed to reduce the load of the first server to a value closer to the average load without causing the load of the first server to be below the average load. This is easier to do if all partitions have reasonably similar load. Splitting partitions as described above can be used to reduce the disparity in loads between partitions.
[0064] As noted above, the load for a partition can be derived from statistics collected by the server role instances. This information is passed to the master role instance, such as via heartbeat messages at regular intervals. Preferably, the load statistics are defined so that a load metric is additive. This allows the load for a server to be determined based on summing the loads for the individual partitions on the server. One or more formulas for determining the load for a partition and/or a server can be stored in a separate blob or other storage area. The rules or expressions for partitioning can be default rules provided by the generic partitioning environment, or a user can provide rules and/or expressions. Further, it is contemplated that an expression may also be comprise d of a piece of code submitted by an application using the GP system.
[0065] For load balancing rules and/or expressions provided by a user, a user can first identify one or more metrics that are desired as load balancing metrics. Examples of suitable metrics include CPU usage, network bandwidth usage, number of requests processed per time period, or any other convenient metric. Some metrics may be specific to a partition while other metrics may correspond to a value for all partitions on a partition server. Based on the desired metrics, a user then provides one or more interfaces for collecting the desired metric on each server. Optionally, common metrics such as CPU usage or number of requests per time period can be provided as standard interfaces that are simply accessed by a user. The collected metrics are then passed from partition servers to the corresponding master using messages, such as a heartbeat message used by a server to verify the current status of a server with the master.
[0066] For each metric identified by a user, a series of values recognized by the general partitioning system can be calculated. As an example, the general partitioning system can recognize "dimensions" as variables that are defined by a user. A dimension in the general partitioning system can have an expected format. The dimension can include an expression for a PartitionMetric corresponding to the value of the metric over a partition. Another expression can be for a ServerMetric, which corresponds to the value of the metric over all partitions on the server. Still another expression can be for a Condition value for the metric, which defines a situation where an action is taken.
[0067] In a simple situation, CPU usage can be defined as a dimension by a user. In this example, the CPU usage dimension is used to determine when a server is sufficiently busy that a partition should be moved to another server. In the dimension defined by the user, the percentage of CPU usage on a virtual machine dedicated for processing requests for a given partition is defined as the PartitionMetric. The sum of CPU usage percentages over all partitions on the virtual machine can be defined as the ServerMetric. In this example, the Condition can be defined as ServerMetric usage being greater than 80% of total CPU usage. When this Condition occurs, a partition can be moved to another server. The partition for movement is selected based on the PartitionMetric. It is noted that the PartitionMetric and ServerMetric are both defined by the user. Thus, a user is not required to have a PartitionMetric that is analogous to the ServerMetric. For example, the ServerMetric could be a combination CPU usage and network bandwidth usage, while the PartitionMetric is only related to request rate.
[0068] In addition to defining dimensions for reassignment of partitions, a user can also define dimensions for triggering a partition split. The definition of the dimension for triggering a partition split can be similar to the dimension for a reassigning a partition, or a different format of dimension can be used. For example, a ServerMetric expression may not be necessary for a dimension for triggering a partition split, since a PartitionMetric expression will more likely be useful for determining when to split a partition. Additionally, a dimension for triggering a partition split could include a dimension for how to split a partition when a Condition for splitting the partition is satisfied. It is noted that the dimensions for triggering a partition split may also be useful for identifying when to merge two partitions. Alternatively, a user can define separate dimensions for partition merging.
[0069] More generally, any convenient number of conditions can be specified for determining when a load balancing action should occur. The conditions can incorporate metrics corresponding to one or more dimensions, such as metrics corresponding to a plurality of dimensions. The conditions can be ordered, so that the conditions are evaluated in a specific order for performing load balancing. For example, a condition related to splitting a partition can be placed earlier in the order than a condition for moving partitions to different partition servers. In such an example, if one partition on a partition server is responsible for a large percentage of the load, moving other partitions might not be effective for balancing load among a plurality of servers. By checking for whether to split the partition first, the partition causing the problem can be split into portions with (presumably) lower loads. Similarly, it may be desirable to merge partitions early in an ordering of conditions. For example, a partition server with a large number of low load partitions might appear to be unavailable due to an excessive number of partitions, even though the overall load on the server is below average. Merging partitions before moving assignments allows such a server to become available for receiving an additional partition assignment.
[0070] When specifying a condition for initiating a load balancing activity, any convenient type of metric may be included in a condition. Thus, metrics for load on a single partition, load on a plurality of partitions, load on a server, or load on a plurality of servers may be used together or separately as desired. For metrics related to a plurality of partitions or plurality of servers, a simple example is to determine the load across all servers in order to define an average load. A condition for performing load balancing could then be related to a difference between load on a server versus average server load, such as a difference in the absolute value of the server load versus average load, or a comparison of the server load with a standard deviation from the average server load. When using a plurality of partition loads, it may be desirable to consider the load for a number of the highest loaded partitions on a server in relation to each other. The desired load balancing action may be different for a partition server with multiple partitions that have a similar load as opposed to a partition server with only one high load partition.
[0071] In addition to defining dimensions for partition reassignment, splitting, and merging, a user can also define one or more filters for constraining actions on partitions based on the dimensions. For example, it may be desirable to prevent a server from a receiving a new partition assignment when the server has either a CPU usage ServerMetric of more than 70% or the number of partitions is greater than 10. An AssignmentFilter value can be defined that prevents such assignments. Other examples of potential filters include filters for preventing the movement of a partition from an existing partition server, preventing a split of a partition, or preventing a merge of a partition. Depending on the type of filter, a filter may prevent a load balancing action from occurring. Alternatively, a filter can modify the order of consideration of conditions, or the filter can cause a condition to be skipped entirely during a load balancing calculation cycle.
[0072] As an example, consider a hypothetical system where all server requests consume the same amount of resources. In such a system, a reasonable load metric could be based on the request rate. In this example, each server role instance collects an average of the request rate over a longer time period (RR SlowMA) and an average of the request rate over a shorter time period (RR FastMA). These request rate averages are transmitted back to the master role instance as name/value pairs in a property bag. A simple load metric could then be defined as a formula in the load balancing rules as PartitionMetric = max(RR_FastMA, RR SlowMA). The left hand side of the PartitionMetric for the dimension "load" corresponds to an identifier recognized by master component within the partitioning system. In this case, the dimension "load" would be identified in advance by the user. The right hand side can be any arbitrary expression that generates a load value that is assigned to the PartitionMetric. In this case, the load corresponds to the number of requests based on one of a plurality of moving averages. Based on the PartitionLoad value and/or other values, one or more Conditions can be defined regarding whether partitions should be split, merged, or moved between servers.
[0073] More generally, a user can define any combination of metrics and expressions (such as conditions) for making load balancing decisions. The expressions defined by a user for making a load balancing decisions can also include conditional logic and/or support for multi-dimensional constraints/optimization goals. Thus, a user can define a decisions tree or use other conditional logic to provide an ordering for how and when expressions are evaluated. For example, a user could have a first expression evaluated, and then select from a plurality of potential additional expressions to evaluate based on the value of the first expression. This could be based on an "if-then-else" type of conditional logic, a lookup table for the next expression based on a determined value, or any other convenient type of conditional logic. As a result, a user has flexibility to specify the types of metrics and expressions for use in load balancing, including whether a given expression is evaluated as well as providing an order for evaluating such expressions. Note that the order of evaluation for expressions can also be determined dynamically based on the value of previously evaluated expressions.
[0074] Another example of expressions defined by a user for making load balancing decisions may include multi-dimensional constraints/optimization goals. For example, a user may defined multiple dimensions (e.g., two), and for each dimensions the user may define a separate optimization goal or constraint. CPU utilization and request latency are two exemplary dimensions. The user may specify rules such that a partition servers' CPU usage is below a first threshold (e.g., 90%) and at the same time minimize an average request latency across a defined set of partition servers (e.g., all partition servers). This approach may differ from an if-then-else type logic where users specify exactly what to do. In this model, users define limits and optimization goals, which allow the system to find a solution automatically.
[0075] In another hypothetical system, a user can provide a plurality of conditions or expressions for evaluation. The expressions are based on various CPU usage metrics for the partition servers that a serving a given namespace. A first expression evaluates whether any partition server has a CPU usage related to the namespace of greater than 60%. For this user, if no CPU has a usage greater than 60%>, the user does not desire load balancing. Thus, if the result of the first expression is false (i.e., no partition server has a CPU usage greater than 60%), then no further expressions are evaluated, as load balancing is not desired. If at least one partition server has a CPU usage greater than 60%, a series of expressions can then be evaluated to determine a load balancing action to perform.
[0076] In a situation where load balancing results in moving a partition, a partition can be moved from a first server to a second server by having the master role instance issue two assignment requests. An assignment request to the first server does not include the partition, which results in the first server stopping service for the partition. A second assignment request to the second server includes the partition.
[0077] In a situation where a partition is split into two or more partitions, a master role instance can initiate a split by determining a split key, which corresponds to a key value that will form the end of the inclusive range for one of the new partitions. The split key can be selected in any convenient manner. A master or server role instance can select a split key based on the partition, such as by selecting a key value at or near the middle of the range of the partition. Alternatively, a server can select a split key based on additional statistics regarding the partition. For example, a sampling-based bucket mechanism could be used to track the load for various portions of the partition in a manner similar to determining the load for the full partition. The split key can then be selected so that the load is similar for the buckets assigned to the new partitions.
[0078] Within a given role, the (active) master is responsible for spreading the load across the partition servers. Preferably the master will prevent one or more servers from becoming overloaded and thus unable to process requests. In an alternative embodiment, adjustments of load could be done by keeping one partition/range per server and adjusting these ranges. By instead moving partitions, an adjustment to the load can be performed while impacting a smaller number of servers.
[0079] It is often desirable to have a minimum number of partitions per server so that load can be moved around smoothly via partition reassignments. When the number of partitions falls to the minimum level, further mergers are not performed. Similarly, it is often also desirable to avoid having too many partitions. As the maximum number of partitions is approached for a server, the likelihood of merging partitions can increase. As an example, it may be desirable to maintain between 5 and 8 partitions per server. Of course, the various embodiments of the invention can work with any number of partitions per server, such as from as few as 1 partition per server to hundreds or more per server.
[0080] Preferably, both the split and merge protocols are stateless. Either the master or the involved server(s) may fail over at any time without causing an error for the partitioning system. In other words, if either the master or a server fails during the split or merge process, the next master or server will be able to construct a valid list of partition assignments regardless of when the failure occurred. In a stateless split protocol, the participating server is not required to perform any of the split actions. As an example, a partition table can include a partition on a server S 1 that ranges from a low key value of D to a high key value of H. In this example, the epoch number for the partition is 2. Based on user-defined load balancing equations, it is determined that the partition should be split, so that part of the partition can be assigned to another server. The master role instance asks server SI for a split key. The server SI returns a key of G as the split key. The master then modifies the partition table. In place of the single entry noted above, the table now contains two partitions. One has a low key value of D and a high key value of G, while the second partition has a low key value of G and a high key value of H. As noted above, the partition range definitions based on a low key value and high key value are inclusive of the low key value and exclusive of the high key value. The change in the partition table can occur by modifying the existing entry and adding a new entry, by removing the existing entry and adding two new entries, or by any other convenient method.
[0081] At the next heartbeat cycle, the master detects a conflict between the partitions served by server SI and the information in the partition table. Because a split has just occurred the master does not terminate the blob lease of the server S 1. Instead, the master sends an assignment to server S 1 with a partition range of D to G and an epoch of 3. This modifies the assignment of the partition at SI to match one of the split partitions in the partition table. After receiving an acknowledgment of the new assignment from server SI, the master can assign the second split partition to another server. The second split partition would also have an epoch number of 3. Alternatively, both of the split partitions can be assigned to server S 1 initially, with one or both partitions being moved at a later time to perform load balancing.
[0082] Merging of two partitions can also be handled in a stateless manner. When partitions are merged, as an initial step the partitions for merger are unassigned from the current server. For example, a first partition on server S2 can have a low key value of K and a high key value of M. In this example, the epoch number for the first partition is 7. A second partition on server S4 can have a low key value of M and a high key value of N. The epoch value for the second partition is 9 in this example. As an initial step, the partitions can be unassigned from their respective servers, so that the partition table shows a non-assigned value for the server. The two partition entries are then replaced with a single entry having a low key of K and a high key of N. The epoch number assigned to this partition is one greater than the highest value of the merged partitions, which corresponds to 10 in this example. The new partition can then be assigned to a server. Additional Examples
[0083] In order to provide context for describing the invention, an example of organizing computing resources in a distributed network or cloud computing environment is provided. The following description of a cloud computing environment is provided as an illustrative example. Those of skill in the art will recognize that the claimed invention can be used in conjunction with distributed network environments with alternative types of organization. The definitions below are used within the illustrative example.
[0084] A "client" is defined as a role that issues one or more requests for action by a application-defined interface against a namespace or domain. A client can correspond to a user or to a process initiated on behalf of a user. For example, a request to a lookup a particular account corresponds to a request directed to an application for account lookup that is made against the domain of all accounts with a key corresponding to the desired account.
[0085] A "work item" is a static representation of a job to be run in the cloud computing environment. A work item can specify various aspects of a job, including job binaries, pointers to the data to be processed, and optionally the command line to launch tasks for performing the job. In addition, a work item may specify the reoccurrence schedule, priority and constraints. For example, a work item can specify to be launched every day at 5 PM.
[0086] A "job" is a running instance of a work item. A job contains a collection of tasks that work together to perform a distributed computation. The tasks can run on one or more virtual machines in the cloud computing environment.
[0087] A "task" is the fundamental execution unit of a job. Each task runs on a virtual machine. Users can specify additional input to the command line and pointers to input data for each task. A task may create a hierarchy of files under its working directory on the virtual machine performing the task during the course of execution of the task.
[0088] A user of a cloud computing environment will typically desire to perform jobs using the cloud computing resources. The jobs will typically involve performing jobs on data that is stored in locations that are accessible via the cloud computing environment. One way for an operator to provide a cloud computing environment is to provide the environment as a number of layers. FIG. 1 schematically shows an example of a system suitable for performing tasks within a cloud computing environment. The system in FIG. 1 includes a task runtime layer 110, a third party task runtime layer 120, a resource management layer 130, and a scheduling and execution layer 140.
[0089] In the embodiment shown in FIG. 1, the task runtime layer 110 is responsible for setting up the execution environment and security context for tasks from a user 105. The task runtime layer 110 can also launch tasks and monitor the status of the tasks. The task runtime layer 110 can take the form of a system agent running on each virtual machine. The task runtime layer may also include a runtime library that can be linked into a users' task executables. Having runtime libraries as part of the task runtime layer 110 can potentially provide richer capability to tasks executed by the system agent. Examples of runtime libraries include one or more efficient communication libraries to allow fast communication among tasks; an efficient remote file access library support to read files from other virtual machines and/or other tasks; a checkpoint library to allow tasks to checkpoint (e.g. into binary large objects) and resume; a logging library; and a library for providing a distributed file system to be used across virtual machines performing a given task within a pool of virtual machines.
[0090] The third party task runtime layer 120 allows additional runtimes to be built and run on top of task runtime layer 110. The third party task runtime layer 120 also can provide additional capabilities for coordinating the running of tasks for a job. Examples may include a MapReduce runtime to a library for providing a distributed file system to be used across virtual machines performing a given task within a pool of virtual machines. This allows a user to organize the cloud computing environment in a manner tailored for the user's jobs or tasks. In some embodiments, a job manager task can facilitate allowing a user to use a third party runtime layer to run and/or control cloud computing resources.
[0091] Resource management layer 130 deals with managing the computing resources available in the cloud computing environment. One option is to have the resource management layer 130 manage the resources at three different levels. At a first level, the resource management layer 130 manages the allocation and deallocation of virtual machines associated with a job (i.e., execution of a work item) as well as the files stored on each virtual machine associated with a task. At a second level, the virtual machines associated with a job can be grouped into pools of machines. A pool can contain virtual machines associated with one or more jobs and/or work items. Depending on the embodiment, a single pool can span across multiple virtual machine clusters, such as all virtual machine clusters in a data center, a plurality of virtual machine clusters across a plurality of data centers within a geographic region, or a plurality of clusters across data centers in a plurality of geographic regions. A single pool can contain a large number of virtual machines, such as millions. The virtual machines can be contained in a large number of pools, such as up to billions. At a third level, the resource management layer manages the amount of virtual machines available for association with jobs or work items in a given group of pools. This allows for dynamic adjustment of the amount of compute resources used based on the current load of the system. Additionally, virtual machines that are not being used by a current group of pools may be released back to the cloud computing environment for incorporation into other groups of pools.
[0092] In the embodiment shown in FIG. 1, scheduling and execution layer 140 manages work items, jobs, and tasks that are being performed by a user. The scheduling and execution layer 140 makes scheduling decisions and is responsible for launching jobs and tasks as well as retries on failures. Such a scheduling and execution layer 140 can include components for managing jobs and/or tasks at various levels.
[0093] The layers described above can be implemented in a cloud computing environment that includes processors at multiple geographic locations. FIG. 2 schematically shows an example of how processors at different locations can be integrated within a single cloud computing architecture.
[0094] In FIG. 2, one or more task tenants 215 can be used to manage pools of virtual machines. A task tenant 215 can maintain a set of virtual machines. The jobs of one or more users can run on the virtual machines within a task tenant 215 as part of one or more pools of virtual machines. One or more task tenants 215 can be used in a given geographic region. The responsibilities of a task tenant 215 can include maintaining the set of virtual machines and dynamically growing or shrink the task tenant based on the resource utilization within the task tenant. This allows a task tenant 215 to increase the number of virtual machines within the task tenant to accommodate increased customer demand. This also allows a task tenant 215 to release unused virtual machines so that the virtual machines can be allocated to other hosted services in the data center handling service for other customers. Another responsibility of a task tenant 215 can be implementing part of the pool allocation/deallocation/management logic. This allows the task tenant 215 to participate in determining how virtual machines are assigned to pools associated with a task for a customer. The task tenant 215 can also be responsible for scheduling and execution of tasks on the virtual machines within the task tenant.
[0095] In the embodiment shown in FIG. 2, one or more task location services 225 are provided that control a plurality of task tenants 215. The plurality of task tenants can correspond to all task tenants in a given geographic region, various task tenants from around the world, or any other convenient grouping of task tenants. In FIG. 2, task location services 225 are shown that serve regions labeled "US North" and US South". The responsibilities of a task location service 225 can include management of task accounts for the given geographic region. The task location services 225 can also provide application programming interfaces (APIs) for allowing users to interact with the cloud computing environment. Such APIs can include handling APIs associated with pools of virtual machines, pool management logic, and coordination of pool management logic across task tenants within a given geographic region. The APIs can also include APIs for handling tasks submitted by a user, as well as maintaining, scheduling, and terminating work items or jobs associated with the user tasks. The APIs can further include APIs for statistics collection, aggregation, and reporting for all work items, jobs, tasks, and pools in a geographic region. Additionally, the APIs can include APIs for allowing auction of available virtual machines as preemptible virtual machines to users on a short term basis based on a spot market for virtual machines. The APIs can also include APIs for metering usage and providing billing support.
[0096] The task location services 225 can be linked together by a global location service 235. The global location service 235 can be responsible for account creation and management of accounts, including managing task accounts in conjunction with the task location service tenants 225. This includes being responsible for disaster recovery and being responsible for availability of work items and jobs if there is a major data center disaster. This may include running a work item or job in a different location due to a data center not being available for any reason. This can also include allowing customers to migrate their work items, jobs, and pools from one data center to another data center. Typically there will be only one active global location service 235. This active global location service 235 is in communication with the various task location services 225 as well as service components for managing data storage (not shown). The global location service can maintain a global account namespace 237.
[0097] FIG. 3 shows a potential configuration for a task location service. In the configuration shown in FIG. 3, a task location service can include one or more account servers 321. The account servers handle account management for accounts in a given geographic region, including creation, deletion, or property updates. Account front ends 322 serve as the front end nodes for account service. The account front ends 322 are behind an account virtual IP address 324 as shown in the figure. The account front ends 322 process the account API requests coming from global location service, such as API requests to create accounts or delete accounts.
[0098] The configuration in FIG. 3 also includes one or more pool servers 331. A pool server 331 handles pool management and pool transactions for pools of virtual machines in a given geographic region. A pool server 331 handles pool creation, deletion and property updates. A pool server 331 also manages the high level virtual machine allocation algorithm across multiple task tenants. Virtual machine allocation can take into consideration the connectivity of a virtual machine with storage for a given user. The pool server may also perform other tasks related to allocation of virtual machines.
[0099] The configuration in FIG. 3 also includes one or more work item or job servers (WIJ) 336. WIJ servers 336 handle creation, deletion, and updates of work items and jobs. In addition, if a user has requested automatic creation and/or destruction of pools when work items or jobs start or finish, the WIJ servers 336 may initiate the creation and deletion of pools associated with the work items or jobs. The WIJ servers 336 also use generic partitioning mechanisms for scaling. In an embodiment, there are multiple WIJ servers 336 in each task location service, and each of the WIJ servers handles a range of work items.
[00100] The pool servers 331 and WIJ servers 336 receive requests from users via task location service front ends 338. The task location service front ends 338 are also responsible for calling corresponding components to process requests from users. The task location service front ends 338 are behind an account virtual IP address 334 as shown in the figure.
[00101] The configuration in FIG. 3 further includes a task location service master
342. In an embodiment, the task location service master 342 has two main responsibilities. First, the task location service master 325 serves as a master system for implementing partitioning logic for the corresponding servers in a task location service 225. Additionally, the task location service master 342 can be responsible for computing the new market price for preemptible virtual machines at the beginning of each spot period for the entire geographic region of the task location service. It can collect current bids and resource availability information from the pool servers and task tenants, and computes the new market price accordingly. Alternatively, the task location service master can send the bid and resource availability information to a spot price market service. It also makes high level allocation guidance to pool servers about preemptible virtual machines across all task tenants in a geographic region.
[00102] In order to track the activity and behavior of the computing environment, a task location service master 342 can communicate with one or more statistics aggregation servers 355. The statistics aggregation servers are responsible for collecting and aggregating detailed statistics for tasks, jobs, work items and pools. The other components in the system emit fine-grained statistics for tasks and virtual machines. The statistics aggregation servers aggregate these fine-grained statistics from task level or virtual machine level statistics into work item, account level, and/or pool level statistics. The statistics can be exposed for use via an API. In addition, the statistics aggregation servers can be responsible for generating hourly metering records for each account for use in billing. [00103] As a more specific example, generic partitioning can be applied to the roles and sub-roles in the task location service shown in FIG. 3. The top level role demonstrated in FIG. 3 is a task location service or tenant. If multiple instances of the task location service are present, one of the instances will correspond to task location service master (or dictator) 342. Within the tenant are an account server role 321, a pool server role 331, and a work items-jobs server role 336. Each of these roles also represent instances of the task location service, but these role instances handle a set of functions within the overall tenant. For example, a request for account information is handled by the account server role within the tenant. If multiple instances of the task location service or tenant are present, the master for each of the roles within the tenant can correspond to a different instance.
[00104] FIG. 6 shows a conventional example of how redundancy can be provided for multiple master roles. In this conventional example, each master role needs to have two additional instances in order to improve availability. A fault domain includes nodes that have common failure patterns and they may fail together. For example, nodes on the same rack sharing the same power supply may be in a common fault domain as they may fail as the result of a common problem. An upgrade domain corresponds to a set of nodes that may be taken offline at the same time during a system upgrade. These roles are spread across different "fault domains" and "upgrade domains" as illustrated in FIG. 6 so as to not be down at a common time as a result of an upgrade or a failure.
[00105] Under the conventional method, providing the additional instances needed for the three roles within a task location service would require having separate additional instances for each role. In FIG. 6, this shown by having an explicit machine that provides the additional instances for each type of master. Thus, the master 621 for the account servers would require additional instances 622 and 623. Similarly, master 632 for the pool servers has backup instances 631 and 633. Master 643 for the WD servers has backup instances 642 and 641.
[00106] FIG. 8 shows an example of how virtual machines in a distributed computing environment using generic partitioning can be organized to provide the various instances and masters for a role. In FIG. 8, separate GP Masters 821, 831, and 841 are shown for account servers, pool servers, and WIJ servers respectively. Since the GP Master module and any fixed interfaces are the same regardless of the role being managed, the backup server needs for the GP Masters 821, 831, and 841 can be combined on a single machine. Thus, a single backup 852 can be provided as the backup for the three GP Masters. If one of the GP Masters 821, 831, or 841 experiences a failure, the same GP Master module and fixed interfaces may be used. The only additional information types, in this example, needed by the failover backup to take over the GP Master role that experienced a failure are the key for the corresponding namespace and any application- defined interfaces. Similarly, a single second backup 853 can be used for all three of GP Masters 821, 831, and 841. Consequently, in this example, only two GP Master backup servers (852 and 853) are used for the at least three GP Master roles. Although three GP Master roles are shown as being backed up by common machines, any convenient number of GP Master roles belonging to the same user or account could be backed up using a common machine.
[00107] FIG. 12 shows an exemplary aspect of providing backup machines for master roles relative to fault domains and upgrade domains, in accordance with aspects of the present invention. Similar to those concepts discussed hereinabove with respect to FIG. 8, multiple GP Master roles may be backed up on a fewer number of servers. For example, FIG. 12 depicts an Account GP Master 1202 in a first fault domain and a first upgrade domain, a Pool GP Master 1204 in a second fault domain and in a second upgrade domain, a WIH GP Master 1210 in a third fault domain and in a third upgrade domain, a first GP Backup 1206 and a second GP Backup 1208. The first GP Backup 1206 and the second GP Backup 1208 are each in different fault domains and upgrade domains from the GP Master roles. In this illustrative example, a single Generic Partitioning system, which in this example only requires five servers (or potentially four servers with a single backup) to host all of the masters for the three roles. In the example illustrated in FIG. 6, nine different servers may be required for those same three master roles. The approach illustrated in FIG. 8 may be accomplished through the utilization of two additional servers that can be used for any type of role being hosted by the system. Consequently, a backup server (e.g., GP Backup 1206) may be used if one or more of the master roles are unavailable do to a failure of a fault domain or an unavailable upgrade domain. While a fewer number servers may be needed in this example (as compared to that which is discussed with respect to FIG. 6), it is contemplated that additional fault domains and upgrade domains may be implemented to ensure availability of the backup servers. As with FIG. 8 discussed above, it is contemplated that any number of master roles may be backed up by common machines, in an exemplary aspect.
[00108] [00109] Having briefly described an overview of various embodiments of the invention, an exemplary operating environment suitable for performing the invention is now described. Referring to the drawings in general, and initially to FIG. 7 in particular, an exemplary operating environment for implementing embodiments of the present invention is shown and designated generally as computing device 700. Computing device 700 is but one example of a suitable computing environment and is not intended to suggest any limitation as to the scope of use or functionality of the invention. Neither should the computing device 700 be interpreted as having any dependency or requirement relating to any one or combination of components illustrated.
[00110] Embodiments of the invention may be described in the general context of computer code or machine-useable instructions, including computer-executable instructions such as program modules, being executed by a computer or other machine, such as a personal data assistant or other handheld device. Generally, program modules, including routines, programs, objects, components, data structures, etc., refer to code that perform particular tasks or implement particular abstract data types. The invention may be practiced in a variety of system configurations, including hand-held devices, consumer electronics, general-purpose computers, more specialty computing devices, and the like. The invention may also be practiced in distributed computing environments where tasks are performed by remote-processing devices that are linked through a communications network.
[00111] With continued reference to FIG. 7, computing device 700 includes a bus
710 that directly or indirectly couples the following devices: memory 712, one or more processors 714, one or more presentation components 716, input/output (I/O) ports 718, I/O components 720, and an illustrative power supply 722. Bus 710 represents what may be one or more busses (such as an address bus, data bus, or combination thereof). Although the various blocks of FIG. 7 are shown with lines for the sake of clarity, in reality, delineating various components is not so clear, and metaphorically, the lines would more accurately be grey and fuzzy. For example, one may consider a presentation component such as a display device to be an I/O component. Additionally, many processors have memory. The inventors hereof recognize that such is the nature of the art, and reiterate that the diagram of FIG. 7 is merely illustrative of an exemplary computing device that can be used in connection with one or more embodiments of the present invention. Distinction is not made between such categories as "workstation," "server," "laptop," "hand-held device," etc., as all are contemplated within the scope of FIG. 7 and reference to "computing device."
[00112] The computing device 700 typically includes a variety of computer- readable media. Computer-readable media can be any available media that can be accessed by computing device 700 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer- readable media may comprise computer storage media and communication media. Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, Random Access Memory (RAM), Read Only Memory (ROM), Electronically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other holographic memory, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to encode desired information and which can be accessed by the computing device 700. In an embodiment, the computer storage media can be selected from tangible computer storage media. In another embodiment, the computer storage media can be selected from non-transitory computer storage media.
[00113] Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and includes any information delivery media. The term "modulated data signal" means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer-readable media.
[00114] The memory 712 can include computer-storage media in the form of volatile and/or nonvolatile memory. The memory may be removable, non-removable, or a combination thereof. Exemplary hardware devices include solid-state memory, hard drives, optical-disc drives, etc. The computing device 700 includes one or more processors that read data from various entities such as the memory 712 or the I/O components 720. The presentation component(s) 716 present data indications to a user or other device. Exemplary presentation components include a display device, speaker, printing component, vibrating component, and the like.
[00115] The I/O ports 718 can allow the computing device 700 to be logically coupled to other devices including the I/O components 720, some of which may be built in. Illustrative components can include a microphone, joystick, game pad, satellite dish, scanner, printer, wireless device, etc.
[00116] FIG. 9 shows an example of a method according to the invention. In FIG.
9, a namespace is received 910 from a client application. The client application can correspond to any convenient type of computation that a client desires to perform in a distributed computing environment. The namespace comprises a plurality of key values that span the namespace. A plurality of partitions based on the namespace are assigned 920 to one or more partition servers. The partition servers can implement one or more application-defined partitioning system interfaces. One or more metrics are received 930 from the partition servers based on activity related to the assigned partitions. One or more application-defined load balancing conditions are also received 940. A load balancing action is subsequently performed 950 based on a corresponding load balancing conditions being satisfied. For example, the load balancing action can include a change in a partition assignment for at least one partition server, a split of a partition, a merger of two partitions, or a combination thereof.
[00117] FIG. 10 shows another example of a method according to the invention. In
FIG. 10, a namespace is received 1010 from a client application. The client application can correspond to any convenient type of computation that a client desires to perform in a distributed computing environment. The namespace comprises a plurality of key values that span the namespace. An application-defined interface is also received 1020. The application-defined interface comprises a plurality of expressions corresponding to load balancing conditions and one or more conditional logic expressions. A value is determined 1030 for at least one conditional logic expression. One or more of the plurality of expressions corresponding to load balancing conditions are then selected 1040 based on the determined value of the at least one conditional logic expression. After evaluation, it is determined that at least one load balancing condition corresponding to a selected expression is satisfied 1050. A load balancing action corresponding to the satisfied condition is then performed 1060. Further, it is contemplated that the method may further be comprised of an evaluation of a plurality of load balancing conditions or expressions or procedures without a predetermined order to reach an optimization target while still obtaining the constraints specified in the load balancing conditions.
[00118] FIG. 11 shows yet another example of a method according to the invention.
In FIG. 11, a namespace is received 1110 from a client application. The client application can correspond to any convenient type of computation that a client desires to perform in a distributed computing environment. The namespace comprises a plurality of key values that span the namespace. A plurality of partitions based on the namespace are assigned 1120 to one or more partition servers. The partition servers can implement one or more application-defined partitioning system interfaces. One or more metrics are received 1130 from the partition servers based on activity related to the assigned partitions. A plurality of application-defined load balancing conditions are also received 1140. The plurality of application-defined load balancing conditions include an ordering for evaluation of the conditions. The plurality of application-defined load balancing conditions are evaluated 1150 according to the received ordering. A load balancing action is performed 1160 that corresponds to a received load balancing condition that is satisfied.
[00119] Embodiments of the present invention have been described in relation to particular embodiments, which are intended in all respects to be illustrative rather than restrictive. Alternative embodiments will become apparent to those of ordinary skill in the art to which the present invention pertains without departing from its scope.
[00120] In an embodiment, a method is provided for performing load balancing in a distributed computing environment. The method includes receiving a namespace from a client application comprising a plurality of key values that span the namespace; assigning, by a fixed partitioning system interface, a plurality of partitions based on the namespace to one or more partition servers, the plurality of partitions corresponding to key ranges that span the namespace, the partition servers implementing one or more application-defined partitioning system interfaces; receiving one or more metrics from the partition servers based on activity related to assigned partitions; receiving one or more application-defined load balancing conditions, a load balancing condition comprising an expression that can be evaluated based on at least one received metric; and performing, by a fixed partitioning system interface, a load balancing action corresponding to a received load balancing condition, the load balancing action being performed based on the corresponding load balancing condition being satisfied, the load balancing action including a change in a partition assignment for at least one partition server. [00121] In another embodiment, a method is provided for performing load balancing in a distributed computing environment. The method includes receiving a namespace from a client comprising a plurality of key values that span the namespace; receiving an application-defined interface comprising a plurality of expressions corresponding to load balancing conditions based on one or more of the application- defined metrics and one or more conditional logic expressions; determining a value for at least one conditional logic expression; selecting one or more of the plurality of expressions corresponding to load balancing conditions for evaluation based on the determined value for the at least one conditional logic expression; determining that at least one load balancing condition corresponding to the selected one or more expressions is satisfied; and performing, by a fixed partitioning system interface, a load balancing action corresponding to the satisfied condition.
[00122] In still another embodiment, a method for performing load balancing in a distributed computing environment is provided. The method includes receiving a namespace from a client application comprising a plurality of key values that span the namespace; assigning, by a fixed partitioning system interface, a plurality of partitions based on the namespace to one or more partition servers, the plurality of partitions corresponding to key ranges that span the namespace, the partition servers implementing one or more application-defined partitioning system interfaces; receiving one or more metrics from the partition servers based activity related to assigned partitions; receiving a plurality of application-defined load balancing conditions, a load balancing condition comprising an expression that can be evaluated based on at least one received metric, the plurality of load balancing conditions further comprising an ordering for the conditions; evaluating the plurality of application-defined load balancing conditions according to the received ordering; and performing, by a fixed partitioning system interface, a load balancing action corresponding to a received load balancing condition, the load balancing action being performed based on a corresponding load balancing condition being satisfied during said evaluation, the load balancing action including a change in a partition assignment for at least one partition server.
[00123] From the foregoing, it will be seen that this invention is one well adapted to attain all the ends and objects hereinabove set forth together with other advantages which are obvious and which are inherent to the structure. [00124] It will be understood that certain features and subcombinations are of utility and may be employed without reference to other features and subcombinations. This is contemplated by and is within the scope of the claims.

Claims

CLAIMS What is claimed is:
1. A method for performing load balancing in a distributed computing environment, comprising:
receiving a namespace from an application comprising a plurality of key values that span the namespace;
assigning, by a fixed partitioning system interface, a plurality of partitions based on the namespace to one or more partition servers, the plurality of partitions corresponding to key ranges that span the namespace, the partition servers implementing one or more application-defined partitioning system interfaces;
receiving one or more metrics from the partition servers based on activity related to assigned partitions;
receiving one or more application-defined load balancing conditions, a load balancing condition comprising an expression that can be evaluated based on at least one received metric; and
performing, by a fixed partitioning system interface, a load balancing action corresponding to a received load balancing condition, the load balancing action being performed based on the corresponding load balancing condition being satisfied, the load balancing action including a change in a partition assignment for at least one partition server.
2. The method of claim 1, wherein the load balancing action is splitting a partition, merging two partitions, changing the partition server assignment for a partition, or a combination thereof.
3. The method of claim 1, wherein the load balancing action is performed in a stateless manner.
4. The method of claim 1, wherein the one or more metrics are received from the partition servers in responses to heartbeat messages.
5. The method of claim 1, wherein at least one of the one or more metrics is an application-defined metric.
6. The method of claim 1, wherein a load balancing condition further comprises a partition load value based on at least one received metric, a server load value based on at least one received metric, or a combination thereof.
7. The method of claim 6, wherein the condition comprises an expression based on at least one of a partition load value or a server load value.
8. The method of claim 7, wherein the load balancing action comprises changing the partition server assignment for a partition, and wherein the load balancing condition is satisfied based on a server load value, a partition for change in assignment is selected based on a partition load value, and at least one metric in the partition load value is different from a metric in the server load value.
9. The method of claim 1, wherein the received one or more load balancing conditions further comprise an order for evaluating the received load balancing conditions.
10. The method of claim 1, wherein the partition servers are implemented as virtual machines in a distributed computing environment.
11. One or more computer-storage media storing computer-useable instructions that, when executed, perform a method for performing load balancing in a distributed computing environment, comprising:
receiving a namespace from a client comprising a plurality of key values that span the namespace;
receiving an application-defined interface comprising a plurality of expressions corresponding to load balancing conditions based on one or more of the application-defined metrics and one or more conditional logic expressions;
determining a value for at least one conditional logic expression; selecting one or more of the plurality of expressions corresponding to load balancing conditions for evaluation based on the determined value for the at least one conditional logic expression;
determining that at least one load balancing condition corresponding to the selected one or more expressions is satisfied; and
performing, by a fixed partitioning system interface, a load balancing action corresponding to the satisfied condition.
12. The method of claim 11, wherein performing a load balancing action comprises: removing the assignment of a partition from a first partition server; selecting, by a fixed partitioning system interface, a partition server for assignment of the removed partition from a group of partition servers, the group of partition servers excluding at least one partition server based on the filter condition; and assigning the removed partition to the selected partition server.
13. The method of claim 11, wherein the conditional logic expression comprises an expression that can be evaluated to determine a value, the conditional logic expression having an associated lookup table for selection of the one or more expressions based on the determined value.
14. The method of claim 11 , wherein the expressions corresponding to load balancing are comprised of a procedure that utilizes two or more metrics as an input and also specify a target or a constraint.
15. The method of claim 11, wherein the dimension of application- defined metrics includes at least one expression for a partition metric, at least one expression for a server metric, and a condition based on one or more of the partition metric or server metric expressions.
16. The method of claim 15, wherein the dimension of application- defined metrics further comprises an order for evaluation of the application-defined metrics.
17. A method for performing load balancing in a distributed computing environment, comprising:
receiving a namespace from a client application comprising a plurality of key values that span the namespace;
assigning, by a fixed partitioning system interface, a plurality of partitions based on the namespace to one or more partition servers, the plurality of partitions corresponding to key ranges that span the namespace, the partition servers implementing one or more application-defined partitioning system interfaces;
receiving one or more metrics from the partition servers based activity related to assigned partitions;
receiving a plurality of application-defined load balancing conditions, a load balancing condition comprising an expression that can be evaluated based on at least one received metric, the plurality of load balancing conditions further comprising an ordering for the conditions; evaluating the plurality of application-defined load balancing conditions according to the received ordering; and
performing, by a fixed partitioning system interface, a load balancing action corresponding to a received load balancing condition, the load balancing action being performed based on a corresponding load balancing condition being satisfied during said evaluation, the load balancing action including a change in a partition assignment for at least one partition server.
18. The method of claim 17, wherein evaluating the plurality of application-defined load balancing conditions according to the received ordering comprises:
determining that a load balancing condition is satisfied;
pausing the evaluation of the load balancing conditions while the load balancing action corresponding to the determined load balancing condition is being performed; and
resuming evaluation of the load-balancing conditions according to the received ordering.
19. The method of claim 17, further comprising receiving at least one filter condition, wherein the status of the filter condition is changed due to performance of a load balancing action during said evaluation of the load-balancing conditions.
20. The method of claim 19, wherein the status of the filter condition is changed due to merging two partitions.
PCT/US2013/024366 2012-02-03 2013-02-01 Dynamic load balancing in a scalable environment WO2013116664A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
CN201380007789.7A CN104081353B (en) 2012-02-03 2013-02-01 Balancing dynamic load in scalable environment
JP2014555763A JP6185486B2 (en) 2012-02-03 2013-02-01 A method for performing load balancing in a distributed computing environment
KR1020147021737A KR102013004B1 (en) 2012-02-03 2013-02-01 Dynamic load balancing in a scalable environment
EP13742889.2A EP2810163A4 (en) 2012-02-03 2013-02-01 Dynamic load balancing in a scalable environment

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/366,052 US8868711B2 (en) 2012-02-03 2012-02-03 Dynamic load balancing in a scalable environment
US13/366,052 2012-02-03

Publications (1)

Publication Number Publication Date
WO2013116664A1 true WO2013116664A1 (en) 2013-08-08

Family

ID=48903896

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2013/024366 WO2013116664A1 (en) 2012-02-03 2013-02-01 Dynamic load balancing in a scalable environment

Country Status (6)

Country Link
US (1) US8868711B2 (en)
EP (1) EP2810163A4 (en)
JP (1) JP6185486B2 (en)
KR (1) KR102013004B1 (en)
CN (1) CN104081353B (en)
WO (1) WO2013116664A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3467652A4 (en) * 2016-06-30 2019-05-15 Huawei Technologies Co., Ltd. Data balancing method and device

Families Citing this family (90)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8782434B1 (en) 2010-07-15 2014-07-15 The Research Foundation For The State University Of New York System and method for validating program execution at run-time
GB2495079A (en) 2011-09-23 2013-04-03 Hybrid Logic Ltd Live migration of applications and file systems in a distributed system
US10331801B2 (en) * 2011-09-23 2019-06-25 Open Invention Network, Llc System for live-migration and automated recovery of applications in a distributed system
US10311027B2 (en) 2011-09-23 2019-06-04 Open Invention Network, Llc System for live-migration and automated recovery of applications in a distributed system
US10860384B2 (en) 2012-02-03 2020-12-08 Microsoft Technology Licensing, Llc Managing partitions in a scalable environment
US9852010B2 (en) 2012-02-03 2017-12-26 Microsoft Technology Licensing, Llc Decoupling partitioning for scalability
US20130219069A1 (en) * 2012-02-22 2013-08-22 Computer Associates Think, Inc. System and method for managing virtual hard disks in cloud environments
US9141430B2 (en) * 2012-04-30 2015-09-22 Hewlett-Packard Development Company, L.P. Scheduling mapreduce job sets
US9774676B2 (en) 2012-05-21 2017-09-26 Google Inc. Storing and moving data in a distributed storage system
US9449006B2 (en) 2012-06-04 2016-09-20 Google Inc. Method and system for deleting obsolete files from a file system
US9659038B2 (en) 2012-06-04 2017-05-23 Google Inc. Efficient snapshot read of a database in a distributed storage system
US9230000B1 (en) * 2012-06-04 2016-01-05 Google Inc. Pipelining Paxos state machines
WO2013184712A2 (en) 2012-06-04 2013-12-12 Google Inc. Systems and methods of increasing database access concurrency using granular timestamps
US9298576B2 (en) 2012-06-04 2016-03-29 Google Inc. Collecting processor usage statistics
US9015721B2 (en) * 2012-07-30 2015-04-21 Hewlett-Packard Development Company, L. P. Managing array computations during programmatic run-time in a distributed computing environment
US10652318B2 (en) * 2012-08-13 2020-05-12 Verisign, Inc. Systems and methods for load balancing using predictive routing
US9122873B2 (en) 2012-09-14 2015-09-01 The Research Foundation For The State University Of New York Continuous run-time validation of program execution: a practical approach
US9069782B2 (en) 2012-10-01 2015-06-30 The Research Foundation For The State University Of New York System and method for security and privacy aware virtual machine checkpointing
CN103809953B (en) * 2012-11-14 2018-09-07 腾讯科技(深圳)有限公司 Multimedia monitoring method and system
US9225638B2 (en) 2013-05-09 2015-12-29 Vmware, Inc. Method and system for service switching using service tags
US9489443B1 (en) * 2013-05-24 2016-11-08 Amazon Technologies, Inc. Scheduling of splits and moves of database partitions
US9197546B2 (en) * 2013-08-06 2015-11-24 Oracle International Corporation System and method for providing a messaging cluster with hybrid partitions
US20150089062A1 (en) * 2013-09-25 2015-03-26 Virtual Bridges, Inc. Methods and systems for dynamically specializing and re-purposing computer servers in an elastically scaling cloud computing infrastructure
US9886310B2 (en) 2014-02-10 2018-02-06 International Business Machines Corporation Dynamic resource allocation in MapReduce
US9513806B2 (en) 2014-05-01 2016-12-06 Microsoft Technology Licensing, Llc Dimension based load balancing
WO2015165545A1 (en) 2014-05-01 2015-11-05 Longsand Limited Embedded processing of structured and unstructured data using a single application protocol interface (api)
US9798303B2 (en) * 2014-05-01 2017-10-24 Rockwell Automation Technologies, Inc. Systems and methods for balancing loads in an industrial automation system
US9609054B2 (en) 2014-05-01 2017-03-28 Microsoft Technology Licensing, Llc Load balancing scalable storage utilizing optimization modules
US9513939B2 (en) 2014-05-19 2016-12-06 International Business Machines Corporation Agile VM load balancing through micro-checkpointing and multi-architecture emulation
US9286001B2 (en) * 2014-06-30 2016-03-15 Microsoft Licensing Technology Llc Effective range partition splitting in scalable storage
CN105302536A (en) 2014-07-31 2016-02-03 国际商业机器公司 Configuration method and apparatus for related parameters of MapReduce application
US11296930B2 (en) 2014-09-30 2022-04-05 Nicira, Inc. Tunnel-enabled elastic service model
US9935827B2 (en) 2014-09-30 2018-04-03 Nicira, Inc. Method and apparatus for distributing load among a plurality of service nodes
US10135737B2 (en) 2014-09-30 2018-11-20 Nicira, Inc. Distributed load balancing systems
US10776373B2 (en) * 2014-10-28 2020-09-15 Salesforce.Com, Inc. Facilitating elastic allocation of organization-specific queue resources in an on-demand services environment
US9798634B2 (en) * 2014-10-30 2017-10-24 Sap Se Distributed failover for multi-tenant server farms based on load times of replicated tenants
US10218779B1 (en) * 2015-02-26 2019-02-26 Google Llc Machine level resource distribution
US10609091B2 (en) 2015-04-03 2020-03-31 Nicira, Inc. Method, apparatus, and system for implementing a content switch
US9826030B1 (en) * 2015-06-04 2017-11-21 Amazon Technologies, Inc. Placement of volume partition replica pairs
US9826041B1 (en) 2015-06-04 2017-11-21 Amazon Technologies, Inc. Relative placement of volume partitions
KR101928529B1 (en) * 2015-06-10 2018-12-13 울산과학기술원 Code Distributed Hash Table based MapReduce System and Method
US10242018B2 (en) * 2016-04-18 2019-03-26 International Business Machines Corporation Page allocations for encrypted files
WO2017209788A1 (en) * 2016-06-03 2017-12-07 Google Llc Weighted auto-sharding
US10001983B2 (en) 2016-07-27 2018-06-19 Salesforce.Com, Inc. Rolling version update deployment utilizing dynamic node allocation
US11604795B2 (en) 2016-09-26 2023-03-14 Splunk Inc. Distributing partial results from an external data system between worker nodes
US12013895B2 (en) 2016-09-26 2024-06-18 Splunk Inc. Processing data using containerized nodes in a containerized scalable environment
US10353965B2 (en) 2016-09-26 2019-07-16 Splunk Inc. Data fabric service system architecture
US11860940B1 (en) 2016-09-26 2024-01-02 Splunk Inc. Identifying buckets for query execution using a catalog of buckets
US10956415B2 (en) 2016-09-26 2021-03-23 Splunk Inc. Generating a subquery for an external data system using a configuration file
US11620336B1 (en) 2016-09-26 2023-04-04 Splunk Inc. Managing and storing buckets to a remote shared storage system based on a collective bucket size
US10268726B1 (en) * 2017-04-20 2019-04-23 Amazon Technologies, Inc. Partition key management for improved throughput
US11921672B2 (en) 2017-07-31 2024-03-05 Splunk Inc. Query execution at a remote heterogeneous data store of a data fabric service
US12118009B2 (en) 2017-07-31 2024-10-15 Splunk Inc. Supporting query languages through distributed execution of query engines
US11989194B2 (en) * 2017-07-31 2024-05-21 Splunk Inc. Addressing memory limits for partition tracking among worker nodes
US10805181B2 (en) 2017-10-29 2020-10-13 Nicira, Inc. Service operation chaining
US11012420B2 (en) 2017-11-15 2021-05-18 Nicira, Inc. Third-party service chaining using packet encapsulation in a flow-based forwarding element
US10797910B2 (en) 2018-01-26 2020-10-06 Nicira, Inc. Specifying and utilizing paths through a network
US10659252B2 (en) 2018-01-26 2020-05-19 Nicira, Inc Specifying and utilizing paths through a network
US10805192B2 (en) 2018-03-27 2020-10-13 Nicira, Inc. Detecting failure of layer 2 service using broadcast messages
US10728174B2 (en) 2018-03-27 2020-07-28 Nicira, Inc. Incorporating layer 2 service between two interfaces of gateway device
US11334543B1 (en) 2018-04-30 2022-05-17 Splunk Inc. Scalable bucket merging for a data intake and query system
US11595250B2 (en) 2018-09-02 2023-02-28 Vmware, Inc. Service insertion at logical network gateway
US10944673B2 (en) 2018-09-02 2021-03-09 Vmware, Inc. Redirection of data messages at logical network gateway
US10977140B2 (en) 2018-11-06 2021-04-13 International Business Machines Corporation Fault tolerant distributed system to monitor, recover and scale load balancers
US11042397B2 (en) 2019-02-22 2021-06-22 Vmware, Inc. Providing services with guest VM mobility
JP7014752B2 (en) 2019-06-03 2022-02-01 矢崎総業株式会社 Heat storage material composition and heat storage system for heating and cooling of buildings
US11528194B2 (en) * 2019-09-06 2022-12-13 Jpmorgan Chase Bank, N.A. Enterprise control plane for data streaming service
US20210073043A1 (en) * 2019-09-11 2021-03-11 Synchronoss Technologies, Inc. Method and system for uniform, consistent, stateless and deterministic consistent hashing for fixed size partitions
US11494380B2 (en) 2019-10-18 2022-11-08 Splunk Inc. Management of distributed computing framework components in a data fabric service system
US11283717B2 (en) 2019-10-30 2022-03-22 Vmware, Inc. Distributed fault tolerant service chain
US11140218B2 (en) 2019-10-30 2021-10-05 Vmware, Inc. Distributed service chain across multiple clouds
KR102251863B1 (en) * 2019-11-01 2021-05-14 경희대학교 산학협력단 Task allocation method for vehicle cloud computing and server and vehicle data processing system performing the same
US11169855B2 (en) * 2019-12-03 2021-11-09 Sap Se Resource allocation using application-generated notifications
US11223494B2 (en) 2020-01-13 2022-01-11 Vmware, Inc. Service insertion for multicast traffic at boundary
US11847478B2 (en) 2020-01-17 2023-12-19 Vmware, Inc. Real-time feedback associated with configuring virtual infrastructure objects using tags
US11153406B2 (en) 2020-01-20 2021-10-19 Vmware, Inc. Method of network performance visualization of service function chains
US11659061B2 (en) 2020-01-20 2023-05-23 Vmware, Inc. Method of adjusting service function chains to improve network performance
US11922222B1 (en) 2020-01-30 2024-03-05 Splunk Inc. Generating a modified component for a data intake and query system using an isolated execution environment image
US11212356B2 (en) 2020-04-06 2021-12-28 Vmware, Inc. Providing services at the edge of a network using selected virtual tunnel interfaces
KR102260549B1 (en) * 2020-09-10 2021-06-04 한국전자기술연구원 Load balancing method based on resource utilization and geographic location in a associative container environment
US11675631B2 (en) 2020-09-29 2023-06-13 Kyndryl, Inc. Balancing mainframe and distributed workloads based on performance and costs
US11704313B1 (en) 2020-10-19 2023-07-18 Splunk Inc. Parallel branch operation using intermediary nodes
US11025710B1 (en) * 2020-10-26 2021-06-01 Verizon Digital Media Services Inc. Systems and methods for dynamic load balancing based on server utilization and content popularity
US11734043B2 (en) 2020-12-15 2023-08-22 Vmware, Inc. Providing stateful services in a scalable manner for machines executing on host computers
US11611625B2 (en) 2020-12-15 2023-03-21 Vmware, Inc. Providing stateful services in a scalable manner for machines executing on host computers
EP4016299A1 (en) * 2020-12-21 2022-06-22 Amadeus S.A.S. A method and a system for managing the computing resources in data regions
US11595469B2 (en) 2021-05-25 2023-02-28 Red Hat, Inc. Balancing data partitions among dynamic services in a cloud environment
US12072939B1 (en) 2021-07-30 2024-08-27 Splunk Inc. Federated data enrichment objects
KR20230070718A (en) * 2021-11-15 2023-05-23 에스케이하이닉스 주식회사 Key-value storage identifying tenants and operating method thereof
US12093272B1 (en) 2022-04-29 2024-09-17 Splunk Inc. Retrieving data identifiers from queue for search of external data system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020178262A1 (en) * 2001-05-22 2002-11-28 David Bonnell System and method for dynamic load balancing
US20060080273A1 (en) 2004-10-12 2006-04-13 International Business Machines Corporation Middleware for externally applied partitioning of applications
US20070067435A1 (en) * 2003-10-08 2007-03-22 Landis John A Virtual data center that allocates and manages system resources across multiple nodes
US20070143442A1 (en) * 2005-12-20 2007-06-21 Nec Laboratories America, Inc. Scalable Publish/Subscribe Broker Network Using Active Load Balancing
US20100262860A1 (en) * 2009-04-09 2010-10-14 Chandramouli Sargor Load balancing and high availability of compute resources

Family Cites Families (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6886035B2 (en) * 1996-08-02 2005-04-26 Hewlett-Packard Development Company, L.P. Dynamic load balancing of a network of client and server computer
US6801949B1 (en) 1999-04-12 2004-10-05 Rainfinity, Inc. Distributed server cluster with graphical user interface
US6415323B1 (en) 1999-09-03 2002-07-02 Fastforward Networks Proximity-based redirection system for robust and scalable service-node location in an internetwork
US6922685B2 (en) 2000-05-22 2005-07-26 Mci, Inc. Method and system for managing partitioned data resources
US7293255B2 (en) 2000-09-25 2007-11-06 Sun Microsystems, Inc. Apparatus and method for automated creation of resource types
US7356592B2 (en) 2002-01-24 2008-04-08 International Business Machines Corporation Method and apparatus for web farm traffic control
US6813637B2 (en) 2002-03-15 2004-11-02 Sun Microsystems, Inc. System and method for automatically partitioning an application between a web server and an application server
US7167678B2 (en) 2002-12-18 2007-01-23 Microsoft Corporation Persistent peer-to-peer networking over a piconet network
US7290260B2 (en) * 2003-02-20 2007-10-30 International Business Machines Corporation Dynamic processor redistribution between partitions in a computing system
US7761514B2 (en) 2003-11-26 2010-07-20 International Business Machines Corporation Method and apparatus for providing dynamic group management for distributed interactive applications
JP2005182702A (en) * 2003-12-24 2005-07-07 Fujitsu Ltd Access control system in ip network
US20060168107A1 (en) 2004-03-16 2006-07-27 Balan Rajesh K Generalized on-demand service architecture for interactive applications
US7979862B2 (en) 2004-12-21 2011-07-12 Hewlett-Packard Development Company, L.P. System and method for replacing an inoperable master workload management process
US7716180B2 (en) * 2005-12-29 2010-05-11 Amazon Technologies, Inc. Distributed storage system with web services client interface
US7958216B2 (en) 2006-03-31 2011-06-07 Sap Ag Method for implementing an unplanned shutdown of a computer system
CN101009648A (en) * 2006-12-27 2007-08-01 北京航空航天大学 Multi-server hierarchical alterative load balance method
US8180747B2 (en) 2007-11-12 2012-05-15 F5 Networks, Inc. Load sharing cluster file systems
US8392482B1 (en) 2008-03-31 2013-03-05 Amazon Technologies, Inc. Versioning of database partition maps
US9021071B2 (en) 2008-06-09 2015-04-28 International Business Machines Corporation Methods of federating applications providing modular data
WO2010041481A1 (en) 2008-10-10 2010-04-15 Hitachi, Ltd. Storage system and method for controlling the same
US8966017B2 (en) 2009-07-09 2015-02-24 Novell, Inc. Techniques for cloud control and management
US20110078303A1 (en) 2009-09-30 2011-03-31 Alcatel-Lucent Usa Inc. Dynamic load balancing and scaling of allocated cloud resources in an enterprise network
US8910176B2 (en) 2010-01-15 2014-12-09 International Business Machines Corporation System for distributed task dispatch in multi-application environment based on consensus for load balancing using task partitioning and dynamic grouping of server instance
US8364775B2 (en) 2010-08-12 2013-01-29 International Business Machines Corporation High availability management system for stateless components in a distributed master-slave component topology
US8665890B2 (en) * 2011-05-20 2014-03-04 The Regents Of The University Of California Hybrid cross-layer routing protocol for MANETs

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020178262A1 (en) * 2001-05-22 2002-11-28 David Bonnell System and method for dynamic load balancing
US20070067435A1 (en) * 2003-10-08 2007-03-22 Landis John A Virtual data center that allocates and manages system resources across multiple nodes
US20060080273A1 (en) 2004-10-12 2006-04-13 International Business Machines Corporation Middleware for externally applied partitioning of applications
US20070143442A1 (en) * 2005-12-20 2007-06-21 Nec Laboratories America, Inc. Scalable Publish/Subscribe Broker Network Using Active Load Balancing
US20100262860A1 (en) * 2009-04-09 2010-10-14 Chandramouli Sargor Load balancing and high availability of compute resources

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2810163A4

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3467652A4 (en) * 2016-06-30 2019-05-15 Huawei Technologies Co., Ltd. Data balancing method and device

Also Published As

Publication number Publication date
JP2015506523A (en) 2015-03-02
EP2810163A4 (en) 2016-08-10
CN104081353A (en) 2014-10-01
JP6185486B2 (en) 2017-08-23
KR20140119090A (en) 2014-10-08
KR102013004B1 (en) 2019-08-21
CN104081353B (en) 2018-08-03
US8868711B2 (en) 2014-10-21
EP2810163A1 (en) 2014-12-10
US20130204991A1 (en) 2013-08-08

Similar Documents

Publication Publication Date Title
US11886925B2 (en) Managing partitions in a scalable environment
US10635500B2 (en) Decoupling partitioning for scalability
US8868711B2 (en) Dynamic load balancing in a scalable environment
US11729113B2 (en) Translating high level requirements policies to distributed storage configurations
US9460185B2 (en) Storage device selection for database partition replicas
US10769036B2 (en) Distributed transaction log
US8386610B2 (en) System and method for automatic storage load balancing in virtual server environments
US8392482B1 (en) Versioning of database partition maps
US20180004777A1 (en) Data distribution across nodes of a distributed database base system
US20160050123A1 (en) Fault tolerant federation of computing clusters
Μπέκας Service management in NoSQL data stores via replica-group reconfigurations
US8898289B1 (en) Distributed event processing method and architecture

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13742889

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 20147021737

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2014555763

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2013742889

Country of ref document: EP