WO2005017783A2 - Hierarchical management of the dynamic allocation of resourses in a multi-node system - Google Patents

Hierarchical management of the dynamic allocation of resourses in a multi-node system Download PDF

Info

Publication number
WO2005017783A2
WO2005017783A2 PCT/US2004/026389 US2004026389W WO2005017783A2 WO 2005017783 A2 WO2005017783 A2 WO 2005017783A2 US 2004026389 W US2004026389 W US 2004026389W WO 2005017783 A2 WO2005017783 A2 WO 2005017783A2
Authority
WO
WIPO (PCT)
Prior art keywords
service
node
database
server
resources
Prior art date
Application number
PCT/US2004/026389
Other languages
French (fr)
Other versions
WO2005017783A3 (en
Inventor
Benny Souder
Debashish Chatterjee
Lakshminarayanan Chidambaran
David Brower
Carol Colrain
Daniel Semler
Christopher Kantarjiev
James W. Stamos
Original Assignee
Oracle International Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/718,747 external-priority patent/US7873684B2/en
Priority claimed from US10/917,873 external-priority patent/US7516221B2/en
Priority claimed from US10/918,056 external-priority patent/US7437460B2/en
Priority claimed from US10/917,715 external-priority patent/US7664847B2/en
Priority claimed from US10/918,055 external-priority patent/US7552171B2/en
Priority claimed from US10/917,687 external-priority patent/US7441033B2/en
Priority claimed from US10/918,054 external-priority patent/US8365193B2/en
Priority claimed from US10/917,661 external-priority patent/US7437459B2/en
Priority claimed from US10/917,953 external-priority patent/US7552218B2/en
Priority to JP2006523412A priority Critical patent/JP4970939B2/en
Priority to CA002533744A priority patent/CA2533744C/en
Application filed by Oracle International Corporation filed Critical Oracle International Corporation
Priority to EP04781129.4A priority patent/EP1654648B1/en
Priority to AU2004266017A priority patent/AU2004266017B2/en
Publication of WO2005017783A2 publication Critical patent/WO2005017783A2/en
Publication of WO2005017783A3 publication Critical patent/WO2005017783A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5011Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals

Definitions

  • the present invention relates to work load management, and in particular, work load management within a multi-node computer system.
  • Enterprises are looking at ways of reducing costs and increasing efficiencies of their data processing system.
  • a typical enterprise data processing system allocates individual resources for each of the enterprise's applications. Enough resources are allocated up front for each application to handle the estimated peak load of the application.
  • Each application has different load characteristics; some applications are busy during the day; some others during the night; some reports are run once a week and some others once a month.
  • Grid computing enables the utilization or elimination of this unutilized capacity. In fact, Grid computing is poised to drastically change the economics of computing.
  • a grid is a collection of computing elements that provide processing and some degree of shared storage; the resources of a grid are allocated dynamically to meet the computational needs and priorities of its clients.
  • Grid computing can dramatically lower the cost of computing, extend the availability of computing resources, and deliver higher productivity and higher quality.
  • the basic idea of Grid computing is the notion of computing as a utility, analogous to the electric power grid or the telephone network.
  • a client of the Grid does not care where its data is or where the computation is performed. All a client wants is to have computation done and have the information delivered to the client when it wants.
  • This view of Grid computing as a utility is, of course, a client side view. From the server side, or behind the scenes, the Grid is about resource allocation, information sharing, and high availability. Resource allocation ensures that all those that need or request resources are getting what they need. Resources are not standing idle while requests are left unserviced. Information sharing makes sure that the information clients and applications need is available where and when it is needed. High availability ensures that all the data and computation must always be there — just as a utility company must always provide electric power.
  • Grid Computing for Databases ensures that all those that need or request resources are getting what they need. Resources are not standing idle while requests are left unserviced. Information sharing makes sure that the information clients and applications need is available where and when it is needed. High availability ensures that all the data and computation must always be there — just as a utility company must always provide electric power.
  • a grid can support multiple databases and dynamically allocate and reallocate resources as needed to support the current demand for each database. As the demand for a database increases, more resources are allocated for that database, while other resources are deallocated from another database. For example, on an enterprise grid, a database is being serviced by one database server running on one server blade on the grid. The number of users requesting data from the database increases. In response to this increase, a database server for another database is removed from one server blade and a database server for the database experiencing increased user requests is provisioned to the server blade.
  • Grid computing for databases requires allocation and management of resources at different levels.
  • the performance provided to the users of the database must be monitored and resources of the database allocated between the users to ensure performance goals for each of the users are met.
  • the allocation of a grid's resources between the databases must be managed to ensure that performance goals for users of all the databases are met.
  • the work to manage allocation of resources at these different levels and the information needed to perform such management is very complex. Therefore, there is a need for a mechanism that simplifies and efficiently handles the management of resources in a Grid computing system for database systems as well as other types of systems that allocate resources at different levels within a Grid.
  • FIG. 1 is a block diagram showing a multi-node computer system on which an embodiment of the present invention may be implemented.
  • FIG. 2 is a block diagram showing a cluster of nodes according to an embodiment of the present invention.
  • FIG. 3 is a block diagram showing components of a cluster of nodes and multi- node database server that participate in providing various services for a database according to an embodiment of the present invention.
  • FIG. 4 is a flow chart showing a procedure for expanding a service to a target database instance and quiescing a service on the target database instance.
  • FIG. 5 is a flow chart showing the procedure for two-phase-quiescing.
  • FIG. 6 is a block diagram of a computer system that may be used in an embodiment of the present invention.
  • a service is work of a particular type or category that is performed for the benefit of one or more clients.
  • the service includes any use or expenditure of computer resources, including, for example, CPU processing time, storing and accessing data in volatile memory, read and writes from and to persistent storage (i.e. disk storage), and use of network or bus bandwidth.
  • a service may be, for example, work that is performed for a particular application on a client of a database server.
  • the approaches manage allocation of resources at different levels. For services that use a particular database, the performance realized by the services is monitored. Resources assigned to the database are allocated between these services to ensure performance goals for each are met. Resources assigned to a cluster of nodes are allocated between the databases to ensure that performance goals for all the services that use the databases are met.
  • the approach uses a hierarchy of directors to manage resources at the different levels.
  • One type of director a database director, manages resources allocated to a database among services that use a database and its database instances.
  • the database director manages the allocation of database instances among the services.
  • a cluster director manages resources of a cluster of nodes between databases whose database servers are hosted on the cluster.
  • Yet another director, a farm director manages resources allocated between clusters.
  • FIG. 1 shows a multi-node computer system that may be used to implement an embodiment of the present invention.
  • FIG. 1 shows cluster farm 101.
  • a cluster farm is a set of nodes that is organized into groups of nodes, referred to as clusters.
  • Clusters provide some degree of shared storage (e.g. shared access to a set of disk drives) between the nodes in the cluster.
  • the nodes in a cluster farm may be in the form of computers (e.g. work stations, personal computers) interconnected via a network.
  • the nodes may be the nodes of a grid.
  • a grid is composed of nodes in the form of server blades interconnected with other server blades on a rack.
  • Each server blade is an inclusive computer system, with processor, memory, network connections, and associated electronics on a single motherboard.
  • server blades do not include onboard storage (other than volatile memory), and they share storage units (e.g. shared disks) along with a power supply, cooling system, and cabling within a rack.
  • a defining characteristic of a cluster in a cluster farm is that the cluster's nodes may be automatically transferred between clusters within the farm through software control without the need to physically reconnect the node from one cluster to another.
  • a cluster is controlled and managed by software utilities referred to herein as clusterware.
  • Clusterware may be executed to remove a node from a cluster and to provision the node to a cluster.
  • Clusterware provides a command line interface that accepts requests from a human administrator, allowing the administrator to enter commands to provision and remove a node from a cluster.
  • the interfaces may also take the form of Application Program Interfaces ("APIs"), which may be called by other software being executed within the cluster farm.
  • APIs Application Program Interfaces
  • Clusterware uses and maintains metadata that defines the configuration of a cluster within a farm, including cluster configuration metadata, which defines the topology of a cluster in a farm, including which particular nodes are in the cluster.
  • the metadata is modified to reflect changes made to a cluster in a cluster farm by the clusterware.
  • An example of clusterware is software developed by OracleTM, such as Oracle9i Real Application Clusters or Oracle Real Application Clusters lOg. Oracle9i Real Application Clusters is described in Oracle9i RAC: Oracle Real Application Clusters Configuration and Internals, by Mike Ault and Madhu Tumma, 2nd edition (August 2, 2003).
  • Cluster farm 101 includes clusters 110, 120, and 130. Each of the clusters hosts one or more multi-node database servers that provide and manage access to databases.
  • FIG. 2 shows a cluster 110 according to an embodiment of the present invention.
  • a defining feature of a cluster is that is it treated as a single unit or entity by clients of the cluster because the clients issue requests for services hosted by the cluster without specifying what particular node or nodes carryout the request, as shall be described in greater detail.
  • Cluster 110 includes multi-node database servers 222, 232, and 242.
  • Multi- node database servers 222, 232, and 242 reside on one or more nodes of cluster 110.
  • a server such as a multi-node server, is a combination of integrated software components and an allocation of computational resources, such as memory, a node, and processes on the node for executing the integrated software components on a processor, where the combination of the software and computational resources are dedicated to performing a particular function on behalf of one or more clients.
  • a multi-node server governs and facilitates access to a particular database, processing requests by clients to access the database.
  • Multi-node servers 222, 232, and 242 govern and provide access to database 220, 230, and 240, respectively.
  • Another example of a server is a web server.
  • Resources from multiple nodes in a multi-node computing system can be allocated to running a server's software.
  • Each combination of the software and allocation of the resources from a node is a server that is referred to herein as a "server instance" or "instance".
  • a multi-node database server comprises multiple server instances that can run on multiple nodes.
  • Several instances of a multi-node database server can in fact run on the same node.
  • a multi-node database server comprises multiple "database instances", each database instance running on a node, and governing and facilitating access to a particular database.
  • each instance can be referred to herein as a database instance of the particular database.
  • Clusters are often used to host multi-node database servers.
  • Clients of cluster 110 as well as multi-node database servers 222, 232, and 242 include clients 203 and 205.
  • Clients 203 and 205 execute applications on computers interconnected to cluster 110 via, for example, a network.
  • An application is a unit of software that is configured to interact with and use the functions of a server. In general, applications are comprised of integrated functions and software modules (e.g. programs comprised of machine executable code or interpretable code, dynamically linked libraries) that perform a set of related functions.
  • Clients 203 for example, include computer processes executing an FIN application and PAY application.
  • the FIN application includes software that generates and analyzes accounting and financial information of an enterprise.
  • the PAY application generates and tracks information about employee compensation of an enterprise.
  • database server 222 may be a client of database server 232.
  • a database such as databases 220, 230, and 240, is a collection of database objects.
  • Database objects include any form of structured data.
  • Structured data is data structured according to a metadata description defining the structure.
  • Structured data includes relational tables, object tables, object-relational tables, and bodies of data structured according to the Extensible Markup Language (“XML”), such as XML documents.
  • XML Extensible Markup Language
  • a session is established for the client.
  • a session such as a database session, is a particular connection established for a client to a server, such as a database instance, through which the client issues a series of requests (requests for execution of database statements).
  • session state data is maintained that reflects the current state of a database session.
  • Such information contains, for example, the identity of the client for which the session was established, and temporary variable values generated by processes executing software within the database session.
  • a client establishes a database session by transmitting a database connection request to cluster 110.
  • a client of cluster 110 may interact with cluster 110 through client side interface components that reside on the same computer of the client.
  • the client side interface components include API functions that are invoked by an application executed by clients 203 and 205.
  • connection information identifying the node is received by the client side interface components and used by them to transmit subsequent requests by the client to the node.
  • the database session assigned to a client may be migrated to another database instance.
  • Migrating a database session entails creating a new database session on another node. While information about the migration and connection is transmitted to the client side interface components, the information is not accessible to the application and the application is "unaware" of the migration. In this way, the migration is performed transparently to the application. Requests that would be associated with the old database session are performed within the new database session. Techniques for migrating database sessions in this way are described in Transparent Session Migration Across Servers (50277-2383). Services
  • a service is work of a particular type or category that is performed for the benefit of one or more clients.
  • Cluster 110 provides to clients 203 and 205 a database service for accessing database 220, a database service for accessing database 230, and a database service for accessing database 240.
  • a database service is work that is performed by a database server for a client, typically including work to process and/or compute queries that require access to a database.
  • the term query as used herein refers to a database statement that conforms to a database language, such as SQL, and includes database statements that specify operations to add, delete, or modify data and create and modify database objects, such as tables, objects views, and executable routines.
  • a database service may be further divided or categorized into subcategories.
  • Database services for database 220 are further divided into the FIN service and PAY service.
  • the FIN service is the database service performed by database server 222 for the FIN application. Typically, this service involves accessing database objects on database 220 that store database data for FIN applications.
  • the PAY services are database services performed by database server 222 for the PAY application. Typically, this service involves accessing database objects on database 220 that store database data for PAY applications.
  • work by a cluster may be divided or categorized in categories and subcategories of services
  • the present invention is not limited to any particular way.
  • the work may be divided into services based on users or groups of users (business enterprises, divisions within an enterprise) for which the work is performed. Participants That Provide Services
  • Fig. 3 shows components of cluster 110 and multi-node database server 222 that participate in providing various services for database 220.
  • multi- node database server 222 includes database instances 322, 332, 342, 352, and 362, which reside on nodes 320, 330, 340, 350, and 360, respectively, and manage access to database 220.
  • Database instances may be provisioned or removed from a particular node using an instance management application.
  • Instance management applications are available, for example, as part of Oracle9i Real Application Clusters or Oracle Real Application Clusters lOg.
  • Instance management applications provide command line interfaces or APIs that may be invoked by administrators or clients to provision or remove a database instance from a node.
  • Database instances of multi-node database server 222 have been allocated to a particular service.
  • Database instances 322 and 332 have been allocated to service FIN.
  • Database instances 342 and 352 have been allocated to service PAY.
  • Instance 362 has not been allocated to any service.
  • a service is referred to as running or residing on or being hosted by an instance, node, or cluster when the instance, node, or cluster has been allocated to perform the service.
  • the FIN service is referred to as running or residing on database instances 322 and 332, and on nodes 320 and 330.
  • Listener 390 is a process running on cluster 110 that receives client database connection requests and directs them to a database instance within cluster 110.
  • the client connection requests received are associated with a service, (e.g. service FIN, PAY)
  • the client request is directed to a database instance hosting the service, where a database session is established for the client. As mentioned previously, the session may be migrated to another database instance.
  • Listener 390 directs the request to the particular database instance and/or node in a way that is transparent to the application. Listener 390 may be running on any node within cluster 110.
  • the client may issue additional requests, which may be in the form of function or remote procedure invocations, and which include requests to begin execution of a transaction, to execute queries, to perform updates and other types of transaction operations, to commit or otherwise terminate a transaction, and to terminate a database session.
  • Resources are allocated and re-allocated to meet levels of performance and cardinality constraints on the resources.
  • Levels of performance and resource availability established for a particular service are referred to herein as service-level agreements.
  • Levels of performance and cardinality constraints on resources that apply to a multi-node system in general and not necessarily to a particular service are referred to herein as policies.
  • a service-level agreement for service FIN maybe require as a level of performance that the average transaction time for service FIN be no more than a given threshold, and as an availability requirement that at least two instances host service FIN.
  • a policy may require that the CPU utilization of any node should not exceed 80%.
  • Policies may also be referred to herein as backend policies because they are used by backend administrators to manage overall system performance and to allocate resources between a set of services when it is deemed there are insufficient resources to meet service-level agreements of all the set of services. For example, a policy assigns a higher priority to a database relative to another database. When there are insufficient resources to meet service-level agreements of services of both databases, the database with the higher priority, and the services that use the database, will be favored when allocating resources.
  • a workload monitor such as workload monitor 388
  • workload monitor 388 is a distributed set of processes that run on nodes of a cluster to monitor and measure workload of the cluster and generate "performance metrics".
  • Workload monitor 388 runs on cluster 110.
  • Performance metrics is data that indicates the level of performance for one or more resources or services based on performance measures. Approaches for performing these functions are described in Measuring Workload by Sendee (50277-2337).
  • the information generated is accessible by various components within multi-node database server 222 that are responsible for managing the allocation of resources to meet service- level agreements, as shall be described in greater detail later.
  • a performance metric of a particular type that can be used to gauge a characteristic or condition that indicates a level of performance or workload is referred to herein as a performance measure.
  • a performance measure includes for example, transaction execution time or percent of CPU utilization.
  • service-level agreements that involve levels of performance can be defined by thresholds and criteria that are based on performance measures.
  • execution time of a transaction is a performance measure.
  • a service-level agreement based on this measure is that a transaction for service FIN should execute within 300 milliseconds.
  • Yet another performance measure is percentage CPU utilization of a node.
  • a backend policy based on this measure is that a node experience no more than 80% utilization.
  • Performance metrics can indicate the performance of a cluster, the performance of a service running on a cluster, a node in the cluster, or a particular database instance.
  • a performance metric or measure particular to a service is referred to herein as a service performance metric or measure.
  • a service performance measure for service FIN is the transaction time for transactions executed for service FIN.
  • Grid computing involves dynamically allocating computer resources to meet service-level agreements.
  • computer resources are balanced or adjusted at one or more levels of a resource allocation hierarchy.
  • Each level of the hierarchy has a different set of resource pools that are allocated between uses (e.g. services).
  • a resource pool is a group of resources of a particular type, for example, nodes and database instances that are available to a service, nodes that are available to a database or nodes that are available to a cluster.
  • the three levels in the resource allocation are the database level, the cluster level, and the farm level.
  • the resources pools allocated are those currently being used for a particular database, including database instances of the database and nodes that host them.
  • the resource pools of the database level i.e. resources that can be allocated at the database level
  • connection-time balancing Under connection-time balancing, listener 390 balances workload between service instances by directing database connection requests that require a particular service to an instance of database 220. For example, assume service FIN on database instance 322 is providing better service performance than other instances. Accordingly, listener 390 directs a greater portion of database connection requests requiring the FIN service to database instance 322.
  • run-time session balancing The second way to place a session is referred to as run-time session balancing.
  • database sessions are migrated from a database instance to another database instance.
  • the database sessions are migrated using transparent session migration.
  • transparent session migration As mentioned before, techniques for performing this are described in Transparent Session Migration Across Servers.
  • Service placement entails expanding and contracting services. Under service expansion and contraction, database instances are allocated to or deallocated from hosting services. When a database instance is assigned to host a service, more database sessions may be created for that service on that instance, thus increasing the number of database sessions associated with and available for the service. For example, to meet service-level agreements for service FIN, instances 322 and 332 are allocated to run service FIN. As demand for the service increases, the service-level agreements are no longer satisfied. When service-level agreements are not met, the service-level agreements are referred to as being violated. In response to this service-level violation, the FIN service is allocated an additional instance, instance 342. An instance can run more than one service. For purposes of illustration, an instance runs only one service. Thus, when FIN is added to instance 342, service PAY is "quiesced" from instance 342, that is, instance 342 is deallocated as a resource that is used for the PAY service and the service on the instance is ceased.
  • resources that are currently allocated to a cluster are balanced between databases (i.e. database services) to meet service-level agreements.
  • the resource pools balanced at this level are database instances and nodes hosting the database instances.
  • balancing resources at this level involves provisioning and quiescing instances among existing nodes in the cluster. For example, to meet service- level agreements in response to a service-level violation, instance 362 is provisioned to node 360, a node already in the cluster for multi-node database server 222.
  • the resources that can be allocated between services are the nodes in a cluster.
  • the pool of nodes for a cluster is dynamic. For example, in response to a service-level violation, node 370 is added to the cluster of multi-node database server 222. Hierarchy Of Actions to Adjust Resource Allocation
  • adjusting the resources at the lower level of the resource allocation hierarchy is less disruptive than at the higher levels.
  • Migrating database sessions and services to and from running instances at the database level is less disruptive and costly than provisioning or quiescing a new database instance at the cluster level. It is less expensive to reshuffle the assigned resources amongst the entities to which the resources are already assigned than to request more resources to be assigned from a higher level.
  • Service placement and session migration at the database level are less expensive than changing the number of database instances for the database at the cluster level.
  • migrating sessions to reshuffle them between instances already hosting a service is less expensive than service expansion and contraction because the latter impacts the service topology and may have a greater overall impact on load distribution.
  • FIN is provisioned to instance 352 and PAY is quiesced from the same instance, but only if quiescing service PAY does not violate service-level agreements for PAY. Otherwise, service FIN is expanded to another database instance. However, if there is no available instance to which to provision service FIN, then a new instance is provisioned. This is done by making allocation at the cluster level, provisioning a new instance 362 to node 360, a node already in the cluster. Then, an allocation may be made at the database level by expanding service FIN to instance 362, which at the time the service is provisioned, is an instance already allocated to database 220.
  • a distributed system component referred to as a distributed director, is responsible for managing workload of resources and resource allocation at each of the levels.
  • a system component is a combination of software, data, and one or more processes that execute that software and that use and maintain the data to perform a particular function.
  • a distributed system component is executed on multiple nodes.
  • a director is a system component of a database instance and is operating under the control of the database instance.
  • a distributed director includes directors executed on multiple nodes of cluster farm 101.
  • a director is responsible for managing the allocation of resources at one or more levels of the resource allocation hierarchy. Specifically, for each database, a director serves as a database director managing the allocation of resources at the database level. Other database instances of the database may have directors that serve as standby database directors, readying to take over as an "active" database director if the current active database director becomes unable to perform this role due to, for example, a system failure.
  • a director serves as the cluster director.
  • the cluster director is responsible for managing allocation of resources at the cluster level for a cluster.
  • Other directors within the cluster serve as standby cluster directors.
  • a director serves as the farm director.
  • the farm director is responsible for managing the allocation of resources at the farm level.
  • Other directors within the cluster serve as standby farm directors.
  • a director receives, maintains, and generates information needed to manage workload at that director's corresponding level.
  • director 380 is running on database instance 342.
  • Director 380 serves as database director for database 220, cluster director for cluster 110, and farm director for cluster farm 101.
  • Other directors act as a database director for databases 230 and 240, respectively, and as a cluster director for clusters 120 and 130, respectively.
  • a database director attempts to remedy a service-level violation by adjusting the allocation of resources at the database level. If the service-level violation requires an adjustment at the next higher level of the resource allocation hierarchy, the cluster level, the database director escalates the resolution of the service-level violation to the cluster director. If the service-level violation requires an adjustment at the highest level of the resource allocation hierarchy, the farm level, the cluster director escalates the resolution of the service-level violation to the farm director.
  • a remedy to a service-level violation may involve an adjustment of resource allocation by all directors.
  • the message queue is a table stored in the database of the database instance hosting the director.
  • the records or rows of the table correspond to a message queue.
  • the record indicates the status of any action taken in response to a message by the director responsible for responding to messages of that type.
  • the database director may request a database instance from the cluster director. The request is added to the queue.
  • the cluster director scans the message queue, detects the request, and acts upon it, updating the record to reflect actions undertaken to respond to the request.
  • the cluster director sends a message to the database director, which is placed in the message queue of the database director.
  • the advantage of using a table in database 220 is that it makes available the power and capability of a transaction oriented database system, such as multi-node database server 222, to store the message queue persistently and recoverably.
  • a transaction oriented database system such as multi-node database server 222
  • the standby director stepping in its place can access the message queue in a state consistent with the way the director left it when it failed.
  • Such a message queue and the use of it are described in Recoverable Asynchronous Message Driven Processing in a Multi-Node System (50277-2414).
  • the database director is responsible for monitoring service performance of services to ensure service-level agreements are met, expanding or contracting services from database instances of a database 220, and for migrating database sessions between database instances of database 220.
  • the database director also has access to and stores service performance metrics and service-level agreements for each service using the database. Based on these service performance metrics and service-level agreements, the database director ensures service-level agreements are met in two ways - (1) maintaining service performance compliance with service-level agreements by generating and sending information to the listener that allows the listener to balance workload between service instances; and (2) remedying service-level violations by detecting them and initiating adjustments to resource allocations in order to remedy the service-level violations.
  • connection-time balancing is used.
  • the director 380 generates and provides information, referred to herein as performance grades, to listener 390.
  • Performance grades guide listener 390 in balancing workload to maintain service performance compliance. Performance grades indicate relative service performance of a service on an instance relative to other instances. Based on the performance grades, listener 390 skews the distribution of connection user requests for the service to database instances providing better service performance. Techniques for generating performance grades and distributing user requests in this way based on performance grades are described in Calculation of Service Performance Grades in a Multi-Node Environment That Hosts the Services (50277-2410).
  • Remedying a service-level violation requires detecting the service-level violation.
  • Director 380 detects a service-level violation for a service by comparing service performance metrics to service-level agreements. For example, director 380 receives from workload monitor 388 service performance metrics that indicate that the average transaction time for service FIN exceeds 30 milliseconds. A service-level agreement for service FIN requires that average transaction time be no more than 20 milliseconds. By comparing the actual average transaction time to the service-level agreement, director 380 detects a service-level violation.
  • database director When the database director detects a service-level violation for a service, it attempts the least disruptive and costly resource allocation adjustments before attempting the more disruptive and costly resource allocation adjustments, in accordance with the resource allocation hierarchy. To this end, database director first determines whether it may remedy the service-level violation by balancing the workload between instances already hosting a service by migrating database sessions allocated to the service to another database instance where the service performance is better. The number of database sessions migrated is chosen in a way that is targeted to achieve a balanced load between the service instances. [0087] If the database director determines that a service-level violation should not be solved by rebalancing load between existing service instances, director 380 attempts to remedy the service-level violation by expanding a service, i.e.
  • the target database instance by allocating another existing database instance to host the service, referred to herein as the target database instance. If the target database instance is not hosting a service, the service is expanded by allocating the target database instance to host the service. If the target database instance hosts another service, the database director may quiesce the service if the database director determines that doing so will not cause a service-level violation for the other service.
  • FIG. 4 depicts a flowchart of a process followed by director 380, as database director for database 220, to expand a service to an additional database instance ("target database instance") when a service on the target database instance must first be quiesced.
  • service PAY is hosted by database instances 342, 352, and 362, the latter being on node 360.
  • Service FIN is being expanded to database instance 342.
  • the service instance of service PAY running on instance 342 is being quiesced.
  • director 380 sends a blocking message to listener 390.
  • the blocking message instructs listener 390 to cease distributing user requests for service PAY to target database instance 342.
  • database director 380 migrates database sessions on a target database instance to the other database instances hosting service FIN.
  • the database sessions are distributed between the other database instances in a way that balances workload between them.
  • director 380 sends a service activation message to listener 390.
  • the service activation message instructs listener 390 that service FIN is running on instance 362.
  • director 380 uses run-time balancing to balance the workload between instances on which service FIN is running (i.e. instances 322, 332, 342).
  • director 380 may determine that no service can contract to make room for the expansion of another service without violating service-level agreements on a target instance. In this case, director 380 may chose to expand the service to a database instance not yet hosting any service. If no database instance is available, director 380 requests one from the cluster director. In response, the cluster director provisions another database instance as requested, and notifies director 380. Director 380 expands the service to the new database instance.
  • Service-level agreements may limit the cardinality of a service.
  • service-level agreements for service FIN require that at least one but no more than three database instances host service FIN, and that service PAY be hosted by at least three database instances.
  • a service may be quiesced for reasons other than to make a database instance available for expansion of another service.
  • cardinality constraints for service FUST may vary based on the time of the day.
  • the cardinality of service FIN may be as high as three but during non-business hours, cardinality may be no higher than one.
  • database director 380 contracts service FIN by quiescing the service on two of the database instances.
  • a database director may need to respond to requests made by a cluster director. Such actions include responding to requests by the cluster director to quiesce a database instance, that is, quiesce the services currently hosted by a database instance. This step is needed when the cluster director wishes to replace a database instance for a database with a database instance for another database.
  • Cluster Director
  • the cluster director is responsible for provisioning and removing database instances from existing nodes in the cluster.
  • the cluster director also enforces database level policies.
  • Database level policies require, for example, that the cardinality of database instances for a database fall within a minimum and/or maximum, or require that when there are insufficient resources to meet all service-level agreements, that the cluster director skew the allocation of database instances to databases designated as having a higher priority for resource allocation. Skewing between databases in this way also skews the allocation of resources to services using the higher prioritized databases.
  • the cluster director has access to and stores data specifying the priority for resource allocation between databases. Such data may be configured by an administrator of a cluster.
  • the cluster director provisions and removes database instances in response to a database director's request for a database instance ("NEED-INSTANCE" request). If there is a node within the cluster that is not hosting a database instance (a "free node"), then the cluster director allocates another node to the database by provisioning a database instance to the free node.
  • cluster director may request one from the farm director by issuing a "NEED-NODE" request to the farm director. If the farm director is unable to provide one, then the cluster director arbitrates allocation of database instances between databases hosted by the cluster. The arbitration may entail removing a database instance of a database from a node in a cluster and provisioning a database instance for the database for which the NEED-INSTANCE request was generated.
  • FIG. 5 is a flow chart depicting a process for arbitrating the allocation of database instances between databases.
  • the database director for database 220 has determined that another database instance is needed for a service, and has generated a NEED-INSTANCE request for the cluster director, which is director 380.
  • Director 380 in its capacity as cluster director, determines that a database instance for another database should be removed from a node within cluster 110 so that the node may be used to provision another database instance for database 220.
  • director 380 transmits a "VOLUNTEER-TO-QUIESCE" request to database directors other than the requesting database director (i.e. the director issuing the NEED-INSTANCE request).
  • the purpose of the VOLUNTEER-TO-QUIESCE request is to ask a database director whether it may quiesce a database instance i.e. may reduce the cardinality of database instances for the director's database.
  • a database director of a database may respond by volunteering, transmitting a message indicating that a database instance for the database may be quiesced.
  • a database director may decline to quiesce a database instance i.e. reduce the cardinality of database instances.
  • One reason a database director may send a message declining the request is that all the director's database instances are needed to satisfy availability requirements for a service.
  • database director 380 receives at least one message from a database director affirming that a database instance may be quiesced, that is, more than one database director volunteers, then at step 520, director 380 selects a database from among those who volunteered. A database with a lower resource allocation priority may be selected in favor of a database with a higher resource allocation priority. For purposes of illustration, director 380 selects database 240. The cluster director 384 sends a message to the database director of database 240 to quiesce a database instance of database 240 from a node hosting the database instance.
  • database director 380 receives no message from a database director affirming that a database instance may be quiesced, that is, no database director volunteers, then at step 530, director 380 selects a database with a lower resource allocation priority. Director 380 then transmits a message to the database director of the selected database to quiesce a database instance. Techniques for selecting a database instance to quiesce are described in further detail in On Demand Node and Server Allocation and Deallocation (50277-2413).
  • the database director of database 240 quiesces a database instance on the node and transmits a notification to the cluster director, director 380, that the database instance is quiesced ("INSTANCE-IDLE" notification).
  • director 380 receives the notification.
  • cluster director 380 removes the database instance from the node and provisions a database instance for database 230 to the node, using, for example, instance provisioning APIs of clusterware.
  • the services offered by the database relinquishing the node may experience service-level violations.
  • the cardinality of nodes in the cluster maybe increased, as shall be explained in greater detail, and thus free nodes may become available and may be used to remedy such service-level violations.
  • Director 380 as cluster director, monitors cluster configuration metadata to detect when more free nodes become available, and may allocate them to a database incurring service-level violations. For example, in the current example, the database director for database 220, after relinquishing the database instance, continues to detect service-level violations, and, in response, transmits NEED- INSTANCE requests to its cluster director, director 380. Eventually, after detecting that more free nodes have been added to cluster 110, director 380 is able to respond to one of the requests by provisioning a database instance to the free node. Farm Director
  • the farm director is responsible for allocating nodes between clusters in a cluster farm, shuffling nodes between them by removing a node from one cluster and provisioning the node to another cluster.
  • the cluster director also enforces cluster-wide policies. Cluster-wide policies may require, for example, that the cardinality of nodes within a cluster fall within a minimum and/or maximum.
  • a farm director In response to a NEED-NODE request from a cluster director, a farm director provisions a node to a cluster and removes a free node from another cluster, subject to cluster service-level agreements.
  • a farm director corresponds with the cluster directors in a cluster farm to arbitrate the allocation of a node. This process entails transmitting a "RELINQUISH-NODE" request to the cluster directors, who respond to indicate whether or not they may relinquish a node.
  • the farm director selects a cluster and interacts with the cluster director of the selected cluster to remove the node from the cluster. Removing the node from the cluster may entail quiescing a database instance. Once the database instance is quiesced, a farm director removes the node from the selected cluster and provisions the node to the cluster needing the node by invoking clusterware using APIs provided for this purpose.
  • a farm director monitors the performance of clusters in the cluster farm using the performance metrics received from workload monitors executing on each of the clusters. If performance metrics indicate that a cluster is performing in violation of service-level agreements for the cluster or is not performing as well as the other cluster, then the farm director shifts one or more nodes from a better performing cluster to the worse performing cluster.
  • election of Directors As mentioned before, within the database instances of a database, there are multiple directors that may serve as the active database director or standby director. Therefore, there is a need for a mechanism to determine which director is the active director. Furthermore, when the active director fails, there is a need to select a standby database director to be the active database director. The same kind of need exists for the cluster directors and farm director. The process of selecting an active director is referred to herein as director election.
  • a database global lock is used to synchronize processes running under control of all the database instances of a database.
  • a database director Upon startup, a database director requests an exclusive lock. If no database director holds the lock, then the database director is granted the lock and assumes the position of database director. Other directors who subsequently request the lock are not granted the lock and assume the position of standby director until the lock is granted, if at all. The requests remain pending until granted or rescinded by the requestor.
  • Multi-node database servers detect when the holders of a database global lock experience system failure. In this case, the database global lock of a failed holder is cancelled or released and a pending request for the database global lock is granted to a standby database director. The standby database director whose lock is granted then assumes the role of database director.
  • a process group is a group that may be joined by processes executing on any node within a cluster. Members of the group are informed when another member stops running (e.g. due to system failure) or leaves the process group. In addition, members are assigned an id when joining the group. [0114] When a director starts up, it joins a process group for its database. If upon joining a director has the highest member id, the director assumes the position of database director for the database. When the active director stops running, the members of the process group are informed and the one with the highest member id assumes the role of database director.
  • Similar techniques may be used for election of cluster directors.
  • a cluster- wide lock may be used for director election or a process group for a cluster may be used.
  • the director that assumes the cluster director role may be a director for a database different than that of the previous active cluster director. Consequently, the message queue may reside on a different database. Tables on a database may be accessed more efficiently by processes residing on database instances for the database than processes residing on a database instance of another database.
  • a technique that can be used is to ensure that the standby director assuming the role of an active director is hosted by a database instance for the same database is the static data designation technique.
  • a database is designated as the one that hosts (i.e. whose database instances host) the cluster director for a cluster. Only directors for that database play the role of cluster director for a cluster. Director election among these directors may be performed using either a global database lock or a process group for the database.
  • the database may be designated by an administrator using an interface provided by clusterware for this purpose. To enhance standby cluster director availability, a database with high priority and high minimum/maximum cardinality requirements can be designated, ensuring a relatively large number of standby directors.
  • the static database designation approach may be used to elect an active farm director.
  • Organizing the management of resources within a farm cluster using the hierarchy of directors facilitates the generation and exchange of information needed to manage the resources. In general, processes running within a particular database (i.e. processes of a database instance for a particular database) are able to communicate with other processes within the database more efficiently than processes not within the database.
  • a director that is serving as a database director for a database and that obtains service performance metrics from one or more work load monitors running within the database, is able to obtain that data more efficiently than directors running within other databases.
  • the work of getting and generating information for the services running a database such as service performance metrics, service-level agreements, message queue data, need only be performed by one director.
  • an embodiment of the present invention has been illustrated by dynamically allocating the resources of a multi-node system among database services and subcategories of database services.
  • the present invention is not so limited.
  • an embodiment of the present invention may be used to allocate computer resources of a multi-node system that hosts an application server among services provided by the application server.
  • An application server is part of, for example, a three tier architecture in which an application server sits between clients and a database server.
  • the application server is used primarily for storing, providing access to, and executing application code, while a database server is used primarily for storing and providing access to a database for the application server.
  • the application server transmits requests for data to the database server.
  • the requests may be generated by an application server in response to executing the application code stored on the application server.
  • An example of an application server is Oracle 9i Application Server or Oracle lOg Application Server. Similar to examples of a multi-node server described herein, an application server may be distributed as multiple server instances executing on multiple nodes, the server instances hosting multiple sessions that may be migrated between the server instances.
  • the present invention is also not limited to homogenous multi-node servers comprised only of server instances that execute copies of the same software product or same version of a software product.
  • a multi-node database server may be comprised of several groups of server instances, each group executing different database server software from a different vendor, or executing a different version of database server software from the same vendor.
  • Transparent Session Migration allows a client to be switched from a session on a server to another session on another server in a way that is a transparent to the application for which the initial session was established.
  • migrate refers to an operation in which a client of an existing session on a server is switched from the existing session to another the session, allowing the existing session to be terminated and the client to use the other session in lieu of the existing session.
  • the existing session is referred to herein as having been migrated.
  • transparent refers to performing an operation in a way that, with respect to a unit of software, does not require execution of instructions in the unit that are tailored to perform the operation.
  • a client-side interface component through which the application interacts with the server, handles details of migration, modifying the internal state of the client-side interface component to effect the same.
  • Legacy applications do not have to be modified in order to institute techniques described herein.
  • the session's state is captured and restored. Capturing a session's state entails producing a stream of bytes, as a veritable copy of a session state of the session, that can be stored in an object, file, or other type of data structure and later accessed to restore the session.
  • the session of a client is captured on a source server, producing a stream of bytes that is stored in a data structure and transported to the destination server, where the destination server restores the session by loading the stream of bytes into a session on the destination server that was established for the client.
  • a database session may be "statefull" when a future application call depends on session state generated by prior application calls. Because of the potential dependency of future applications calls upon session state of a statefull session, migration of statefull database sessions entails determining whether portions of session state stored on the source instance may be transferred to another database instance, and if so, transferring copies of the session state between a source database instance and destination instance. In an embodiment, to determine whether portions of session state stored on the source instance may be transferred to another database instance, various migration checks are made to determine whether session state may be transferred to another database instance. These migration checks include detenriining whether a database session is at a transaction boundary, is at a call boundary, or is at a component boundary.
  • a database session is at a transaction boundary if there are currently no active transactions being executed for the session.
  • a transaction is a logical unit of work that is performed as an atomic unit.
  • the database must reflect all the changes made by a transaction, or none of the changes made by the transaction to ensure the integrity of the database. Consequently, none of the changes made by a transaction are permanently applied to a database until the transaction has been fully executed.
  • a transaction is said to "commit" when the changes made by the transaction are made permanent.
  • a transaction is active if the transaction has not been committed, aborted, or otherwise terminated.
  • a session is at a call boundary if a database instance has finished executing a client call rather being at an intermediate stage of processing the call.
  • a database instance goes through stages, each stage corresponding to a particular type of operation.
  • the stages are (1) creating a cursor, (2) parsing the database statement and binding its variables, (3) executing the database statement, (4) fetching the rows to return for the query, and (5) closing the cursor.
  • These stages are described in greater detail in Oracle ⁇ Server Concepts, Release 8.0, Volume 3 (the contents of which incorporated herein by reference), at chapter 23.
  • Intermediate stages are the operations performed before processing of the call is complete.
  • the intermediate stages are stages (1) - (5). After an instance performs step (5) in response to a call, a source session is at a call boundary.
  • a session is at a component boundary if each "database component" of a session is at its respective component boundary.
  • a database component is a set of software modules on a database server that provide specialized and related functions.
  • a component session state is specifically generated and used by a database component. According to an embodiment, session state can be viewed as a union or combination of component session states.
  • the following are examples of database components.
  • a cursor component manages cursors within a database instance.
  • a cursor is an area of memory used to store information about a parsed database statement and other information related to the processing of the database statement.
  • a PL/SQL component is a database component responsible for executing code (e.g.
  • Session parameters component is a database component responsible for managing attributes that generally control how calls and requests associated with a session are processed. The attributes are stored in a component session state. For example, session parameters can include an attribute that controls the particular human language for results returned by executing queries.
  • a Java component is a database component that is responsible for executing code (e.g. class and object methods) written in Java. The Java component uses component session state to store information related to the execution of Java code. [0132] A database session is at a component boundary for a particular database component if the component session state of the database component can be migrated to another session.
  • a database component provides a function that returns a value indicating whether or not the database component's respective component session state may be migrated.
  • a reason component session state of a database session may not be migrated is that component session state includes a file descriptor of an open file. The file descriptor contains information that is only valid for an instance hosting the database session.
  • the function is part of an interface of callback functions supported by each database component. The functions are invoked to capture and restore component session state, as well as to determine whether a component session state permits a session to be migrated.
  • Performance metrics are data that indicates the quality of service realized by services for one or more resources.
  • a background process generates performance metrics from performance statistics that are generated for each session and service hosted on a database instance.
  • performance statistics can indicate a quality of performance.
  • performance statistics in general, include more detailed information about specific uses of specific resources.
  • Performance statistics include, for example, how much time CPU time was used by a session, the velocity of a call, the number of calls a session made, the response time required to complete the calls for a session, how much CPU processing time was used to parse queries for the session, how much CPU processing time was used to execute queries, how many logical and physical reads were performed for the session, and wait times for input and output operations to various resources, such as wait times to read or write to a particular set of data blocks. Performance statistics generated for a session are aggregated by services associated with the session.
  • a session object For each session established on a database instance, a session object is created as part of establishing the session on the database instance.
  • a session object contains items of information used by a database instance to manage a session. Among these items of information are a service identifier defined for the session.
  • the service id is a hash value generated from the service name of the service.
  • Performance statistics are generated and aggregated by processes performing work requested within a session. For example, a database session is established for a client, who transmits a request to execute a query. A process for the session executes the database server software, parses the query and formulates an execution plan for computing the query in parallel.
  • the database server software that parses the database query, formulates the execution plan and computes the query also generates and aggregates performance statistics.
  • the database server software aggregates the performance statistics by session and service of the session.
  • the performance statistics are stored in a performance statistics repository, which may be, for example, an in- memory fixed table that is associated with a session or that is associated with a service disconnected from any session.
  • Performance statistics for a session are aggregated and stored within an in-memory table for the session.
  • Performance statistics for the service are aggregated and stored within the in-memory table for the service.
  • a session is associated with the service PAY.
  • To compute a query a database process assigned to the session uses .4 seconds of CPU time.
  • a workload monitor accesses the performance statistics repository to generate performance metrics and stores the generated performance metrics in a performance metrics repository, which is preferably an in- memory database table.
  • a workload monitor may every five seconds access the performance statistics repository to generate performance metrics, including performance grades.
  • the performance grades are transmitted to a listener for connection-time balancing. The listener uses the performance grades to perform connection-time balancing, establishing sessions for connection requests to services on database instances offering superior performance for the services.
  • a workload monitor can also compare performance metrics as they are generated to detect violations of service-level agreements. When a performance violation for a service is detected, the workload monitor sends messages to alert a director (e.g. database director) responsible for responding to the service-level violation.
  • a director e.g. database director
  • a service and its various attributes and quality of service e.g. service-level agreements
  • a database server provides a command line interface that accepts commands from human administrators to create and modify definitions for services of a database server. The definitions are stored as database configuration data in a dictionary in a database as a "service profile".
  • FIG. 6 is a block diagram that illustrates a computer system 600 upon which an embodiment of the invention may be implemented.
  • Computer system 600 includes a bus 602 or other communication mechanism for communicating information, and a processor 604 coupled with bus 602 for processing information.
  • Computer system 600 also includes a main memory 606, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 602 for storing information and instructions to be executed by processor 604.
  • Main memory 606 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 604.
  • Computer system 600 further includes a read only memory (ROM) 608 or other static storage device coupled to bus 602 for storing static information and instructions for processor 604.
  • ROM read only memory
  • a storage device 610 such as a magnetic disk or optical disk, is provided and coupled to bus 602 for storing information and instructions.
  • Computer system 600 may be coupled via bus 602 to a display 612, such as a cathode ray tube (CRT), for displaying information to a computer user.
  • An input device 614 is coupled to bus 602 for communicating information and command selections to processor 604.
  • cursor control 616 such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 604 and for controlling cursor movement on display 612.
  • This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane.
  • the invention is related to the use of computer system 600 for implementing the techniques described herein. According to one embodiment of the invention, those techniques are performed by computer system 600 in response to processor 604 executing one or more sequences of one or more instructions contained in main memory 606. Such instructions may be read into main memory 606 from another computer-readable medium, such as storage device 610. Execution of the sequences of instructions contained in main memory 606 causes processor 604 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions to implement the invention. Thus, embodiments of the invention are not limited to any specific combination of hardware circuitry and software.
  • Non- volatile media includes, for example, optical or magnetic disks, such as storage device 610.
  • Volatile media includes dynamic memory, such as main memory 606.
  • Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 602. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
  • Computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, or any other magnetic medium, a CD- ROM, any other optical medium, punchcards, papertape, any other physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read.
  • Various forms of computer readable media may be involved in carrying one or more sequences of one or more instructions to processor 604 for execution.
  • the instructions may initially be carried on a magnetic disk of a remote computer.
  • the remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem.
  • a modem local to computer system 600 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal.
  • An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 602.
  • Bus 602 carries the data to main memory 606, from which processor 604 retrieves and executes the instructions.
  • Computer system 600 also includes a communication interface 618 coupled to bus 602.
  • Communication interface 618 provides a two-way data communication coupling to a network link 620 that is connected to a local network 622.
  • communication interface 618 maybe an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of telephone line.
  • ISDN integrated services digital network
  • communication interface 618 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN.
  • LAN local area network
  • Wireless links may also be implemented, h any such implementation, communication interface 618 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
  • Network link 620 typically provides data communication through one or more networks to other data devices.
  • network link 620 may provide a connection through local network 622 to a host computer 624 or to data equipment operated by an Internet Service Provider (ISP) 626.
  • ISP 626 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the "Internet" 628.
  • Internet 628 uses electrical, electromagnetic or optical signals that carry digital data streams.
  • the signals through the various networks and the signals on network link 620 and through communication interface 618, which carry the digital data to and from computer system 600, are exemplary forms of carrier waves transporting the information.
  • Computer system 600 can send messages and receive data, including program code, through the network(s), network link 620 and communication interface 618.
  • a server 630 might transmit a requested code for an application program through Internet 628, ISP 626, local network 622 and communication interface 618.
  • the received code may be executed by processor 604 as it is received, and/or stored in storage device 610, or other non- volatile storage for later execution, h this manner, computer system 600 may obtain application code in the form of a carrier wave.

Abstract

Approaches are used for efficiently and effectively managing the dynamic allocation of resources of multi-node database systems between services provided by the multi-node database server. A service is a category of work that is hosted on the database server. The approaches manage allocation of resources at different levels. For services that use a particular database, the performance realized by the services is monitored. Resources assigned to the database are allocated between these services to ensure performance goals for each are met. Resources assigned to a cluster of nodes are allocated between the databases to ensure that performance goals for all the services that use the databases are met. Resources assigned to a farm of clusters are assigned amongst clusters based on service level agreements and back-end policies. The approach uses a hierarchy of directors to manage resources at the different levels.

Description

HffiRARCHICAL MANAGEMENT OF THE DYNAMIC ALLOCATION OF RESOURCES IN A MULTI-NODE SYSTEM
RELATED APPLICATIONS
[0001] The present application claims priority to U.S. Provisional Application No.
60/495,368, Computer Resource Provisioning, filed on August 14, 2003, which is incorporated herein by reference; the present application claims priority to U.S.
Provisional Application No. 60/500,096, Service Based Workload Management and
Measurement in a Distributed System, filed on September 3, 2003, which is incorporated herein by reference; the present application claims priority to U.S. Provisional
Application No. 60/500,050, Automatic And Dynamic Provisioning Of Databases, filed on September 3, 2003, which is incorporated herein by reference.
[0002] The present application is related to the following U.S. Applications:
[0003] U.S. Application No. XX/XXX,XXX, Hierarchical Management of the
Dynamic Allocation of Resources in a Multi-Node System (Attorney Docket No. 50277-
2382), filed by Benny Souder, et al. on August 12, 2004, and incorporated herein by reference;
[0004] U.S. Application No. 1011 \%,141 , Automatic and Dynamic Provisioning of
Databases (Attorney Docket No. 50277-2343), filed on 11/21/2003, which is incorporated herein by reference;
[0005] U.S. Application No. XX/XXX,XXX, Transparent Session Migration Across
Sei-vers (Attorney Docket No. 50277-2383), filed by Sanjay Kaluskar, et al. on August 12,
2004 and incorporated herein by reference;
[0006] U.S. Application No. XX/XXX,XXX, Calculation of Service Performance
Grades in a Multi-Node Environment That Hosts the Services (Attorney Docket No. 50277-2410), filed by Lakshminarayanan Chidambaran, et al. on August 12, 2004 and incorporated herein by reference;
[0007] U. S . Application No . XX/XXX,XXX, Incremental Run-Time Session
Balancing in a Multi-Node System (Attorney Docket No. 50277-2411) filed by
Lakshminarayanan Chidambaran, et al. on August 12, 2004 and incorporated herein by reference;
[0008] U.S. Application No. XX/XXX,XXX, Service Placement for Enforcing
Performance and Availability Levels in a Multi-Node System (Attorney Docket No.
50277-2412), filed by Lakshminarayanan Chidambaran, et al. on August 12, 2004 and incorporated herein by reference;
[0009] U.S. Application No. XX/XXX,XXX, On Demand Node and Server Instance
Allocation and De-Allocation (Attorney Docket No. 50277-2413), filed by
Lakshminarayanan Chidambaran, et al. on August 12, 2004 and incorporated herein by reference;
[0010] U.S. Application No. XX/XXX,XXX, Recoverable Asynchronous Message
Driven Processing in a Multi-Node System (Attorney Docket No. 50277-2414), filed by
Lakshminarayanan Chidambaran, et al. on August 12, 2004 and incorporated herein by reference; and
[0011] U.S. Application No. XX/XXX,XXX, Managing Workload by Service
(Attorney Docket No. 50277-2337), filed by Carol Colrain, et al. on August 12, 2004 and incorporated herein by reference.
[0012] The present application is related to the following International Applications:
[0013] International Application No. PCT/XXXX/XXXXX, entitled Automatic and
Dynamic Provisioning of Databases (Attorney Docket No. 50277-2571), filed on August 9, 2004 by Oracle International Corporation at the United States Receiving Office and which is incorporated herein by reference;
[0014] International Application No. PCT/XXXX/XXXXX, Transparent Session
Migration Across Servers (Attorney Docket No. 50277-2593), filed on August 13, 2004 by Oracle International Corporation at the United States Receiving Office and which is incorporated herein by reference;
[0015] International Application No. PCT/XXXX/XXXXX, Transparent Migration of Stateless Sessions Across Servers (Attorney Docket No. 50277-2594), filed on August
13, 2004 by Oracle International Corporation at the United States Receiving Office and which is incorporated herein by reference; and
[0016] International Application No. PCT/XXXX/XXXXX, On Demand Node and
Server Instance Allocation and De-Allocation (Attorney Docket No. 50277-2595), filed on
August 13, 2004 by Oracle International Corporation at the United States Receiving
Office and which is incorporated herein by reference.
FIELD OF THE INVENTION
[0017] The present invention relates to work load management, and in particular, work load management within a multi-node computer system.
BACKGROUND OF THE INVENTION
[0018] Enterprises are looking at ways of reducing costs and increasing efficiencies of their data processing system. A typical enterprise data processing system allocates individual resources for each of the enterprise's applications. Enough resources are allocated up front for each application to handle the estimated peak load of the application. Each application has different load characteristics; some applications are busy during the day; some others during the night; some reports are run once a week and some others once a month. As a result, there is a lot of resource capacity that is left unutilized. Grid computing enables the utilization or elimination of this unutilized capacity. In fact, Grid computing is poised to drastically change the economics of computing.
[0019] A grid is a collection of computing elements that provide processing and some degree of shared storage; the resources of a grid are allocated dynamically to meet the computational needs and priorities of its clients. Grid computing can dramatically lower the cost of computing, extend the availability of computing resources, and deliver higher productivity and higher quality. The basic idea of Grid computing is the notion of computing as a utility, analogous to the electric power grid or the telephone network. A client of the Grid does not care where its data is or where the computation is performed. All a client wants is to have computation done and have the information delivered to the client when it wants.
[0020] This is analogous to the way electric utilities work; a customer does not know where the generator is, or how the electric grid is wired. The customer just asks for electricity and gets it. The goal is to make computing a utility ~ a ubiquitous commodity. Hence it has the name, the Grid.
[0021] This view of Grid computing as a utility is, of course, a client side view. From the server side, or behind the scenes, the Grid is about resource allocation, information sharing, and high availability. Resource allocation ensures that all those that need or request resources are getting what they need. Resources are not standing idle while requests are left unserviced. Information sharing makes sure that the information clients and applications need is available where and when it is needed. High availability ensures that all the data and computation must always be there — just as a utility company must always provide electric power. Grid Computing for Databases
[0022] One area of computer technology that can benefit -from Grid computing is database technology. A grid can support multiple databases and dynamically allocate and reallocate resources as needed to support the current demand for each database. As the demand for a database increases, more resources are allocated for that database, while other resources are deallocated from another database. For example, on an enterprise grid, a database is being serviced by one database server running on one server blade on the grid. The number of users requesting data from the database increases. In response to this increase, a database server for another database is removed from one server blade and a database server for the database experiencing increased user requests is provisioned to the server blade.
[0023] Grid computing for databases requires allocation and management of resources at different levels. At a level corresponding to a single database, the performance provided to the users of the database must be monitored and resources of the database allocated between the users to ensure performance goals for each of the users are met. Between databases, the allocation of a grid's resources between the databases must be managed to ensure that performance goals for users of all the databases are met. The work to manage allocation of resources at these different levels and the information needed to perform such management is very complex. Therefore, there is a need for a mechanism that simplifies and efficiently handles the management of resources in a Grid computing system for database systems as well as other types of systems that allocate resources at different levels within a Grid.
[0024] Approaches described in this section are approaches that could be pursued, but not necessarily approaches that have been previously conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.
BRIEF DESCRIPTION OF THE DRAWINGS
[0025] The present invention is illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
[0026] FIG. 1 is a block diagram showing a multi-node computer system on which an embodiment of the present invention may be implemented. [0027] FIG. 2 is a block diagram showing a cluster of nodes according to an embodiment of the present invention.
[0028] Fig. 3 is a block diagram showing components of a cluster of nodes and multi- node database server that participate in providing various services for a database according to an embodiment of the present invention.
[0029] FIG. 4 is a flow chart showing a procedure for expanding a service to a target database instance and quiescing a service on the target database instance. [0030] FIG. 5 is a flow chart showing the procedure for two-phase-quiescing. [0031] FIG. 6 is a block diagram of a computer system that may be used in an embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0032] A method and apparatus for managing the allocation of resources in a multi- node environment is described. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
[0033] Described herein are approaches used for efficiently and effectively managing the dynamic allocation of resources of multi-node database systems between services provided by the multi-node database system. A service is work of a particular type or category that is performed for the benefit of one or more clients. The service includes any use or expenditure of computer resources, including, for example, CPU processing time, storing and accessing data in volatile memory, read and writes from and to persistent storage (i.e. disk storage), and use of network or bus bandwidth. A service may be, for example, work that is performed for a particular application on a client of a database server.
[0034] The approaches manage allocation of resources at different levels. For services that use a particular database, the performance realized by the services is monitored. Resources assigned to the database are allocated between these services to ensure performance goals for each are met. Resources assigned to a cluster of nodes are allocated between the databases to ensure that performance goals for all the services that use the databases are met.
[0035] The approach uses a hierarchy of directors to manage resources at the different levels. One type of director, a database director, manages resources allocated to a database among services that use a database and its database instances. The database director manages the allocation of database instances among the services. A cluster director manages resources of a cluster of nodes between databases whose database servers are hosted on the cluster. Yet another director, a farm director, manages resources allocated between clusters.
[0036] FIG. 1 shows a multi-node computer system that may be used to implement an embodiment of the present invention. Referring to Fig. 1, it shows cluster farm 101. A cluster farm is a set of nodes that is organized into groups of nodes, referred to as clusters. Clusters provide some degree of shared storage (e.g. shared access to a set of disk drives) between the nodes in the cluster.
[0037] The nodes in a cluster farm may be in the form of computers (e.g. work stations, personal computers) interconnected via a network. Alternately, the nodes may be the nodes of a grid. A grid is composed of nodes in the form of server blades interconnected with other server blades on a rack. Each server blade is an inclusive computer system, with processor, memory, network connections, and associated electronics on a single motherboard. Typically, server blades do not include onboard storage (other than volatile memory), and they share storage units (e.g. shared disks) along with a power supply, cooling system, and cabling within a rack. [0038] A defining characteristic of a cluster in a cluster farm is that the cluster's nodes may be automatically transferred between clusters within the farm through software control without the need to physically reconnect the node from one cluster to another. A cluster is controlled and managed by software utilities referred to herein as clusterware. Clusterware may be executed to remove a node from a cluster and to provision the node to a cluster. Clusterware provides a command line interface that accepts requests from a human administrator, allowing the administrator to enter commands to provision and remove a node from a cluster. The interfaces may also take the form of Application Program Interfaces ("APIs"), which may be called by other software being executed within the cluster farm. Clusterware uses and maintains metadata that defines the configuration of a cluster within a farm, including cluster configuration metadata, which defines the topology of a cluster in a farm, including which particular nodes are in the cluster. The metadata is modified to reflect changes made to a cluster in a cluster farm by the clusterware. An example of clusterware is software developed by Oracle™, such as Oracle9i Real Application Clusters or Oracle Real Application Clusters lOg. Oracle9i Real Application Clusters is described in Oracle9i RAC: Oracle Real Application Clusters Configuration and Internals, by Mike Ault and Madhu Tumma, 2nd edition (August 2, 2003).
[0039] Cluster farm 101 includes clusters 110, 120, and 130. Each of the clusters hosts one or more multi-node database servers that provide and manage access to databases.
Clusters and Multi-node Database Servers
[0040] FIG. 2 shows a cluster 110 according to an embodiment of the present invention. A defining feature of a cluster is that is it treated as a single unit or entity by clients of the cluster because the clients issue requests for services hosted by the cluster without specifying what particular node or nodes carryout the request, as shall be described in greater detail.
[0041] Cluster 110 includes multi-node database servers 222, 232, and 242. Multi- node database servers 222, 232, and 242 reside on one or more nodes of cluster 110. A server, such as a multi-node server, is a combination of integrated software components and an allocation of computational resources, such as memory, a node, and processes on the node for executing the integrated software components on a processor, where the combination of the software and computational resources are dedicated to performing a particular function on behalf of one or more clients. Among other functions of database management, a multi-node server governs and facilitates access to a particular database, processing requests by clients to access the database. Multi-node servers 222, 232, and 242 govern and provide access to database 220, 230, and 240, respectively. Another example of a server is a web server.
[0042] Resources from multiple nodes in a multi-node computing system can be allocated to running a server's software. Each combination of the software and allocation of the resources from a node is a server that is referred to herein as a "server instance" or "instance". Thus, a multi-node database server comprises multiple server instances that can run on multiple nodes. Several instances of a multi-node database server can in fact run on the same node. A multi-node database server comprises multiple "database instances", each database instance running on a node, and governing and facilitating access to a particular database. Hence, each instance can be referred to herein as a database instance of the particular database. Clusters are often used to host multi-node database servers.
[0043] Clients of cluster 110 as well as multi-node database servers 222, 232, and 242 include clients 203 and 205. Clients 203 and 205 execute applications on computers interconnected to cluster 110 via, for example, a network. An application, as the term is used herein, is a unit of software that is configured to interact with and use the functions of a server. In general, applications are comprised of integrated functions and software modules (e.g. programs comprised of machine executable code or interpretable code, dynamically linked libraries) that perform a set of related functions. [0044] Clients 203, for example, include computer processes executing an FIN application and PAY application. The FIN application includes software that generates and analyzes accounting and financial information of an enterprise. The PAY application generates and tracks information about employee compensation of an enterprise.
[0045] The clients of database servers 222, 232, and 242 are not limited to computers interconnected to cluster 110 via a network. For example, database server 222 may be a client of database server 232.
[0046] A database, such as databases 220, 230, and 240, is a collection of database objects. Database objects include any form of structured data. Structured data is data structured according to a metadata description defining the structure. Structured data includes relational tables, object tables, object-relational tables, and bodies of data structured according to the Extensible Markup Language ("XML"), such as XML documents.
Sessions
[0047] In order for a client to interact with a database server on cluster 110, a session is established for the client. A session, such as a database session, is a particular connection established for a client to a server, such as a database instance, through which the client issues a series of requests (requests for execution of database statements). For each database session established on a database instance, session state data is maintained that reflects the current state of a database session. Such information contains, for example, the identity of the client for which the session was established, and temporary variable values generated by processes executing software within the database session. [0048] A client establishes a database session by transmitting a database connection request to cluster 110.
[0049] A client of cluster 110, such as client 203 and 205, may interact with cluster 110 through client side interface components that reside on the same computer of the client. The client side interface components include API functions that are invoked by an application executed by clients 203 and 205. When a connection to a database server is assigned, connection information identifying the node is received by the client side interface components and used by them to transmit subsequent requests by the client to the node.
[0050] The database session assigned to a client may be migrated to another database instance. Migrating a database session entails creating a new database session on another node. While information about the migration and connection is transmitted to the client side interface components, the information is not accessible to the application and the application is "unaware" of the migration. In this way, the migration is performed transparently to the application. Requests that would be associated with the old database session are performed within the new database session. Techniques for migrating database sessions in this way are described in Transparent Session Migration Across Servers (50277-2383). Services
[0051] As mentioned before, a service is work of a particular type or category that is performed for the benefit of one or more clients. Cluster 110 provides to clients 203 and 205 a database service for accessing database 220, a database service for accessing database 230, and a database service for accessing database 240. In general, a database service is work that is performed by a database server for a client, typically including work to process and/or compute queries that require access to a database. The term query as used herein refers to a database statement that conforms to a database language, such as SQL, and includes database statements that specify operations to add, delete, or modify data and create and modify database objects, such as tables, objects views, and executable routines. [0052] Like any service, a database service may be further divided or categorized into subcategories. Database services for database 220 are further divided into the FIN service and PAY service. The FIN service is the database service performed by database server 222 for the FIN application. Typically, this service involves accessing database objects on database 220 that store database data for FIN applications. The PAY services are database services performed by database server 222 for the PAY application. Typically, this service involves accessing database objects on database 220 that store database data for PAY applications.
[0053] There are various ways in which work by a cluster may be divided or categorized in categories and subcategories of services, the present invention is not limited to any particular way. For example, the work may be divided into services based on users or groups of users (business enterprises, divisions within an enterprise) for which the work is performed. Participants That Provide Services
[0054] Fig. 3 shows components of cluster 110 and multi-node database server 222 that participate in providing various services for database 220. Referring to Fig. 3, multi- node database server 222 includes database instances 322, 332, 342, 352, and 362, which reside on nodes 320, 330, 340, 350, and 360, respectively, and manage access to database 220. Database instances may be provisioned or removed from a particular node using an instance management application. Instance management applications are available, for example, as part of Oracle9i Real Application Clusters or Oracle Real Application Clusters lOg. Instance management applications provide command line interfaces or APIs that may be invoked by administrators or clients to provision or remove a database instance from a node. [0055] Database instances of multi-node database server 222 have been allocated to a particular service. Database instances 322 and 332 have been allocated to service FIN. Database instances 342 and 352 have been allocated to service PAY. Instance 362 has not been allocated to any service. A service is referred to as running or residing on or being hosted by an instance, node, or cluster when the instance, node, or cluster has been allocated to perform the service. Thus, the FIN service is referred to as running or residing on database instances 322 and 332, and on nodes 320 and 330. [0056] Listener 390 is a process running on cluster 110 that receives client database connection requests and directs them to a database instance within cluster 110. The client connection requests received are associated with a service, (e.g. service FIN, PAY) The client request is directed to a database instance hosting the service, where a database session is established for the client. As mentioned previously, the session may be migrated to another database instance. Listener 390 directs the request to the particular database instance and/or node in a way that is transparent to the application. Listener 390 may be running on any node within cluster 110. Once the database session is established for the client, the client may issue additional requests, which may be in the form of function or remote procedure invocations, and which include requests to begin execution of a transaction, to execute queries, to perform updates and other types of transaction operations, to commit or otherwise terminate a transaction, and to terminate a database session.
Monitoring Workload
[0057] Resources are allocated and re-allocated to meet levels of performance and cardinality constraints on the resources. Levels of performance and resource availability established for a particular service are referred to herein as service-level agreements. Levels of performance and cardinality constraints on resources that apply to a multi-node system in general and not necessarily to a particular service are referred to herein as policies. For example, a service-level agreement for service FIN maybe require as a level of performance that the average transaction time for service FIN be no more than a given threshold, and as an availability requirement that at least two instances host service FIN. A policy may require that the CPU utilization of any node should not exceed 80%. [0058] Policies may also be referred to herein as backend policies because they are used by backend administrators to manage overall system performance and to allocate resources between a set of services when it is deemed there are insufficient resources to meet service-level agreements of all the set of services. For example, a policy assigns a higher priority to a database relative to another database. When there are insufficient resources to meet service-level agreements of services of both databases, the database with the higher priority, and the services that use the database, will be favored when allocating resources.
[0059] To meet service-level agreements, a mechanism is needed to monitor and measure workload placed on various resources. These measures of workload are used to determine whether service-level agreements are being met and to adjust the allocation of resources as needed to meet the service-level agreements.
[0060] A workload monitor, such as workload monitor 388, is a distributed set of processes that run on nodes of a cluster to monitor and measure workload of the cluster and generate "performance metrics". Workload monitor 388 runs on cluster 110. Performance metrics is data that indicates the level of performance for one or more resources or services based on performance measures. Approaches for performing these functions are described in Measuring Workload by Sendee (50277-2337). The information generated is accessible by various components within multi-node database server 222 that are responsible for managing the allocation of resources to meet service- level agreements, as shall be described in greater detail later. [0061] A performance metric of a particular type that can be used to gauge a characteristic or condition that indicates a level of performance or workload is referred to herein as a performance measure. A performance measure includes for example, transaction execution time or percent of CPU utilization. In general, service-level agreements that involve levels of performance can be defined by thresholds and criteria that are based on performance measures.
[0062] For example, execution time of a transaction is a performance measure. A service-level agreement based on this measure is that a transaction for service FIN should execute within 300 milliseconds. Yet another performance measure is percentage CPU utilization of a node. A backend policy based on this measure is that a node experience no more than 80% utilization.
[0063] Performance metrics can indicate the performance of a cluster, the performance of a service running on a cluster, a node in the cluster, or a particular database instance. A performance metric or measure particular to a service is referred to herein as a service performance metric or measure. For example, a service performance measure for service FIN is the transaction time for transactions executed for service FIN. HIERARCHICAL RESOURCE ALLOCATION
[0064] Grid computing involves dynamically allocating computer resources to meet service-level agreements. In an embodiment, computer resources are balanced or adjusted at one or more levels of a resource allocation hierarchy. Each level of the hierarchy has a different set of resource pools that are allocated between uses (e.g. services). A resource pool is a group of resources of a particular type, for example, nodes and database instances that are available to a service, nodes that are available to a database or nodes that are available to a cluster. The three levels in the resource allocation are the database level, the cluster level, and the farm level.
Database Level
[0065] At the database level, the resources pools allocated are those currently being used for a particular database, including database instances of the database and nodes that host them. The resource pools of the database level (i.e. resources that can be allocated at the database level) are allocated between services of a database to meet service-level agreements. Generally, this involves placing services on instances and placing sessions on instances.
[0066] Sessions can be placed in several ways. The first way is referred to herein as connection-time balancing. Under connection-time balancing, listener 390 balances workload between service instances by directing database connection requests that require a particular service to an instance of database 220. For example, assume service FIN on database instance 322 is providing better service performance than other instances. Accordingly, listener 390 directs a greater portion of database connection requests requiring the FIN service to database instance 322.
[0067] The second way to place a session is referred to as run-time session balancing. Under run-time session balancing, database sessions are migrated from a database instance to another database instance. The database sessions are migrated using transparent session migration. As mentioned before, techniques for performing this are described in Transparent Session Migration Across Servers.
[0068] Service placement entails expanding and contracting services. Under service expansion and contraction, database instances are allocated to or deallocated from hosting services. When a database instance is assigned to host a service, more database sessions may be created for that service on that instance, thus increasing the number of database sessions associated with and available for the service. For example, to meet service-level agreements for service FIN, instances 322 and 332 are allocated to run service FIN. As demand for the service increases, the service-level agreements are no longer satisfied. When service-level agreements are not met, the service-level agreements are referred to as being violated. In response to this service-level violation, the FIN service is allocated an additional instance, instance 342. An instance can run more than one service. For purposes of illustration, an instance runs only one service. Thus, when FIN is added to instance 342, service PAY is "quiesced" from instance 342, that is, instance 342 is deallocated as a resource that is used for the PAY service and the service on the instance is ceased.
Cluster Level
[0069] At the cluster level, resources that are currently allocated to a cluster are balanced between databases (i.e. database services) to meet service-level agreements. The resource pools balanced at this level are database instances and nodes hosting the database instances. Generally, balancing resources at this level involves provisioning and quiescing instances among existing nodes in the cluster. For example, to meet service- level agreements in response to a service-level violation, instance 362 is provisioned to node 360, a node already in the cluster for multi-node database server 222.
Farm Level
[0070] At this level the resources that can be allocated between services are the nodes in a cluster. The pool of nodes for a cluster is dynamic. For example, in response to a service-level violation, node 370 is added to the cluster of multi-node database server 222. Hierarchy Of Actions to Adjust Resource Allocation
[0071] In general, adjusting the resources at the lower level of the resource allocation hierarchy is less disruptive than at the higher levels. Migrating database sessions and services to and from running instances at the database level is less disruptive and costly than provisioning or quiescing a new database instance at the cluster level. It is less expensive to reshuffle the assigned resources amongst the entities to which the resources are already assigned than to request more resources to be assigned from a higher level. Service placement and session migration at the database level are less expensive than changing the number of database instances for the database at the cluster level. Within the database level, migrating sessions to reshuffle them between instances already hosting a service is less expensive than service expansion and contraction because the latter impacts the service topology and may have a greater overall impact on load distribution. [0072] To remedy a service-level violation, resource allocations are adjusted beginning at the lowest levels of the resource allocation hierarchy, hi this way, the service-level violations are in general remedied in a less disruptive and costly manner. Resort to a higher level of resource allocation is not made if a service-level violation may be solved by adjusting resource allocations at the database level. Some service-level violations require adjustments at some or all levels.
[0073] For example, in response to a service-level violation, FIN is provisioned to instance 352 and PAY is quiesced from the same instance, but only if quiescing service PAY does not violate service-level agreements for PAY. Otherwise, service FIN is expanded to another database instance. However, if there is no available instance to which to provision service FIN, then a new instance is provisioned. This is done by making allocation at the cluster level, provisioning a new instance 362 to node 360, a node already in the cluster. Then, an allocation may be made at the database level by expanding service FIN to instance 362, which at the time the service is provisioned, is an instance already allocated to database 220.
Hierarchy of Directors
[0074] According to an embodiment, a distributed system component, referred to as a distributed director, is responsible for managing workload of resources and resource allocation at each of the levels. As the term is used herein, a system component is a combination of software, data, and one or more processes that execute that software and that use and maintain the data to perform a particular function. A distributed system component is executed on multiple nodes. Preferably, but not necessarily, a director is a system component of a database instance and is operating under the control of the database instance. A distributed director includes directors executed on multiple nodes of cluster farm 101.
[0075] According to an embodiment, a director is responsible for managing the allocation of resources at one or more levels of the resource allocation hierarchy. Specifically, for each database, a director serves as a database director managing the allocation of resources at the database level. Other database instances of the database may have directors that serve as standby database directors, readying to take over as an "active" database director if the current active database director becomes unable to perform this role due to, for example, a system failure.
[0076] For each cluster, a director serves as the cluster director. The cluster director is responsible for managing allocation of resources at the cluster level for a cluster. Other directors within the cluster serve as standby cluster directors.
[0077] Finally, a director serves as the farm director. The farm director is responsible for managing the allocation of resources at the farm level. Other directors within the cluster serve as standby farm directors. [0078] A director receives, maintains, and generates information needed to manage workload at that director's corresponding level.
[0079] Referring to FIG. 3, director 380 is running on database instance 342. Director 380 serves as database director for database 220, cluster director for cluster 110, and farm director for cluster farm 101. Other directors act as a database director for databases 230 and 240, respectively, and as a cluster director for clusters 120 and 130, respectively. [0080] In general, to resolve a service-level violation, directors at all levels of the resource allocation hierarchy may participate to remedy the violation. A database director attempts to remedy a service-level violation by adjusting the allocation of resources at the database level. If the service-level violation requires an adjustment at the next higher level of the resource allocation hierarchy, the cluster level, the database director escalates the resolution of the service-level violation to the cluster director. If the service-level violation requires an adjustment at the highest level of the resource allocation hierarchy, the farm level, the cluster director escalates the resolution of the service-level violation to the farm director. A remedy to a service-level violation may involve an adjustment of resource allocation by all directors.
[0081] Directors communicate to each other using a messaging queue. According to an embodiment, the message queue is a table stored in the database of the database instance hosting the director. The records or rows of the table correspond to a message queue. The record indicates the status of any action taken in response to a message by the director responsible for responding to messages of that type. For example, the database director may request a database instance from the cluster director. The request is added to the queue. The cluster director scans the message queue, detects the request, and acts upon it, updating the record to reflect actions undertaken to respond to the request. The cluster director sends a message to the database director, which is placed in the message queue of the database director.
[0082] The advantage of using a table in database 220 is that it makes available the power and capability of a transaction oriented database system, such as multi-node database server 222, to store the message queue persistently and recoverably. When the active database, cluster, or farm director fails, the standby director stepping in its place can access the message queue in a state consistent with the way the director left it when it failed. Such a message queue and the use of it are described in Recoverable Asynchronous Message Driven Processing in a Multi-Node System (50277-2414). Database Director
[0083] The database director is responsible for monitoring service performance of services to ensure service-level agreements are met, expanding or contracting services from database instances of a database 220, and for migrating database sessions between database instances of database 220. The database director also has access to and stores service performance metrics and service-level agreements for each service using the database. Based on these service performance metrics and service-level agreements, the database director ensures service-level agreements are met in two ways - (1) maintaining service performance compliance with service-level agreements by generating and sending information to the listener that allows the listener to balance workload between service instances; and (2) remedying service-level violations by detecting them and initiating adjustments to resource allocations in order to remedy the service-level violations. [0084] To maintain service performance compliance with service-level agreements, connection-time balancing is used. Specifically, the director 380 generates and provides information, referred to herein as performance grades, to listener 390. Performance grades guide listener 390 in balancing workload to maintain service performance compliance. Performance grades indicate relative service performance of a service on an instance relative to other instances. Based on the performance grades, listener 390 skews the distribution of connection user requests for the service to database instances providing better service performance. Techniques for generating performance grades and distributing user requests in this way based on performance grades are described in Calculation of Service Performance Grades in a Multi-Node Environment That Hosts the Services (50277-2410).
[0085] Remedying a service-level violation requires detecting the service-level violation. Director 380 detects a service-level violation for a service by comparing service performance metrics to service-level agreements. For example, director 380 receives from workload monitor 388 service performance metrics that indicate that the average transaction time for service FIN exceeds 30 milliseconds. A service-level agreement for service FIN requires that average transaction time be no more than 20 milliseconds. By comparing the actual average transaction time to the service-level agreement, director 380 detects a service-level violation.
[0086] When the database director detects a service-level violation for a service, it attempts the least disruptive and costly resource allocation adjustments before attempting the more disruptive and costly resource allocation adjustments, in accordance with the resource allocation hierarchy. To this end, database director first determines whether it may remedy the service-level violation by balancing the workload between instances already hosting a service by migrating database sessions allocated to the service to another database instance where the service performance is better. The number of database sessions migrated is chosen in a way that is targeted to achieve a balanced load between the service instances. [0087] If the database director determines that a service-level violation should not be solved by rebalancing load between existing service instances, director 380 attempts to remedy the service-level violation by expanding a service, i.e. by allocating another existing database instance to host the service, referred to herein as the target database instance. If the target database instance is not hosting a service, the service is expanded by allocating the target database instance to host the service. If the target database instance hosts another service, the database director may quiesce the service if the database director determines that doing so will not cause a service-level violation for the other service.
[0088] FIG. 4 depicts a flowchart of a process followed by director 380, as database director for database 220, to expand a service to an additional database instance ("target database instance") when a service on the target database instance must first be quiesced. For purposes of illustration, service PAY is hosted by database instances 342, 352, and 362, the latter being on node 360. Service FIN is being expanded to database instance 342. The service instance of service PAY running on instance 342 is being quiesced. [0089] Referring to FIG.4, at step 410, director 380 sends a blocking message to listener 390. The blocking message instructs listener 390 to cease distributing user requests for service PAY to target database instance 342.
[0090] At step 420, database director 380 migrates database sessions on a target database instance to the other database instances hosting service FIN. The database sessions are distributed between the other database instances in a way that balances workload between them.
[0091] At step 430, director 380 sends a service activation message to listener 390. The service activation message instructs listener 390 that service FIN is running on instance 362. [0092] At step 440, director 380 uses run-time balancing to balance the workload between instances on which service FIN is running (i.e. instances 322, 332, 342). [0093] In some cases, director 380 may determine that no service can contract to make room for the expansion of another service without violating service-level agreements on a target instance. In this case, director 380 may chose to expand the service to a database instance not yet hosting any service. If no database instance is available, director 380 requests one from the cluster director. In response, the cluster director provisions another database instance as requested, and notifies director 380. Director 380 expands the service to the new database instance.
[0094] Service-level agreements may limit the cardinality of a service. For example, service-level agreements for service FIN require that at least one but no more than three database instances host service FIN, and that service PAY be hosted by at least three database instances.
[0095] A service may be quiesced for reasons other than to make a database instance available for expansion of another service. For example, cardinality constraints for service FUST may vary based on the time of the day. During normal business hours, the cardinality of service FIN may be as high as three but during non-business hours, cardinality may be no higher than one. At the start of non-business hours, three database instances are hosting service FIN. Database director 380 contracts service FIN by quiescing the service on two of the database instances.
[0096] A database director may need to respond to requests made by a cluster director. Such actions include responding to requests by the cluster director to quiesce a database instance, that is, quiesce the services currently hosted by a database instance. This step is needed when the cluster director wishes to replace a database instance for a database with a database instance for another database. Cluster Director
[0097] The cluster director is responsible for provisioning and removing database instances from existing nodes in the cluster. The cluster director also enforces database level policies. Database level policies require, for example, that the cardinality of database instances for a database fall within a minimum and/or maximum, or require that when there are insufficient resources to meet all service-level agreements, that the cluster director skew the allocation of database instances to databases designated as having a higher priority for resource allocation. Skewing between databases in this way also skews the allocation of resources to services using the higher prioritized databases. The cluster director has access to and stores data specifying the priority for resource allocation between databases. Such data may be configured by an administrator of a cluster. [0098] The cluster director provisions and removes database instances in response to a database director's request for a database instance ("NEED-INSTANCE" request). If there is a node within the cluster that is not hosting a database instance (a "free node"), then the cluster director allocates another node to the database by provisioning a database instance to the free node.
[0099] If there is no free node in the cluster, cluster director may request one from the farm director by issuing a "NEED-NODE" request to the farm director. If the farm director is unable to provide one, then the cluster director arbitrates allocation of database instances between databases hosted by the cluster. The arbitration may entail removing a database instance of a database from a node in a cluster and provisioning a database instance for the database for which the NEED-INSTANCE request was generated. [0100] FIG. 5 is a flow chart depicting a process for arbitrating the allocation of database instances between databases. The database director for database 220, has determined that another database instance is needed for a service, and has generated a NEED-INSTANCE request for the cluster director, which is director 380. Director 380, in its capacity as cluster director, determines that a database instance for another database should be removed from a node within cluster 110 so that the node may be used to provision another database instance for database 220.
[0101] Referring to FIG. 5, at step 510, director 380, as cluster director, transmits a "VOLUNTEER-TO-QUIESCE" request to database directors other than the requesting database director (i.e. the director issuing the NEED-INSTANCE request). The purpose of the VOLUNTEER-TO-QUIESCE request is to ask a database director whether it may quiesce a database instance i.e. may reduce the cardinality of database instances for the director's database. A database director of a database may respond by volunteering, transmitting a message indicating that a database instance for the database may be quiesced. A database director may decline to quiesce a database instance i.e. reduce the cardinality of database instances. One reason a database director may send a message declining the request is that all the director's database instances are needed to satisfy availability requirements for a service.
[0102] If database director 380 receives at least one message from a database director affirming that a database instance may be quiesced, that is, more than one database director volunteers, then at step 520, director 380 selects a database from among those who volunteered. A database with a lower resource allocation priority may be selected in favor of a database with a higher resource allocation priority. For purposes of illustration, director 380 selects database 240. The cluster director 384 sends a message to the database director of database 240 to quiesce a database instance of database 240 from a node hosting the database instance.
[0103] If database director 380 receives no message from a database director affirming that a database instance may be quiesced, that is, no database director volunteers, then at step 530, director 380 selects a database with a lower resource allocation priority. Director 380 then transmits a message to the database director of the selected database to quiesce a database instance. Techniques for selecting a database instance to quiesce are described in further detail in On Demand Node and Server Allocation and Deallocation (50277-2413).
[0104] At step 540, the database director of database 240 quiesces a database instance on the node and transmits a notification to the cluster director, director 380, that the database instance is quiesced ("INSTANCE-IDLE" notification). At step 550, director 380 receives the notification.
[0105] At step 560, cluster director 380 removes the database instance from the node and provisions a database instance for database 230 to the node, using, for example, instance provisioning APIs of clusterware.
[0106] As a result of arbitrating the allocation of database instances between databases, the services offered by the database relinquishing the node may experience service-level violations. The cardinality of nodes in the cluster maybe increased, as shall be explained in greater detail, and thus free nodes may become available and may be used to remedy such service-level violations. Director 380, as cluster director, monitors cluster configuration metadata to detect when more free nodes become available, and may allocate them to a database incurring service-level violations. For example, in the current example, the database director for database 220, after relinquishing the database instance, continues to detect service-level violations, and, in response, transmits NEED- INSTANCE requests to its cluster director, director 380. Eventually, after detecting that more free nodes have been added to cluster 110, director 380 is able to respond to one of the requests by provisioning a database instance to the free node. Farm Director
[0107] The farm director is responsible for allocating nodes between clusters in a cluster farm, shuffling nodes between them by removing a node from one cluster and provisioning the node to another cluster. The cluster director also enforces cluster-wide policies. Cluster-wide policies may require, for example, that the cardinality of nodes within a cluster fall within a minimum and/or maximum.
[0108] In response to a NEED-NODE request from a cluster director, a farm director provisions a node to a cluster and removes a free node from another cluster, subject to cluster service-level agreements. A farm director corresponds with the cluster directors in a cluster farm to arbitrate the allocation of a node. This process entails transmitting a "RELINQUISH-NODE" request to the cluster directors, who respond to indicate whether or not they may relinquish a node. Based on the responses, the farm director selects a cluster and interacts with the cluster director of the selected cluster to remove the node from the cluster. Removing the node from the cluster may entail quiescing a database instance. Once the database instance is quiesced, a farm director removes the node from the selected cluster and provisions the node to the cluster needing the node by invoking clusterware using APIs provided for this purpose.
[0109] In addition, a farm director monitors the performance of clusters in the cluster farm using the performance metrics received from workload monitors executing on each of the clusters. If performance metrics indicate that a cluster is performing in violation of service-level agreements for the cluster or is not performing as well as the other cluster, then the farm director shifts one or more nodes from a better performing cluster to the worse performing cluster.
Election of Directors [0110] As mentioned before, within the database instances of a database, there are multiple directors that may serve as the active database director or standby director. Therefore, there is a need for a mechanism to determine which director is the active director. Furthermore, when the active director fails, there is a need to select a standby database director to be the active database director. The same kind of need exists for the cluster directors and farm director. The process of selecting an active director is referred to herein as director election.
[0111] There are various ways in which database directors may be elected. The first involves the use of a database global lock. A database global lock is used to synchronize processes running under control of all the database instances of a database. Upon startup, a database director requests an exclusive lock. If no database director holds the lock, then the database director is granted the lock and assumes the position of database director. Other directors who subsequently request the lock are not granted the lock and assume the position of standby director until the lock is granted, if at all. The requests remain pending until granted or rescinded by the requestor.
[0112] Multi-node database servers detect when the holders of a database global lock experience system failure. In this case, the database global lock of a failed holder is cancelled or released and a pending request for the database global lock is granted to a standby database director. The standby database director whose lock is granted then assumes the role of database director.
[0113] Another technique for director election involves the use of process groups. A process group is a group that may be joined by processes executing on any node within a cluster. Members of the group are informed when another member stops running (e.g. due to system failure) or leaves the process group. In addition, members are assigned an id when joining the group. [0114] When a director starts up, it joins a process group for its database. If upon joining a director has the highest member id, the director assumes the position of database director for the database. When the active director stops running, the members of the process group are informed and the one with the highest member id assumes the role of database director.
[0115] Similar techniques may be used for election of cluster directors. A cluster- wide lock may be used for director election or a process group for a cluster may be used. [0116] Under these techniques, it is possible that the director that assumes the cluster director role may be a director for a database different than that of the previous active cluster director. Consequently, the message queue may reside on a different database. Tables on a database may be accessed more efficiently by processes residing on database instances for the database than processes residing on a database instance of another database.
[0117] A technique that can be used is to ensure that the standby director assuming the role of an active director is hosted by a database instance for the same database is the static data designation technique. Under this technique, a database is designated as the one that hosts (i.e. whose database instances host) the cluster director for a cluster. Only directors for that database play the role of cluster director for a cluster. Director election among these directors may be performed using either a global database lock or a process group for the database.
[0118] The database may be designated by an administrator using an interface provided by clusterware for this purpose. To enhance standby cluster director availability, a database with high priority and high minimum/maximum cardinality requirements can be designated, ensuring a relatively large number of standby directors. The static database designation approach may be used to elect an active farm director. [0119] Organizing the management of resources within a farm cluster using the hierarchy of directors facilitates the generation and exchange of information needed to manage the resources. In general, processes running within a particular database (i.e. processes of a database instance for a particular database) are able to communicate with other processes within the database more efficiently than processes not within the database. Thus, a director that is serving as a database director for a database, and that obtains service performance metrics from one or more work load monitors running within the database, is able to obtain that data more efficiently than directors running within other databases. In addition, since there is only one database director active within a database, the work of getting and generating information for the services running a database, such as service performance metrics, service-level agreements, message queue data, need only be performed by one director.
[0120] Similarly for a cluster director, only one director within the cluster need perform the work of accessing the message queue and getting cluster service-level, agreements and information needed to track which nodes are in the cluster, and what database instances of what database the nodes are hosting.
[0121] The way the information and information exchange is distributed is used to define the actions a director can itself take to manage service performance and which actions the director escalates or delegates to another director. For example, reallocating database instances between databases requires knowing such information as what nodes are available in the cluster, and what nodes have what database instances. Therefore, when a database director, which does not know such information, detects a service-level violation that requires action in the form of reallocation of database instances between databases, the database director escalates the action to the cluster director, which knows such information. Examples of Alternate Embodiments
[0122] An embodiment of the present invention has been illustrated by dynamically allocating the resources of a multi-node system among database services and subcategories of database services. However, the present invention is not so limited. [0123] For example, an embodiment of the present invention may be used to allocate computer resources of a multi-node system that hosts an application server among services provided by the application server. An application server is part of, for example, a three tier architecture in which an application server sits between clients and a database server. The application server is used primarily for storing, providing access to, and executing application code, while a database server is used primarily for storing and providing access to a database for the application server. The application server transmits requests for data to the database server. The requests may be generated by an application server in response to executing the application code stored on the application server. An example of an application server is Oracle 9i Application Server or Oracle lOg Application Server. Similar to examples of a multi-node server described herein, an application server may be distributed as multiple server instances executing on multiple nodes, the server instances hosting multiple sessions that may be migrated between the server instances.
[0124] The present invention is also not limited to homogenous multi-node servers comprised only of server instances that execute copies of the same software product or same version of a software product. For example, a multi-node database server may be comprised of several groups of server instances, each group executing different database server software from a different vendor, or executing a different version of database server software from the same vendor. Transparent Session Migration [0125] Transparent session migrations allows a client to be switched from a session on a server to another session on another server in a way that is a transparent to the application for which the initial session was established. The term migrate refers to an operation in which a client of an existing session on a server is switched from the existing session to another the session, allowing the existing session to be terminated and the client to use the other session in lieu of the existing session. The existing session is referred to herein as having been migrated. The term "transparent" refers to performing an operation in a way that, with respect to a unit of software, does not require execution of instructions in the unit that are tailored to perform the operation. Thus, under transparent session migration, a client is switched between sessions without executing application instructions tailored to accomplish the migration. Instead, a client-side interface component, through which the application interacts with the server, handles details of migration, modifying the internal state of the client-side interface component to effect the same. Legacy applications do not have to be modified in order to institute techniques described herein.
[0126] To migrate a session, the session's state is captured and restored. Capturing a session's state entails producing a stream of bytes, as a veritable copy of a session state of the session, that can be stored in an object, file, or other type of data structure and later accessed to restore the session. Under transparent session migration, the session of a client is captured on a source server, producing a stream of bytes that is stored in a data structure and transported to the destination server, where the destination server restores the session by loading the stream of bytes into a session on the destination server that was established for the client. [0127] Migration of Sessions
[0128] A database session may be "statefull" when a future application call depends on session state generated by prior application calls. Because of the potential dependency of future applications calls upon session state of a statefull session, migration of statefull database sessions entails determining whether portions of session state stored on the source instance may be transferred to another database instance, and if so, transferring copies of the session state between a source database instance and destination instance. In an embodiment, to determine whether portions of session state stored on the source instance may be transferred to another database instance, various migration checks are made to determine whether session state may be transferred to another database instance. These migration checks include detenriining whether a database session is at a transaction boundary, is at a call boundary, or is at a component boundary.
[0129] A database session is at a transaction boundary if there are currently no active transactions being executed for the session. A transaction is a logical unit of work that is performed as an atomic unit. In the context of database systems, the database must reflect all the changes made by a transaction, or none of the changes made by the transaction to ensure the integrity of the database. Consequently, none of the changes made by a transaction are permanently applied to a database until the transaction has been fully executed. A transaction is said to "commit" when the changes made by the transaction are made permanent. A transaction is active if the transaction has not been committed, aborted, or otherwise terminated.
[0130] A session is at a call boundary if a database instance has finished executing a client call rather being at an intermediate stage of processing the call. For example, to process a call to execute a database statement, a database instance goes through stages, each stage corresponding to a particular type of operation. The stages are (1) creating a cursor, (2) parsing the database statement and binding its variables, (3) executing the database statement, (4) fetching the rows to return for the query, and (5) closing the cursor. These stages are described in greater detail in Oracleδ Server Concepts, Release 8.0, Volume 3 (the contents of which incorporated herein by reference), at chapter 23. Intermediate stages are the operations performed before processing of the call is complete. The intermediate stages are stages (1) - (5). After an instance performs step (5) in response to a call, a source session is at a call boundary.
[0131] A session is at a component boundary if each "database component" of a session is at its respective component boundary. A database component is a set of software modules on a database server that provide specialized and related functions. A component session state is specifically generated and used by a database component. According to an embodiment, session state can be viewed as a union or combination of component session states. The following are examples of database components. A cursor component manages cursors within a database instance. A cursor is an area of memory used to store information about a parsed database statement and other information related to the processing of the database statement. A PL/SQL component is a database component responsible for executing code (e.g. procedures) written in PL/SQL, a procedural database language promulgated by Oracle Corporation. Session parameters component is a database component responsible for managing attributes that generally control how calls and requests associated with a session are processed. The attributes are stored in a component session state. For example, session parameters can include an attribute that controls the particular human language for results returned by executing queries. A Java component is a database component that is responsible for executing code (e.g. class and object methods) written in Java. The Java component uses component session state to store information related to the execution of Java code. [0132] A database session is at a component boundary for a particular database component if the component session state of the database component can be migrated to another session. A database component provides a function that returns a value indicating whether or not the database component's respective component session state may be migrated. A reason component session state of a database session may not be migrated is that component session state includes a file descriptor of an open file. The file descriptor contains information that is only valid for an instance hosting the database session. [0133] The function is part of an interface of callback functions supported by each database component. The functions are invoked to capture and restore component session state, as well as to determine whether a component session state permits a session to be migrated.
Generating Performance Metrics
[0134] Performance metrics are data that indicates the quality of service realized by services for one or more resources. A background process generates performance metrics from performance statistics that are generated for each session and service hosted on a database instance. Like performance metrics, performance statistics can indicate a quality of performance. However, performance statistics, in general, include more detailed information about specific uses of specific resources. Performance statistics include, for example, how much time CPU time was used by a session, the velocity of a call, the number of calls a session made, the response time required to complete the calls for a session, how much CPU processing time was used to parse queries for the session, how much CPU processing time was used to execute queries, how many logical and physical reads were performed for the session, and wait times for input and output operations to various resources, such as wait times to read or write to a particular set of data blocks. Performance statistics generated for a session are aggregated by services associated with the session.
[0135] For each session established on a database instance, a session object is created as part of establishing the session on the database instance. A session object contains items of information used by a database instance to manage a session. Among these items of information are a service identifier defined for the session. In an embodiment, the service id is a hash value generated from the service name of the service. [0136] Performance statistics are generated and aggregated by processes performing work requested within a session. For example, a database session is established for a client, who transmits a request to execute a query. A process for the session executes the database server software, parses the query and formulates an execution plan for computing the query in parallel. The database server software that parses the database query, formulates the execution plan and computes the query also generates and aggregates performance statistics. The database server software aggregates the performance statistics by session and service of the session. The performance statistics are stored in a performance statistics repository, which may be, for example, an in- memory fixed table that is associated with a session or that is associated with a service disconnected from any session. Performance statistics for a session are aggregated and stored within an in-memory table for the session. Performance statistics for the service are aggregated and stored within the in-memory table for the service. [0137] For example, a session is associated with the service PAY. To compute a query, a database process assigned to the session uses .4 seconds of CPU time. The process adds .4 to the subtotals in the in-memory table for service PAY. [0138] Periodically, (e.g. every 5 seconds) a workload monitor accesses the performance statistics repository to generate performance metrics and stores the generated performance metrics in a performance metrics repository, which is preferably an in- memory database table. A workload monitor may every five seconds access the performance statistics repository to generate performance metrics, including performance grades. The performance grades are transmitted to a listener for connection-time balancing. The listener uses the performance grades to perform connection-time balancing, establishing sessions for connection requests to services on database instances offering superior performance for the services.
[0139] A workload monitor can also compare performance metrics as they are generated to detect violations of service-level agreements. When a performance violation for a service is detected, the workload monitor sends messages to alert a director (e.g. database director) responsible for responding to the service-level violation. [0140] A service and its various attributes and quality of service (e.g. service-level agreements) need to be defined. According to an embodiment, a database server provides a command line interface that accepts commands from human administrators to create and modify definitions for services of a database server. The definitions are stored as database configuration data in a dictionary in a database as a "service profile".
HARDWARE OVERVIEW [0141] Figure 6 is a block diagram that illustrates a computer system 600 upon which an embodiment of the invention may be implemented. Computer system 600 includes a bus 602 or other communication mechanism for communicating information, and a processor 604 coupled with bus 602 for processing information. Computer system 600 also includes a main memory 606, such as a random access memory (RAM) or other dynamic storage device, coupled to bus 602 for storing information and instructions to be executed by processor 604. Main memory 606 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 604. Computer system 600 further includes a read only memory (ROM) 608 or other static storage device coupled to bus 602 for storing static information and instructions for processor 604. A storage device 610, such as a magnetic disk or optical disk, is provided and coupled to bus 602 for storing information and instructions. [0142] Computer system 600 may be coupled via bus 602 to a display 612, such as a cathode ray tube (CRT), for displaying information to a computer user. An input device 614, including alphanumeric and other keys, is coupled to bus 602 for communicating information and command selections to processor 604. -Another type of user input device is cursor control 616, such as a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 604 and for controlling cursor movement on display 612. This input device typically has two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane.
[0143] The invention is related to the use of computer system 600 for implementing the techniques described herein. According to one embodiment of the invention, those techniques are performed by computer system 600 in response to processor 604 executing one or more sequences of one or more instructions contained in main memory 606. Such instructions may be read into main memory 606 from another computer-readable medium, such as storage device 610. Execution of the sequences of instructions contained in main memory 606 causes processor 604 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions to implement the invention. Thus, embodiments of the invention are not limited to any specific combination of hardware circuitry and software. [0144] The term "computer-readable medium" as used herein refers to any medium that participates in providing instructions to processor 604 for execution. Such a medium may take many forms, including but not limited to, non- volatile media, volatile media, and transmission media. Non- volatile media includes, for example, optical or magnetic disks, such as storage device 610. Volatile media includes dynamic memory, such as main memory 606. Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise bus 602. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
[0145] Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, or any other magnetic medium, a CD- ROM, any other optical medium, punchcards, papertape, any other physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read.
[0146] Various forms of computer readable media may be involved in carrying one or more sequences of one or more instructions to processor 604 for execution. For example, the instructions may initially be carried on a magnetic disk of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a telephone line using a modem. A modem local to computer system 600 can receive the data on the telephone line and use an infra-red transmitter to convert the data to an infra-red signal. An infra-red detector can receive the data carried in the infra-red signal and appropriate circuitry can place the data on bus 602. Bus 602 carries the data to main memory 606, from which processor 604 retrieves and executes the instructions. The instructions received by main memory 606 may optionally be stored on storage device 610 either before or after execution by processor 604. [0147] Computer system 600 also includes a communication interface 618 coupled to bus 602. Communication interface 618 provides a two-way data communication coupling to a network link 620 that is connected to a local network 622. For example, communication interface 618 maybe an integrated services digital network (ISDN) card or a modem to provide a data communication connection to a corresponding type of telephone line. As another example, communication interface 618 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN. Wireless links may also be implemented, h any such implementation, communication interface 618 sends and receives electrical, electromagnetic or optical signals that carry digital data streams representing various types of information.
[0148] Network link 620 typically provides data communication through one or more networks to other data devices. For example, network link 620 may provide a connection through local network 622 to a host computer 624 or to data equipment operated by an Internet Service Provider (ISP) 626. ISP 626 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the "Internet" 628. Local network 622 and Internet 628 both use electrical, electromagnetic or optical signals that carry digital data streams. The signals through the various networks and the signals on network link 620 and through communication interface 618, which carry the digital data to and from computer system 600, are exemplary forms of carrier waves transporting the information. [0149] Computer system 600 can send messages and receive data, including program code, through the network(s), network link 620 and communication interface 618. In the Internet example, a server 630 might transmit a requested code for an application program through Internet 628, ISP 626, local network 622 and communication interface 618. [0150] The received code may be executed by processor 604 as it is received, and/or stored in storage device 610, or other non- volatile storage for later execution, h this manner, computer system 600 may obtain application code in the form of a carrier wave. [0151] In the foregoing specification, embodiments of the invention have been described with reference to numerous specific details that may vary from implementation to implementation. Thus, the sole and exclusive indicator of what is the invention, and is intended by the applicants to be the invention, is the set of claims that issue from this application, in the specific form in which such claims issue, including any subsequent correction. Any definitions expressly set forth herein for terms contained in such claims shall govern the meaning of such terms as used in the claims. Hence, no limitation, element, property, feature, advantage or attribute that is not expressly recited in a claim should limit the scope of such claim in any way. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims

CLAIMSWhat is claimed is:
1. A method for dynamically allocating computer resources of a multi-node computer system, the method comprising computer implemented steps of: monitoring performance realized by a plurality of services running on the multi-node computer system, wherein said plurality of services includes a first service and a second service; based on said monitoring the performance of a plurality of services, generating performance metrics that indicate performance realized by each service of said plurality of services; based on the performance metrics, said multi-node computer system detecting a violation of service-level agreements for said first service; and in response to detecting said violation of said service-level agreements, said multi-node computer system adjusting allocation of computer resources of said multi-node system between said first service and said second service.
2. The method of claim 1, wherein the step of adjusting allocation of resources includes allocating another node of said multi-node system to host said first service.
3. The method of claim 1, wherein a first server runs on a first node of said multi-node system, wherein the step of adjusting allocation of resources includes expanding said first service to said first server.
4. The method of claim 3, wherein adjusting allocation of resources includes provisioning said first server to said first node to perform the step of expanding said first service to said first server.
5. The method of claim 1 , wherein: said multi-node system includes a first node and a second node; a first group of sessions are established for said first service on said first node; a second group of sessions are established for said first service on said second node; and the step of allocating resources includes migrating at least one session from said second group to said first node.
6. The method of claim 1 , wherein: said computer resources contain pools of resources; the step of adjusting allocation of computer resources includes attempting to resolve said performance violation by adjusting, for said first service, allocation of a first pool of resources that is lower in a hierarchy before attempting to adjust allocation of a second pool of resources that is higher in said hierarchy; and the pools of resources include a first pool of resources and a second pool of resources.
7. The method of claim 6, wherein: the pools of resources include a first pool of servers as said first pool of resources and a second pool of nodes as said second pool of resources; and the step of attempting to resolve said performance violation includes attempting to allocate another server from said first pool to host said first service before attempting to add another node from said second pool to host said first service.
8. The method of claim 7, wherein the steps further include, in response to adding said another node, provisioning another server to said another node.
9. The method of claim 6, wherein: the first multi-node server includes a first server, a second server, and a third server; the first server hosts a plurality of sessions for said first service; and the step of attempting to resolve said performance violation includes attempting to migrate at least one of said plurality of sessions to said second server before allocating said third server to host said first service.
10. The method of claim 1 , wherein: each service of said plurality of services is a category of work performed by said multi-node system; and the step of adjusting allocation of computer resources of said multi-node system includes adjusting allocation of computer resources to each category of said plurality of category of work.
11. The method of claim 10, wherein: a plurality of units of software execute on a set of client computers interconnected to said multi-node system; and each category of work corresponds to a particular unit of software from said plurality of units.
12. The method of claim 11, wherein each category of work conesponds to work performed in response to a request generated by a certain particular unit of software from said plurality of units.
13. A method for dynamically allocating computer resources of a multi-node computer system that includes a first set of nodes and a second set of nodes, the method comprising the steps of: monitoring performance of a plurality of services hosted on said multi-node system to generate performance metrics; based on said monitoring the performance of a plurality of services, generating performance metrics that indicate performance realized by each service of said plurality of services; wherein said multi-node system includes a first set of nodes and a second set of nodes; running a first multi-node server and a second multi-node server on said first set of nodes; wherein said plurality of services includes a first service and a second service hosted by said first multi-node server; a first system component running on said first set of nodes adjusting an allocation of resources of the first multi-node server between said first service and said second service based on said performance metrics; and a second system component running on said first set of nodes adjusting an allocation of resources of the first set of nodes between said first multi- node server and said second multi-node server based on said performance metrics.
14. The method of claim 13, wherein the steps further include: the first system component detecting a service-level violation for said first service; and in response to detecting the service-level violation for said first service, said first system component transmitting a first request to said second system component to adjust the allocation of resources of the first set of nodes between said first multi-node server and said second multi- node server.
15. The method of claim 13, wherein said second system component adjusting an allocation of resources of the first set of nodes between said first multi-node server and said second multi-node server includes said second system component allocating nodes between said first multi-node server and said second multi-node server.
16. The method of claim 13, the steps further include: the first system component detecting a service-level violation for said first service; and in response to detecting the service-level violation for said first service, the first system component adjusting allocation of resources of the first multi-node server between said first service and said second service.
17. The method of claim 16, wherein: the first multi-node server includes a first server instance and a second server instance; and wherein the step of adjusting allocation of resources of the first multi-node server includes causing said second server to host said first service.
18. The method of claim 16, wherein: the first multi-node server includes a first server instance and a second server instance; the first server instance hosts at least one session for said first service; the second server instance hosts at least one session for said first service; and the step of adjusting allocation of resources of the first multi-node server includes adjusting the balance of sessions for said first service between said first server instance and said second server instance.
19. The method of claim 14, wherein in response to receiving said first request, said second system component provisioning another server instance of the first multi-node server to another node.
20. The method of claim 19, wherein: the first set of nodes is a first cluster of nodes that share access to at least one persistent storage device; the second system component transmits a second request to a third system component running on the multi-node system to request another node; and the third system component allocates nodes between the first cluster and another cluster within the multi-node system based on said performance metrics.
21. The method of claim 19, wherein provisioning another server instance causes said second system component to transmit a message to said first system component to indicate said another server instance is available.
PCT/US2004/026389 2003-08-14 2004-08-13 Hierarchical management of the dynamic allocation of resourses in a multi-node system WO2005017783A2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2006523412A JP4970939B2 (en) 2003-08-14 2004-08-13 Hierarchical management of dynamic resource allocation in multi-node systems
CA002533744A CA2533744C (en) 2003-08-14 2004-08-13 Hierarchical management of the dynamic allocation of resources in a multi-node system
AU2004266017A AU2004266017B2 (en) 2003-08-14 2004-08-13 Hierarchical management of the dynamic allocation of resources in a multi-node system
EP04781129.4A EP1654648B1 (en) 2003-08-14 2004-08-13 Hierarchical management of the dynamic allocation of resources in a multi-node system

Applications Claiming Priority (24)

Application Number Priority Date Filing Date Title
US49536803P 2003-08-14 2003-08-14
US60/495,368 2003-08-14
US50005003P 2003-09-03 2003-09-03
US50009603P 2003-09-03 2003-09-03
US60/500,050 2003-09-03
US60/500,096 2003-09-03
US10/718,747 US7873684B2 (en) 2003-08-14 2003-11-21 Automatic and dynamic provisioning of databases
US10/718,747 2003-11-21
US10/918,054 2004-08-12
US10/917,661 2004-08-12
US10/918,054 US8365193B2 (en) 2003-08-14 2004-08-12 Recoverable asynchronous message driven processing in a multi-node system
US10/917,953 US7552218B2 (en) 2003-08-14 2004-08-12 Transparent session migration across servers
US10/918,055 2004-08-12
US10/917,661 US7437459B2 (en) 2003-08-14 2004-08-12 Calculation of service performance grades in a multi-node environment that hosts the services
US10/918,056 US7437460B2 (en) 2003-08-14 2004-08-12 Service placement for enforcing performance and availability levels in a multi-node system
US10/917,873 2004-08-12
US10/918,056 2004-08-12
US10/917,873 US7516221B2 (en) 2003-08-14 2004-08-12 Hierarchical management of the dynamic allocation of resources in a multi-node system
US10/917,687 US7441033B2 (en) 2003-08-14 2004-08-12 On demand node and server instance allocation and de-allocation
US10/918,055 US7552171B2 (en) 2003-08-14 2004-08-12 Incremental run-time session balancing in a multi-node system
US10/917,715 US7664847B2 (en) 2003-08-14 2004-08-12 Managing workload by service
US10/917,953 2004-08-12
US10/917,715 2004-08-12
US10/917,687 2004-08-12

Publications (2)

Publication Number Publication Date
WO2005017783A2 true WO2005017783A2 (en) 2005-02-24
WO2005017783A3 WO2005017783A3 (en) 2005-09-29

Family

ID=34199514

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/US2004/026389 WO2005017783A2 (en) 2003-08-14 2004-08-13 Hierarchical management of the dynamic allocation of resourses in a multi-node system
PCT/US2004/026405 WO2005017745A2 (en) 2003-08-14 2004-08-13 On demand node and server instance allocation and de-allocation

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/US2004/026405 WO2005017745A2 (en) 2003-08-14 2004-08-13 On demand node and server instance allocation and de-allocation

Country Status (5)

Country Link
EP (2) EP1654648B1 (en)
JP (2) JP4805150B2 (en)
AU (2) AU2004266019B2 (en)
CA (2) CA2534807C (en)
WO (2) WO2005017783A2 (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006131437A2 (en) * 2005-06-09 2006-12-14 International Business Machines Corporation Adjusting configuration parameters for a server when a different server fails
EP1872249A2 (en) * 2005-04-07 2008-01-02 Cluster Resources, Inc. On-demand access to compute resources
EP1974270A2 (en) * 2006-01-06 2008-10-01 International Business Machines Corporation Method for autonomic system management using adaptive allocation of resources
US7698430B2 (en) 2005-03-16 2010-04-13 Adaptive Computing Enterprises, Inc. On-demand compute environment
US8151103B2 (en) 2004-03-13 2012-04-03 Adaptive Computing Enterprises, Inc. System and method for providing object triggers
US8176490B1 (en) 2004-08-20 2012-05-08 Adaptive Computing Enterprises, Inc. System and method of interfacing a workload manager and scheduler with an identity manager
CN102521056A (en) * 2011-12-28 2012-06-27 用友软件股份有限公司 Task allocation device and task allocation method
EP2775400A1 (en) * 2013-03-06 2014-09-10 Fujitsu Limited Ressource management system and method
US10133330B2 (en) 2013-07-17 2018-11-20 Fujitsu Limited Cluster system, controller, method for controlling, and computer-readable recording medium having stored therein controlling program that operate node at the combination of the respective load setting values that satisfy required performance and lowers power consumption
US10608949B2 (en) 2005-03-16 2020-03-31 Iii Holdings 12, Llc Simple integration of an on-demand compute environment
US10977090B2 (en) 2006-03-16 2021-04-13 Iii Holdings 12, Llc System and method for managing a hybrid compute environment
US11467883B2 (en) 2004-03-13 2022-10-11 Iii Holdings 12, Llc Co-allocating a reservation spanning different compute resources types
US11494235B2 (en) 2004-11-08 2022-11-08 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11522952B2 (en) 2007-09-24 2022-12-06 The Research Foundation For The State University Of New York Automatic clustering for self-organizing grids
US11526304B2 (en) 2009-10-30 2022-12-13 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US11652706B2 (en) 2004-06-18 2023-05-16 Iii Holdings 12, Llc System and method for providing dynamic provisioning within a compute environment
US11720290B2 (en) 2009-10-30 2023-08-08 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7272609B1 (en) * 2004-01-12 2007-09-18 Hyperion Solutions Corporation In a distributed hierarchical cache, using a dependency to determine if a version of the first member stored in a database matches the version of the first member returned
US9225663B2 (en) 2005-03-16 2015-12-29 Adaptive Computing Enterprises, Inc. System and method providing a virtual private cluster
US9015324B2 (en) 2005-03-16 2015-04-21 Adaptive Computing Enterprises, Inc. System and method of brokering cloud computing resources
US8543686B2 (en) * 2009-07-23 2013-09-24 University-Industry Cooperation Group Of Kyung Hee University Dynamic resource collaboration between network service providers
US9733983B2 (en) * 2011-09-27 2017-08-15 Oracle International Corporation System and method for surge protection and rate acceleration in a traffic director environment
CN107229628B (en) * 2016-03-23 2022-02-01 金篆信科有限责任公司 Distributed database preprocessing method and device
CN107040582B (en) * 2017-02-17 2020-08-14 创新先进技术有限公司 Data processing method and device
CN112395045A (en) * 2020-11-13 2021-02-23 深圳力维智联技术有限公司 Virtual machine recovery and resource adjustment method thereof

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6088728A (en) 1997-06-11 2000-07-11 Oracle Corporation System using session data stored in session data storage for associating and disassociating user identifiers for switching client sessions in a server
WO2003062983A2 (en) 2002-01-16 2003-07-31 Sun Microsystems, Inc. Method, system, and program for determining a modification of a system resource configuration

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3241214B2 (en) * 1994-09-07 2001-12-25 株式会社東芝 Distributed processing apparatus and process execution method
JPH0962624A (en) * 1995-08-28 1997-03-07 Hitachi Ltd Processing method and processing system for on-line transaction
JP3754481B2 (en) * 1996-02-02 2006-03-15 富士通株式会社 Compound computer system
JPH09231143A (en) * 1996-02-23 1997-09-05 Sony Corp Communication control method
JPH09269903A (en) * 1996-04-02 1997-10-14 Hitachi Ltd Process managing system
US6230183B1 (en) * 1998-03-11 2001-05-08 International Business Machines Corporation Method and apparatus for controlling the number of servers in a multisystem cluster
JP4292693B2 (en) * 2000-07-07 2009-07-08 株式会社日立製作所 Computer resource dividing apparatus and resource dividing method
WO2002007037A1 (en) 2000-07-17 2002-01-24 Galactic Computing Corporation Method and system for providing dynamic hosted service management
US6880156B1 (en) * 2000-07-27 2005-04-12 Hewlett-Packard Development Company. L.P. Demand responsive method and apparatus to automatically activate spare servers
JP2002150000A (en) * 2000-11-13 2002-05-24 Dainippon Screen Mfg Co Ltd Server system
US7734676B2 (en) * 2001-06-27 2010-06-08 International Business Machines Corporation Method for controlling the number of servers in a hierarchical resource environment
US7174379B2 (en) * 2001-08-03 2007-02-06 International Business Machines Corporation Managing server resources for hosted applications
US7266099B2 (en) * 2002-01-23 2007-09-04 Hewlett-Packard Development Company, L.P. Method for hand-off of a data session

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6088728A (en) 1997-06-11 2000-07-11 Oracle Corporation System using session data stored in session data storage for associating and disassociating user identifiers for switching client sessions in a server
WO2003062983A2 (en) 2002-01-16 2003-07-31 Sun Microsystems, Inc. Method, system, and program for determining a modification of a system resource configuration

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JEFFREY S. CHASE ET AL.: "Dynamic virtual clusters in a grid site manager", PROCEEDINGS OF THE 12TH IEEE SYMPOSIUM ON HIGH PERFORMANCE DISTRIBUTED COMPUTING, 6000320, pages 90 - 100

Cited By (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11467883B2 (en) 2004-03-13 2022-10-11 Iii Holdings 12, Llc Co-allocating a reservation spanning different compute resources types
USRE45593E1 (en) 2004-03-13 2015-06-30 Adaptive Computing Enterprises, Inc. System and method for providing object triggers
US8151103B2 (en) 2004-03-13 2012-04-03 Adaptive Computing Enterprises, Inc. System and method for providing object triggers
US11652706B2 (en) 2004-06-18 2023-05-16 Iii Holdings 12, Llc System and method for providing dynamic provisioning within a compute environment
US9038078B2 (en) 2004-08-20 2015-05-19 Adaptive Computing Enterprises, Inc. System and method for a workload management and scheduling module to manage access to a compute environment according to local and non-local user identity information
US8806492B2 (en) 2004-08-20 2014-08-12 Adaptive Computing Enterprises, Inc. System and method of interfacing a workload manager and scheduler with an identity manager
US11630704B2 (en) 2004-08-20 2023-04-18 Iii Holdings 12, Llc System and method for a workload management and scheduling module to manage access to a compute environment according to local and non-local user identity information
US8176490B1 (en) 2004-08-20 2012-05-08 Adaptive Computing Enterprises, Inc. System and method of interfacing a workload manager and scheduler with an identity manager
US11494235B2 (en) 2004-11-08 2022-11-08 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11537434B2 (en) 2004-11-08 2022-12-27 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11886915B2 (en) 2004-11-08 2024-01-30 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11656907B2 (en) 2004-11-08 2023-05-23 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11709709B2 (en) 2004-11-08 2023-07-25 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11537435B2 (en) 2004-11-08 2022-12-27 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11861404B2 (en) 2004-11-08 2024-01-02 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US11762694B2 (en) 2004-11-08 2023-09-19 Iii Holdings 12, Llc System and method of providing system jobs within a compute environment
US8930536B2 (en) 2005-03-16 2015-01-06 Adaptive Computing Enterprises, Inc. Virtual private cluster
US7698430B2 (en) 2005-03-16 2010-04-13 Adaptive Computing Enterprises, Inc. On-demand compute environment
US10608949B2 (en) 2005-03-16 2020-03-31 Iii Holdings 12, Llc Simple integration of an on-demand compute environment
US11658916B2 (en) 2005-03-16 2023-05-23 Iii Holdings 12, Llc Simple integration of an on-demand compute environment
US9961013B2 (en) 2005-03-16 2018-05-01 Iii Holdings 12, Llc Simple integration of on-demand compute environment
US9979672B2 (en) 2005-03-16 2018-05-22 Iii Holdings 12, Llc System and method providing a virtual private cluster
US11356385B2 (en) 2005-03-16 2022-06-07 Iii Holdings 12, Llc On-demand compute environment
US11134022B2 (en) 2005-03-16 2021-09-28 Iii Holdings 12, Llc Simple integration of an on-demand compute environment
US10333862B2 (en) 2005-03-16 2019-06-25 Iii Holdings 12, Llc Reserving resources in an on-demand compute environment
US11533274B2 (en) 2005-04-07 2022-12-20 Iii Holdings 12, Llc On-demand access to compute resources
US9075657B2 (en) 2005-04-07 2015-07-07 Adaptive Computing Enterprises, Inc. On-demand access to compute resources
US10986037B2 (en) 2005-04-07 2021-04-20 Iii Holdings 12, Llc On-demand access to compute resources
US10277531B2 (en) 2005-04-07 2019-04-30 Iii Holdings 2, Llc On-demand access to compute resources
EP1872249A2 (en) * 2005-04-07 2008-01-02 Cluster Resources, Inc. On-demand access to compute resources
US11831564B2 (en) 2005-04-07 2023-11-28 Iii Holdings 12, Llc On-demand access to compute resources
US11496415B2 (en) 2005-04-07 2022-11-08 Iii Holdings 12, Llc On-demand access to compute resources
EP3203374A1 (en) * 2005-04-07 2017-08-09 III Holdings 12, LLC On-demand access to compute resources
US11522811B2 (en) 2005-04-07 2022-12-06 Iii Holdings 12, Llc On-demand access to compute resources
US11765101B2 (en) 2005-04-07 2023-09-19 Iii Holdings 12, Llc On-demand access to compute resources
EP1872249A4 (en) * 2005-04-07 2009-07-29 Cluster Resources Inc On-demand access to compute resources
US7464303B2 (en) 2005-06-09 2008-12-09 International Business Machines Corporation Autonomically adjusting configuration parameters for a server when a different server fails
WO2006131437A2 (en) * 2005-06-09 2006-12-14 International Business Machines Corporation Adjusting configuration parameters for a server when a different server fails
WO2006131437A3 (en) * 2005-06-09 2007-08-23 Ibm Adjusting configuration parameters for a server when a different server fails
EP1974270A2 (en) * 2006-01-06 2008-10-01 International Business Machines Corporation Method for autonomic system management using adaptive allocation of resources
US7719983B2 (en) 2006-01-06 2010-05-18 International Business Machines Corporation Method for autonomic system management using adaptive allocation of resources
EP1974270A4 (en) * 2006-01-06 2010-01-06 Ibm Method for autonomic system management using adaptive allocation of resources
US11650857B2 (en) 2006-03-16 2023-05-16 Iii Holdings 12, Llc System and method for managing a hybrid computer environment
US10977090B2 (en) 2006-03-16 2021-04-13 Iii Holdings 12, Llc System and method for managing a hybrid compute environment
US11522952B2 (en) 2007-09-24 2022-12-06 The Research Foundation For The State University Of New York Automatic clustering for self-organizing grids
US11526304B2 (en) 2009-10-30 2022-12-13 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US11720290B2 (en) 2009-10-30 2023-08-08 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
CN102521056A (en) * 2011-12-28 2012-06-27 用友软件股份有限公司 Task allocation device and task allocation method
US9742687B2 (en) 2013-03-06 2017-08-22 Fujitsu Limited Management system and method for execution of virtual machines
EP2775400A1 (en) * 2013-03-06 2014-09-10 Fujitsu Limited Ressource management system and method
US10133330B2 (en) 2013-07-17 2018-11-20 Fujitsu Limited Cluster system, controller, method for controlling, and computer-readable recording medium having stored therein controlling program that operate node at the combination of the respective load setting values that satisfy required performance and lowers power consumption

Also Published As

Publication number Publication date
EP1654648A2 (en) 2006-05-10
AU2004266019A2 (en) 2005-02-24
CA2534807C (en) 2010-12-14
EP1654648B1 (en) 2015-11-11
EP1654649A2 (en) 2006-05-10
WO2005017783A3 (en) 2005-09-29
CA2534807A1 (en) 2005-02-24
JP2007502468A (en) 2007-02-08
JP2007506157A (en) 2007-03-15
CA2533744C (en) 2010-01-12
AU2004266019A1 (en) 2005-02-24
WO2005017745A2 (en) 2005-02-24
CA2533744A1 (en) 2005-02-24
JP4970939B2 (en) 2012-07-11
AU2004266017A1 (en) 2005-02-24
EP1654649B1 (en) 2015-12-23
AU2004266019B2 (en) 2009-11-05
JP4805150B2 (en) 2011-11-02
AU2004266017B2 (en) 2009-12-03
WO2005017745A3 (en) 2005-10-13

Similar Documents

Publication Publication Date Title
US7516221B2 (en) Hierarchical management of the dynamic allocation of resources in a multi-node system
AU2004266017B2 (en) Hierarchical management of the dynamic allocation of resources in a multi-node system
US7930344B2 (en) Incremental run-time session balancing in a multi-node system
US7441033B2 (en) On demand node and server instance allocation and de-allocation
US7437460B2 (en) Service placement for enforcing performance and availability levels in a multi-node system
CN100527090C (en) Method for dynamically distributing computer resource
US8365193B2 (en) Recoverable asynchronous message driven processing in a multi-node system
US7664847B2 (en) Managing workload by service
US8626890B2 (en) Connection pool use of runtime load balancing service performance advisories
US9460185B2 (en) Storage device selection for database partition replicas
US7953860B2 (en) Fast reorganization of connections in response to an event in a clustered computing system
JP6190389B2 (en) Method and system for performing computations in a distributed computing environment
US7502824B2 (en) Database shutdown with session migration
JP2015506523A (en) Dynamic load balancing in a scalable environment
JP2007503628A (en) Fast application notification in clustered computing systems
US7437459B2 (en) Calculation of service performance grades in a multi-node environment that hosts the services
WO2021057108A1 (en) Data reading method, data writing method, and server
Arpaci-Dusseau et al. An Information-Based Approach to Distributed Systems Design

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200480023057.8

Country of ref document: CN

AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DPEN Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101)
ENP Entry into the national phase

Ref document number: 2533744

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 2004781129

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2006523412

Country of ref document: JP

Ref document number: 529/CHENP/2006

Country of ref document: IN

WWE Wipo information: entry into national phase

Ref document number: 2004266017

Country of ref document: AU

ENP Entry into the national phase

Ref document number: 2004266017

Country of ref document: AU

Date of ref document: 20040813

Kind code of ref document: A

WWP Wipo information: published in national office

Ref document number: 2004266017

Country of ref document: AU

WWP Wipo information: published in national office

Ref document number: 2004781129

Country of ref document: EP