US10057139B2 - Maintain a service on a cloud network based on a scale rule - Google Patents

Maintain a service on a cloud network based on a scale rule Download PDF

Info

Publication number
US10057139B2
US10057139B2 US14/909,584 US201314909584A US10057139B2 US 10057139 B2 US10057139 B2 US 10057139B2 US 201314909584 A US201314909584 A US 201314909584A US 10057139 B2 US10057139 B2 US 10057139B2
Authority
US
United States
Prior art keywords
service
cloud
status data
scale
condition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US14/909,584
Other versions
US20160191343A1 (en
Inventor
Jing Dong
John M Green
Phyllis Gallagher
Aswini Kumar
Anupriya Ramraj
Burton Akira Hipp
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hewlett Packard Enterprise Development LP
Original Assignee
Hewlett Packard Enterprise Development LP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hewlett Packard Enterprise Development LP filed Critical Hewlett Packard Enterprise Development LP
Assigned to HEWLETT PACKARD ENTERPRISE DEVELOPMENT LP reassignment HEWLETT PACKARD ENTERPRISE DEVELOPMENT LP ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P.
Publication of US20160191343A1 publication Critical patent/US20160191343A1/en
Assigned to HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. reassignment HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GALLAGHER, Phyllis, RAMRAJ, ANUPRIYA, DONG, JING, GREEN, JOHN M, HIPP, BURTON AKIRA, KUMAR, Aswini
Application granted granted Critical
Publication of US10057139B2 publication Critical patent/US10057139B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/5003Managing SLA; Interaction between SLA and QoS
    • H04L41/5019Ensuring fulfilment of SLA
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/70Admission control; Resource allocation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]

Definitions

  • Cloud computing commonly utilizes a shared pool of computing resources.
  • the computing environment including the shared pool of resources, may be referred to as a “cloud.”
  • the resources may be virtual instances of resources running on a virtual machine.
  • the pool of computing resource may provide a service. Cloud-based services are said to execute or reside “on the cloud.”
  • the computing environment, or cloud may adapt to an increase in service usage by providing additional resource instances to the pool.
  • the shared pool of resources may reside internal to a corporate network or external as part of the public Internet, respectively known as the private cloud and the public cloud.
  • a hybrid cloud environment may utilize both a private cloud and a public cloud.
  • a service provider may provide a private cloud for non-peak service usage and the service provider may offload service requests to the public cloud during peak times, or times of high request rates.
  • Cloud resource providers may provide resources, such as a central processing unit (“CPU”) or computer memory, for services based on the type of cloud computing environment, which may include virtual machines.
  • Cloud resource providers may offer variations on resource provisioning for services, including infrastructure as a service, platform as a service, storage as a service, and software as a service. Cloud resource providers may rent space on the cloud to service providers, or customers, for the service based on a cost per time period, cost per resource, subscription, or any other appropriate cost method.
  • FIGS. 1 and 2 are block diagrams depicting example methods for maintaining a service on a cloud network.
  • FIG. 3 depicts example operations for maintaining a service on a cloud network.
  • FIGS. 4 and 5 are block diagrams depicting example systems for maintaining a service on a cloud network.
  • FIG. 6 depicts example environments in which various examples for maintaining a service on a cloud network may be implemented.
  • a service may be any appropriate supplying of communication, transmissions, information, software, storage, or any other product or activity that may be capable of running on the cloud, or a cloud-based service.
  • a cloud may be any appropriate combination of physical and virtual resources pooled for computing and/or storage purposes.
  • a cloud may include any appropriate number of individual resources, servers, and server groups including virtual instances of resources, servers, and server groups.
  • a cloud network may include any appropriate number of clouds and/or other network of resources accessible by a network device.
  • Cloud resource providers may endeavor to maintain a quality of service for a customer based on resource utilization, or the amount and/or degree of resources utilized by the service. For example, the amount of resources assigned to a service may be highly utilized or underutilized based on any given time of the day. Cloud service providers may maintain the service by scaling the cloud based on the amount of resources utilized by the service. Scaling may refer to increasing or decreasing the shared pool of resources. Cloud resource providers may offer the ability to dynamically scale the cloud to adapt, to service spikes and lulls based on resource utilization. A service spike may occur when there is an increase in service requests or requests associated with providing the service. Accordingly, scaling based on resource utilization may enable an increase in assigning resource for utilizations that may be indirectly associated to the service. For example, the cloud resource provider may add cloud resources every day for a daily virus scan. The customer may be paying for an increase of cloud resources at times of unnecessary or unwanted scaling and may not know until arrival of the bill from the cloud resource provider.
  • a cloud may be scaled while accounting for cost in the scale rule.
  • QoS quality of service
  • workload volume the cloud may be scaled while accounting for cost in the scale rule.
  • a cloud may be scaled based on resource utilization, QoS, the amount of service requests and a budget to reduce insufficient or excessive allocation of cloud resources to a service.
  • the scale rule may account for and/or be adaptable to the desires and/or needs of the customer, in particular the customer's budget.
  • FIGS. 1 and 2 are block diagrams depicting example methods for maintaining a service on a cloud network.
  • an example method for maintaining a service on a cloud network may generally include receiving a set of status data and performing a scale action based on the set of status data.
  • a set of status data associated with a service may be received.
  • the set of status data may describe the state of the service being provided on a cloud network.
  • the set of status data may include a set of utilization data related to a set of resources utilized by the service, a set of quality data related to a performance level of the service, and a set of workload data related to a workload volume of the service.
  • the cloud network may include a cloud, which may include any combination of a resource, a server, and a server group.
  • the cloud network may include any number of additional clouds. The additional clouds may be of varying sizes.
  • the set of utilization may include or otherwise describe the amount of resources being zed by a service, how much of each of the resources are being utilized by a service, the frequency they are utilized by the service, etc.
  • the resources may be any appropriate resource useable in a computing environment whether physical or virtual, including a CPU, a storage medium, etc.
  • the set of quality data may include or otherwise describe data related to QoS, including service loss, time of completion of requests, latency, etc.
  • the set of workload data may include or otherwise describe the rate of workload performed or requested to be performed by the service. For example, the set of workload data may include the rate of workload per second and the rate of workload may be the rate of service requests received.
  • a more complete decision can be made on whether to scale and in which dimension to scale in comparison to considering solely resource utilization. For example, if the service conditions show a set of resources as having high resource utilization and low quality of service, a resource may be added to the service. However, if workload volume is also high, a server group or an entire cloud may be added. Metrics may be monitored at any appropriate level of the cloud, including au individual resource, a server device, a server group, and/or an entire cloud. A resource, a server device, a server group, and/or a cloud may be referred to as a cloud unit and are discussed in more detail below and in the description associated with FIGS. 2-4 .
  • a scale action may be performed on a cloud network based on a scale rule applied to the set of status data.
  • a scale action may be any appropriate change in the cloud network, such as adding a resource to a service or removing a resource from a service.
  • the scale rule may include a set of conditions to satisfy for the scale action to occur on the cloud network.
  • the set of conditions may include a resource utilization condition associated with the set of resources utilized by the service, a quality condition associated with the performance level of the service, a workload condition associated with the workload volume of the service, and a budget condition associated with a cost of the service.
  • the scale action may involve a cloud unit.
  • a cloud unit may be at least one of a resource, a server, a server group, and a cloud.
  • a cloud unit may be physical, virtual, or a combination thereof.
  • the scale action may include one of adding a cloud unit to the service and removing the cloud unit from the service. For example, a scale action may add a server to the database service and a second scale action may remove a server from the web service.
  • the scale action may occur in any appropriate dimension of scaling, including scaling at a resource level, a server level, a server group level, or a cloud level.
  • a resource level scale may include adding a resource to the service and a cloud level scale may include adding a cloud to the service.
  • a scale action may also scale in a combination of dimensions.
  • the cloud may scale by removing a resource from the service and adding a server group to the service.
  • the scale action may relate to the scale rule and the set of conditions may be satisfied by metrics associated with a cloud network, a service, and/or a customer.
  • a set of conditions may be met, or satisfied, when each of the metrics achieves an associated condition. For example, if a scale action is associated with a scale rule that has a resource utilization condition, a quality condition, a workload condition, and a budget condition, then the scale action may be performed if the set of status data associated with the metrics satisfy the resource utilization condition, the quality condition, the workload condition, and the budget condition.
  • Each condition may be a minimum value, a maximum value, a range of values, or a combination thereof.
  • the workload condition may include a range of rates that satisfy the scale rule.
  • Each condition may be satisfied depending on the type of condition used. For example, if the condition is a minimum value than the condition may be satisfied by a value that is equal to or greater than the minimum value or if the condition is a range than the condition is satisfied by a value within the range.
  • a customer may have a budget to spend on the service and/or the cloud resource provider.
  • a scale action may be performed if the cost of the scaling action is within the budget allotted to the service by the customer. For example, even if the resource utilization condition, the quality condition, and the workload volume condition are satisfied, the scale action may not perform if the budget condition is not satisfied (i.e. the scale action would cost more than what the customer has budgeted for the service).
  • the budget and associated budget condition may be provided in a service level agreement (“SLA”).
  • SLA service level agreement
  • the SLA may also provide the resource utilization condition, the quality condition, and the workload condition.
  • the description of the example method for maintaining a service on a cloud network of FIG. 1 may be applied to the description of the example method for maintaining a service on a cloud network of FIG. 2 , and vice versa.
  • the complete description of one is not repeated in the description of the other and vice versa. Further details of the example methods are provided in the descriptions of FIGS. 3-6 .
  • the set of status data may be derived from the metrics provided by a monitor that may analyze the cloud network.
  • the scale rule may provide a threshold or a set of thresholds to monitor.
  • the monitor may send an event notice when the metric surpasses a threshold of the scale rule.
  • the set of thresholds may include a resource utilization threshold associated with a utilization level of a set of resources of the service, a quality threshold associated with a performance level of the service, and a workload threshold associated with a workload volume of the service.
  • the monitor may monitor individual resources and/or groups of resources during runtime. Thresholds are discussed in more detail in the description associated with FIG. 4 .
  • the set of status data may provide information associated with a group of servers providing the service or a group of servers to provide the service.
  • the group of servers may be grouped based on a characteristic or a relationship.
  • a set of servers may be grouped based on a database service, a web service, and an application service into a database server group, a web server group, and an application group.
  • the scale rule may be associated with a plurality of servers of a server group and may base the set of conditions on the server group.
  • at least one of the set of conditions may include a number of servers of the plurality of servers to satisfy the one of the set of conditions.
  • the set of status data and the scale rule may be in the form of an aggregated amount of individual servers or a percentage of the servers of the group.
  • the scale rule may describe adding an additional web server to a group when 60% of the web server group has a high resource utilization, 20% of the web server group may have high quality of service, and the web server request workload volume per server is high on 70% of the
  • applying a scale rule to a set of status data may include comparing a metric to an associated condition.
  • a set of status data may include a set of workload volume data and the set of workload volume data may be compared to the workload condition.
  • the set of conditions may be described by a domain specific language and may be associated with threshold values used by the monitors in sending event notices, including the set of status data.
  • a domain specific language may have a variable designated for “high” and that variable may be associated with a percentage of a cloud unit at that metric, such as 90% resource utilization.
  • the range of a condition of the scale rule may be defined by a domain specific language where the conditions are determined beforehand or during runtime of the service.
  • the domain specific language may allow the service to scale if the resource utilization is “high,” and the term “high” may be defined by a predetermined condition range.
  • the condition ranges may be based on a machine learning method or otherwise dynamically adaptive during runtime.
  • a machine learning method may be any appropriate method that analyzes data associated with the service on the network that is adaptable to use with a scale rule.
  • the domain specific language may use primitives describing any appropriate value or range of a metric associated with a cloud unit, including resources, servers, server groups, or clouds, that may meet a condition.
  • the primitive may be any appropriate amount of cloud units from none to all and may be described any appropriate manner, such as a percentage, ratio, or a range of values.
  • a duration rule may be applied to an observation duration.
  • the duration rule may include an evaluation window and a set of conditions associated with the time that metrics and/or the set of status data satisfy conditions of the scale rule.
  • the observation duration may be related to an accumulated length of time the set of status data satisfies the scale rule within the evaluation window. For example, if a high workload condition initiates the scale action, the duration rule may set that the workload condition be satisfied for an accumulated one minute of a five minute window before a scale action may occur.
  • the duration rule may include an observation duration to assist against high frequency scaling for temporary conditions. For example, the scale rule may be met for thirty seconds as employees automatically log in to the service at the beginning of the work day, but the customer may not want to scale for this reason.
  • the evaluation window may be a time frame to evaluate the observation duration. For example, the evaluation window may be a sliding five minute window or a static time from six o'clock to eleven o'clock in which the time the scale rule is satisfied may be stored, aggregated, or otherwise accumulated to be the observed duration.
  • the duration rule may assist scaling that may not occur because of the false negatives produced from using a continuous duration. For example, if a system without the duration rule scaled when a workload condition is satisfied for a continuous, one minute time length before initiating a scale action and the workload is high for thirty seconds every two minutes, a scale action may not perform because the scale rule may not be satisfied continually; however, the customer may want to scale the cloud because of the consistency of the requests. If a system using the duration rule is placed in a similar scenario where the observed duration is one minute and the evaluation window is five minutes, the duration rule may be satisfied and the scale action may occur.
  • the evaluation window and the observed duration may change dynamically.
  • the evaluation window may be a time frame of two minutes and the observed duration may be thirty seconds and, during non-peak hours, the evaluation window may change to a time frame of thirty minutes and the observed duration may change to be five minutes.
  • the evaluation window and the observed duration may change dynamically based on a machine-learning method. A machine-learning method is discussed in more detail in the description associated with FIG. 4 .
  • the scale action may be performed on a cloud network and may include adding an entire cloud to the service.
  • a cloud network may be executing service privately and may assign service requests to a public cloud for peak service times.
  • the scale rule may consider varying costs in cloud units, such as resources, of different forms of clouds or cloud networks. For example, performing the scale action in a private cloud may be cheaper than the performing the scale action in a public cloud.
  • a scale action may execute using a network device and/or programming instructions.
  • a server device may instantiate virtual machines to add a cloud unit and may terminate a virtual machine to remove a cloud unit.
  • the scale action may also initiate another action related to maintenance of the cloud.
  • a load balancer may distribute service requests across a server group according to a balancing method by reassigning the service or a service request from a first cloud unit to a second cloud unit.
  • Hardware and programming instructions to implement the example methods of FIGS. 1 and 2 are discussed in more detail in the description associated with FIGS. 4 and 5 .
  • FIG. 3 depicts example operations for maintaining a service on a cloud network.
  • the example operations described herein may be examples of a method and/or a system 300 for maintaining a service on a cloud network.
  • the service may be performed on a cloud network 320 .
  • the cloud network 320 may include a resource 322 , a server device 324 , a server group 326 , a cloud 328 , and/or any additional cloud units, such as cloud units 322 , 324 , 326 , and 328 .
  • the service may execute on any appropriate combination of cloud units.
  • the resource 322 may be any appropriate resource used in computing, such as a hard disk, a random-access memory (“RAM”), a CPU, or virtual representation thereof.
  • the server device 324 may contain multiple resources, including virtual resources.
  • a server device 324 may execute a plurality of virtual machines having a set of resources 322 .
  • the server group 326 may be a group of server devices 324 that provide a related service.
  • a plurality of database servers may be grouped into a database server group.
  • the cloud 328 may be a type of cloud in the cloud network 320 or assigned to a particular service.
  • the cloud network 320 may represent a hybrid cloud that contains a public cloud 328 and a private server group 326 .
  • the cloud 328 may include a plurality of server groups 326 , server devices 324 , and/or resources 322 .
  • the monitor operator 302 may monitor a status of the cloud network 320 and/or a cloud unit of the cloud network 320 .
  • a monitor operator 302 may track resource utilization associated with a resource 322 and workload volume associated with a server group 326 .
  • the monitor operator 302 may retrieve a set of status data based on the status of the cloud network 320 and/or a cloud unit of the cloud network 320 .
  • the monitor operator 302 may be coupled to a data store 310 .
  • the data store 310 may store a set of status data.
  • the monitor operator 302 may work in conjunction with other operators to store a set of status data in the data store 310 .
  • the monitor operator 302 may manage a utilization operator 312 for monitoring resource utilization of a service, a quality operator 314 for monitoring the quality of a service, a workload operator 316 for monitory the workload volume of a service, and a cost operator 318 for retrieving and/or verifying the cost of a service.
  • the operators 302 , 312 , 314 , 316 , and 318 may place data in the data store 310 based on a set of thresholds.
  • the set of thresholds may be based on the set of conditions of the scale rule.
  • the data store 310 may be any appropriate storage medium capable of containing the data retrieved from operators 302 , 312 , 314 , 316 , and 318 , such as the set of status data.
  • the rule operator 304 may apply a scale rule to the set of status data.
  • the rule operator 304 may receive a scale rule, including a set of conditions, from the data store 310 or an application programming interface (“API”) accessible by the customer or the cloud service provider.
  • the rule operator 304 may receive the set of status data from the data store 310 .
  • the set of status data may be applied to the set of conditions and/or set of thresholds associated with the data monitored and/or retrieved by monitor operator 302 .
  • the scale rule may include a resource utilization condition, a quality condition, a workload condition, and a budget condition and a set of status data may be retrieved by the utilization operator 312 , the quality operator 314 , the workload operator 316 , and the cost operator 318 to compare the set of status data to the set of conditions.
  • the scale operator 306 may perform a scale action on the cloud network 320 if the scale rule is satisfied.
  • the scale action may add a cloud unit to the service or remove a cloud unit to the service.
  • the scale operator 306 may add a public cloud 328 to assist completion of the requests.
  • a second scale rule may scale in the cloud network 320 by removing the public cloud 328 from the service.
  • the cloud unit may be virtual. Therefore, a server device 324 or a server group 326 may dynamically instantiate or terminate a virtual resource 322 , a virtual machine (for example a server device 324 ), or any other virtualization of a cloud unit.
  • the duration operator 308 may apply a duration rule to the set of status data.
  • the duration operator 308 may ensure that the scale action occurs when the set of status data satisfies the set of conditions of the scale rule for the observed duration within an evaluation window.
  • the scale operator 306 may not perform a scale action on the cloud network 320 if the duration rule is not satisfied, even if the scale rule is satisfied.
  • the observed duration may be an accumulated length of time the set of status data have met the set of conditions of the scale rule.
  • the operators 302 , 304 , 306 , and 308 of FIG. 3 described above may represent operations, processes, interactions, and/or other actions performed by or in connection with the engines 402 , 404 , 406 , and 408 of FIG. 4 .
  • FIGS. 4 and 5 are block diagrams depicting example systems for maintaining a service on a cloud network.
  • an example system 400 for maintaining a service on a cloud network may generally include a monitor engine 402 , a rule engine 404 , and a scale engine 406 .
  • the example system 400 may also include a duration engine 408 and a data store 410 .
  • the data storage 410 may be the same as the data store 310 of FIG. 3 and the description associated with the data store 310 may be applied to the data store 410 , and vice versa, as appropriate.
  • the monitor engine 402 may represent any appropriate combination of hardware and programming configured to maintain a set of status data associated with a resource utilization of the service, a quality of the service, and a workload volume of the service.
  • the monitor engine 402 may maintain the set of status data continuously, dynamically, or based on a schedule.
  • the monitor engine 402 may receive a set of thresholds from the rule engine 404 based on a set of conditions of a scale rule.
  • the set of thresholds may include a resource utilization threshold associated with a set of resources utilized by the service, a quality threshold associated with a performance level of the service, and a workload threshold associated with the workload volume of the service.
  • the resource utilization threshold may include a utilization level of the set of resources of the service, a percentage of a resource used by the service, or other metric related to resource utilization.
  • the quality threshold may include a request completion rate, a packet loss rate, latency, or other metrics related to performance quality.
  • the workload threshold may include service request per time period or other metric related to a rate of workload, or workload volume.
  • the set of status data may include data related to the set of thresholds.
  • the set of status data may include a set of resource utilization data associated with a set of resources utilized by the service, a set of quality data associated with a performance level of the service, and a set of workload data associated with the workload volume of the service.
  • the monitor engine 402 may be coupled to a data store 410 .
  • the monitor engine 402 may maintain the set of status data by adding, modifying, updating, and/or deleting a set of status data contained in a data store 410 .
  • the data store 410 may be a storage medium capable of containing data associated with the service, such as a set of resource utilization data, a set of quality data, a set of workload data, and a set of cost data.
  • the cost of the service and the budget of the customer may be made available to the monitor engine 402 by the rule engine 404 described below.
  • the set of status data may include the cost of the service and the budget condition associated with the cost of the service.
  • the cost of the service may be attainable from a public server, a website, an API an SLA, or any other document or accessible medium.
  • the cost of the service may include the cloud unit cost associated with one of the resource, the server, the server group, and the cloud.
  • the monitor engine 402 may monitor the cloud network based on individual cloud units, service transactions, or a cloud unit group.
  • the cloud unit group may include a plurality of cloud units utilized by the service where each of the plurality of cloud units is related by type or another appropriate relationship for group the cloud unit.
  • each of the plurality of cloud units may have a service relationship to each of the plurality of cloud units, such as a plurality of web servers.
  • the monitor engine 402 may aggregate the set of status data of the cloud unit group.
  • the set of conditions may be based on a ratio of a portion of the plurality of the cloud units that satisfy the set of conditions to the plurality of cloud units. For example, if there are ten cloud units and the scale action will be performed if the set of conditions of the scale rule is satisfied by 70% of the cloud units, then the monitor engine 402 may send out a set of status data when seven of the ten cloud units satisfy the set of conditions.
  • the monitor engine 402 may monitor the cloud network and/or maintain the data store 410 continuously, dynamically, or based on a schedule. For example, the monitor engine 402 may obtain status updates every minute.
  • the monitor engine 402 may send an event notification containing the set of status data to the rule engine 404 continuously, dynamically, or based on a schedule. For example, the monitor engine 402 may send an event notification to the rule engine 404 when a threshold is achieved.
  • the rule engine 404 may receive the event notification and compare the set of status data to the scale rule.
  • the rule engine 404 may represent any appropriate combination of hardware and programming configured to apply a scale rule to the set of status data.
  • the scale rule may include a set of conditions to be satisfied by the set of status data.
  • the set of conditions may include a resource utilization condition to compare to the resource utilization of the service, a quality condition to compare to the quality of the service, a workload condition to compare to the workload volume of the service, and a budget condition to compare to a cost of the service.
  • the rule engine 404 may receive the set of conditions based on a domain specific language and may use the set of conditions to determine a set of thresholds to be monitored by the monitor engine 402 .
  • a domain specific language may provide primitives of high, medium, and low, that may be determined by the rule engine 404 to monitor a set of thresholds that are greater than 71%, between 70% and 41%, and less than 40% respectively.
  • the domain specific language may allow the rule engine 404 to scale based on maximum values, minimum values, ratios, and/or percentages of a cloud unit group.
  • the rule engine 404 may receive the set of status data from the monitor engine 402 .
  • the set of status data may be the level of the metrics monitored by the monitor engine 402 .
  • the set of status data may include the level of resource utilization, the level of performance, and the level of workload volume.
  • the set of status data may be a primitive of the domain specific language associated with the metrics monitored by the monitor engine 402 .
  • the monitor engine 402 may send an event to the rule engine 404 when the web server resource utilization is at 80% and the set of status data may be the high primitive; the rule engine 404 may then compare the set of conditions of the scale rule to the high primitive.
  • At least one of the resource utilization condition, the quality condition, the workload condition, the budget condition, and the observed duration condition may be maintained dynamically based on a machine-learning method.
  • the machine-learning method may observe the service and learn the environment and times when the set of conditions are satisfied or times when the cloud network scales.
  • the machine-learning method may apply the knowledge learned an update the set of conditions and/or the set of thresholds, accordingly.
  • the scale engine 406 may represent any appropriate combination of hardware and programming configured to maintain a cloud unit of a cloud network based on a scale rule and a set of status data.
  • the scale engine 406 may maintain a service by performing a scale action on the cloud network.
  • the scale engine 406 may perform a scale action based on the scale rule applied to the set of status data. For example, based on the scale rule, the scale engine 406 may be configured to at least one of add the cloud unit to the service and remove the cloud unit from the service.
  • the scale engine 406 may be configured to receive a scale notification from the rule engine 404 .
  • the rule engine 404 may send a scale notification to the scale engine 406 when the set of conditions of the scale rule have been satisfied.
  • the scale notification may include the dimension(s) to scale and the cloud unit(s) to scale.
  • the scale notification may contain a request to add a server group to the service.
  • the scale engine 406 may request a cloud network action in conjunction with the scale actions. For example, a load balancing action may balance the service load when a cloud unit is added to or removed from the service.
  • the duration engine 408 may represent any appropriate combination of hardware and programming configured to apply a duration rule to an accumulated length of time the set of status data satisfies the scale rule within an evaluated window.
  • the duration engine 408 may track the accumulated length of time the set of status data satisfies the scale rule within an evaluation window. For example, the duration engine 408 may analyze the set of status data in the data store 410 based on an evaluation window to compare whether the accumulated length of time within the evaluation window achieves an observed duration threshold to satisfy an observed duration condition.
  • the duration engine 408 may work in conjunction with the monitor engine 402 and the rule engine 404 to increase a time variable, such as an observed duration, as the set of status data satisfies the scale rule, and decrease the time variable as the evaluation window changes.
  • the scale engine 406 may maintain the cloud unit if the accumulated length of time satisfies an observed duration condition.
  • the duration engine 408 may be configured to send a duration notification to the scale engine 406 and/or the rule engine 404 .
  • the scale engine 406 may wait to perform a scale action until both a scale notification and a duration notification are received.
  • the rule engine 404 may wait to send a scale notification until a duration notification is received.
  • FIG. 5 depicts that the example system 500 for maintaining a service on a cloud network may be implemented on a memory resource 530 operatively coupled to a processor resource 532 .
  • the processor resource 532 may be operatively coupled to the data store 510 .
  • the data store 510 may be the same as the data store 310 of FIG. 3 and/or the data store 410 of FIG. 4 and the description associated with the data store 310 and/or the data store 410 may be applied to the data store 510 , and vice versa, as appropriate.
  • the memory resource 530 may contain a set of instructions that may be executable by the processor resource 532 .
  • the set of instructions may implement the system 500 when executed by the processor resource 532 .
  • the set of instructions stored on the memory resource 530 may be represented as a monitor module 502 , a rule module 504 , a scale module 506 , and a duration module 508 .
  • the processor resource 532 may carry out the set of instructions to execute the monitor module 502 , the rule module 504 , the scale module 506 , the duration module 508 , and/or any appropriate operations between or otherwise associated with the modules of the system 500 .
  • the processor resource 532 may carry out a set of instructions to maintain a set of status data associated with a resource utilization of the service, apply the set of status data to a scale rule, and maintain a cloud unit based on the scale rule.
  • the monitor module 502 may represent program instructions that when executed function as the monitor engine 402 of FIG. 4 .
  • the rule module 504 may represent program instructions that when executed function as the rule engine 404 of FIG. 4 .
  • the scale module 506 may represent program instructions that when executed function as the scale engine 406 of FIG. 4 .
  • the duration module 508 may represent program instructions that when executed as the duration engine 408 of FIG. 4 .
  • the processor resource 532 may be one or multiple CPUs capable of retrieving instructions from the memory resource 530 and executing those instructions.
  • the processor resource 532 may process the instructions serially, concurrently, or in partial concurrence, unless described otherwise herein.
  • the memory resource 530 and the data store 510 may represent a medium to store data utilized by the system 500 .
  • the medium may be any non-transitory medium or combination of non-transitory mediums able to electronically store data and/or capable of storing the modules of the system 500 and/or data used by the system 500 .
  • a storage medium, such as memory resource 530 may be distinct from a transmission medium, such as a signal.
  • the medium may be machine-readable, such as computer-readable.
  • the data of the data store 510 may include representations of resource utilization, QoS, workload volume, cost, and/or other data or information mentioned herein, such as a duration values and/or time values.
  • engines 402 , 404 , 406 , and 408 and the module 502 , 504 , 506 , and 508 have been described as combinations of hardware and programming. Such components may be implemented in a number of fashions.
  • the programming may be processor executable instructions stored on the memory resource 530 , which is a tangible, non-transitory computer readable storage medium, and the hardware may include processor resource 532 for executing those instructions.
  • the processor resource 532 may include one or multiple processors. Such multiple processors may be integrated in a single device or distributed across devices. For example in reference to FIG. 6 , the processor resource 532 may be distributed across any combination of server devices 692 and client devices 694 .
  • the memory resource 530 may be said to store program instructions that when executed by processor resource 532 implements the system 500 in FIG. 5 .
  • the memory resource 530 may be integrated in the same device as processor resource 532 or it may be separate but accessible to that device and processor resource 532 .
  • the memory resource 530 may be distributed across devices, such as server devices 692 and client devices 694 of FIG. 6 .
  • the memory resource 530 and the data store 510 may represent the same physical medium unless otherwise described herein.
  • the program instructions may be part of an installation package that when installed may be executed by processor resource 532 to implement the system 500 .
  • memory resource 530 may be a portable medium such as a CD, DVD, or flash drive or memory maintained by a server device, such as server device 692 of FIG. 6 , from which the installation package may be downloaded and installed.
  • the program instructions may be part of an application or applications already installed.
  • the memory resource 530 may include integrated memory such as a hard drive, solid state drive, or the like.
  • FIG. 6 depicts example environments in which various examples for maintaining a service on a cloud network may be implemented.
  • the environment 690 is shown to include a cloud scale system 600 for maintaining a service on a cloud network.
  • the system 600 described herein with respect to FIGS. 4 and 5 may represent generally any combination of hardware and programming configured to maintain a service on a cloud network.
  • the system 600 may be integrated into a server device 692 or a client device 694 .
  • the system 600 may be distributed across server devices 692 , client devices 694 , or a combination of server devices 692 and client devices 694 .
  • the environment 690 may be a cloud computing environment.
  • any combination of the system 600 , server device 692 , and client devices 694 may be a virtual instance and/or may reside and/or execute on a cloud.
  • the environment 690 may include a cloud computing environment, such as a cloud network 620 .
  • the cloud network 620 may be the same as the cloud network 320 of FIG. 3 and the description associated with the cloud network 320 may be applied to the cloud network 620 , and vice versa, as appropriate.
  • a server device 692 may access a client device 694 .
  • the server devices 692 may represent generally any computing devices configured to respond to a network request received from the client device 694 .
  • a server device 692 may be a cloud unit of the cloud network 620 providing the service and the client device 694 may be a computing device configured to access the cloud network 620 and receive and/or communicate with the service.
  • a server device 692 may include a web server, an application server, or a data server.
  • the client devices 694 may represent generally any computing devices configured with a browser or other application to communicate such requests and receive and/or process the corresponding responses.
  • a link 696 may represent generally one or any combination of a cable, wireless, fiber optic, or remote connections via a telecommunications link, an infrared link, a radio frequency link, or any other connectors of systems that provide electronic communication.
  • the link 696 may include, at least in part, intranet, the Internet, or a combination of both.
  • the link 696 may also include intermediate proxies, routers, switches, load balancers, and the like.
  • the engines 402 , 404 , 406 , and 408 and/or the module 502 , 504 , 506 , and 508 may be distributed across any combination of server devices, client devices, and storage mediums.
  • the engines 402 , 404 , 406 , and 408 and/or the module 502 , 504 , 506 , and 508 may complete or assist completion of operations performed in describing another engine 402 , 404 , 406 and 408 and/or the module 502 , 504 , 506 , and 508 .
  • the engines 402 , 404 , 406 , and 408 and/or the module 502 , 504 , 506 , and 508 may perform the example operations of operators 302 , 304 , 306 , and 308 and/or the example methods described in connection with FIGS. 1 and 2 .
  • FIGS. 1 and 2 illustrate specific orders of execution
  • the order of execution may differ from that which is illustrated.
  • the order of execution of the blocks may be scrambled relative to the order shown.
  • the blocks shown in succession may be executed concurrently or with partial concurrence. All such variations are within the scope of the present invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer And Data Communications (AREA)
  • Debugging And Monitoring (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

Maintaining a service on a cloud network may include receiving a set of status data associated with the service and performing a scale action on the cloud network based on a scale rule applied to the set of status data. The set of status data may be related to a set of resources utilized by the service, a performance level of the service, and a workload volume of the service. The scale rule may include a utilization condition, a quality condition, a workload condition, and a budget condition.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application is a U.S. National Stage Application of and claims priority to international Patent Application No. PCT/US2013/057738, filed on Aug. 30, 2013, and entitled “MAINTAIN A SERVICE ON A CLOUD NETWORK BASED ON A SCALE RULE”.
BACKGROUND
Cloud computing commonly utilizes a shared pool of computing resources. The computing environment, including the shared pool of resources, may be referred to as a “cloud.” The resources may be virtual instances of resources running on a virtual machine. The pool of computing resource may provide a service. Cloud-based services are said to execute or reside “on the cloud.” The computing environment, or cloud, may adapt to an increase in service usage by providing additional resource instances to the pool. The shared pool of resources may reside internal to a corporate network or external as part of the public Internet, respectively known as the private cloud and the public cloud. A hybrid cloud environment may utilize both a private cloud and a public cloud. For example, a service provider may provide a private cloud for non-peak service usage and the service provider may offload service requests to the public cloud during peak times, or times of high request rates. Cloud resource providers may provide resources, such as a central processing unit (“CPU”) or computer memory, for services based on the type of cloud computing environment, which may include virtual machines. Cloud resource providers may offer variations on resource provisioning for services, including infrastructure as a service, platform as a service, storage as a service, and software as a service. Cloud resource providers may rent space on the cloud to service providers, or customers, for the service based on a cost per time period, cost per resource, subscription, or any other appropriate cost method.
BRIEF DESCRIPTION OF THE DRAWINGS
FIGS. 1 and 2 are block diagrams depicting example methods for maintaining a service on a cloud network.
FIG. 3 depicts example operations for maintaining a service on a cloud network.
FIGS. 4 and 5 are block diagrams depicting example systems for maintaining a service on a cloud network.
FIG. 6 depicts example environments in which various examples for maintaining a service on a cloud network may be implemented.
DETAILED DESCRIPTION
In the following description and figures, some example implementations of systems and/or methods for maintaining a service on a cloud network are described. As used herein, a service may be any appropriate supplying of communication, transmissions, information, software, storage, or any other product or activity that may be capable of running on the cloud, or a cloud-based service. As used herein, a cloud may be any appropriate combination of physical and virtual resources pooled for computing and/or storage purposes. For example, a cloud may include any appropriate number of individual resources, servers, and server groups including virtual instances of resources, servers, and server groups. As used herein, a cloud network may include any appropriate number of clouds and/or other network of resources accessible by a network device.
Cloud resource providers may endeavor to maintain a quality of service for a customer based on resource utilization, or the amount and/or degree of resources utilized by the service. For example, the amount of resources assigned to a service may be highly utilized or underutilized based on any given time of the day. Cloud service providers may maintain the service by scaling the cloud based on the amount of resources utilized by the service. Scaling may refer to increasing or decreasing the shared pool of resources. Cloud resource providers may offer the ability to dynamically scale the cloud to adapt, to service spikes and lulls based on resource utilization. A service spike may occur when there is an increase in service requests or requests associated with providing the service. Accordingly, scaling based on resource utilization may enable an increase in assigning resource for utilizations that may be indirectly associated to the service. For example, the cloud resource provider may add cloud resources every day for a daily virus scan. The customer may be paying for an increase of cloud resources at times of unnecessary or unwanted scaling and may not know until arrival of the bill from the cloud resource provider.
Various examples described below relate to maintaining a service on a cloud network based on a scale rule. By monitoring resource utilization, quality of service (“QoS”), and workload volume, the cloud may be scaled while accounting for cost in the scale rule. For example, a cloud may be scaled based on resource utilization, QoS, the amount of service requests and a budget to reduce insufficient or excessive allocation of cloud resources to a service. The scale rule may account for and/or be adaptable to the desires and/or needs of the customer, in particular the customer's budget.
FIGS. 1 and 2 are block diagrams depicting example methods for maintaining a service on a cloud network. Referring to FIG. 1, an example method for maintaining a service on a cloud network may generally include receiving a set of status data and performing a scale action based on the set of status data.
In block 102, a set of status data associated with a service may be received. The set of status data may describe the state of the service being provided on a cloud network. For example, the set of status data may include a set of utilization data related to a set of resources utilized by the service, a set of quality data related to a performance level of the service, and a set of workload data related to a workload volume of the service. The cloud network may include a cloud, which may include any combination of a resource, a server, and a server group. The cloud network may include any number of additional clouds. The additional clouds may be of varying sizes.
The set of utilization may include or otherwise describe the amount of resources being zed by a service, how much of each of the resources are being utilized by a service, the frequency they are utilized by the service, etc. The resources may be any appropriate resource useable in a computing environment whether physical or virtual, including a CPU, a storage medium, etc. The set of quality data may include or otherwise describe data related to QoS, including service loss, time of completion of requests, latency, etc. The set of workload data may include or otherwise describe the rate of workload performed or requested to be performed by the service. For example, the set of workload data may include the rate of workload per second and the rate of workload may be the rate of service requests received.
By considering metrics, or analysis data, associated with resource utilization, QoS, and workload volume, a more complete decision can be made on whether to scale and in which dimension to scale in comparison to considering solely resource utilization. For example, if the service conditions show a set of resources as having high resource utilization and low quality of service, a resource may be added to the service. However, if workload volume is also high, a server group or an entire cloud may be added. Metrics may be monitored at any appropriate level of the cloud, including au individual resource, a server device, a server group, and/or an entire cloud. A resource, a server device, a server group, and/or a cloud may be referred to as a cloud unit and are discussed in more detail below and in the description associated with FIGS. 2-4.
In block 104, a scale action may be performed on a cloud network based on a scale rule applied to the set of status data. A scale action may be any appropriate change in the cloud network, such as adding a resource to a service or removing a resource from a service. The scale rule may include a set of conditions to satisfy for the scale action to occur on the cloud network. For example, the set of conditions may include a resource utilization condition associated with the set of resources utilized by the service, a quality condition associated with the performance level of the service, a workload condition associated with the workload volume of the service, and a budget condition associated with a cost of the service.
The scale action may involve a cloud unit. A cloud unit may be at least one of a resource, a server, a server group, and a cloud. A cloud unit may be physical, virtual, or a combination thereof. The scale action may include one of adding a cloud unit to the service and removing the cloud unit from the service. For example, a scale action may add a server to the database service and a second scale action may remove a server from the web service.
The scale action may occur in any appropriate dimension of scaling, including scaling at a resource level, a server level, a server group level, or a cloud level. For example, a resource level scale may include adding a resource to the service and a cloud level scale may include adding a cloud to the service. A scale action may also scale in a combination of dimensions. For example, the cloud may scale by removing a resource from the service and adding a server group to the service.
The scale action may relate to the scale rule and the set of conditions may be satisfied by metrics associated with a cloud network, a service, and/or a customer. A set of conditions may be met, or satisfied, when each of the metrics achieves an associated condition. For example, if a scale action is associated with a scale rule that has a resource utilization condition, a quality condition, a workload condition, and a budget condition, then the scale action may be performed if the set of status data associated with the metrics satisfy the resource utilization condition, the quality condition, the workload condition, and the budget condition. Each condition may be a minimum value, a maximum value, a range of values, or a combination thereof. For example, the workload condition may include a range of rates that satisfy the scale rule. Each condition may be satisfied depending on the type of condition used. For example, if the condition is a minimum value than the condition may be satisfied by a value that is equal to or greater than the minimum value or if the condition is a range than the condition is satisfied by a value within the range.
A customer may have a budget to spend on the service and/or the cloud resource provider. A scale action may be performed if the cost of the scaling action is within the budget allotted to the service by the customer. For example, even if the resource utilization condition, the quality condition, and the workload volume condition are satisfied, the scale action may not perform if the budget condition is not satisfied (i.e. the scale action would cost more than what the customer has budgeted for the service). The budget and associated budget condition may be provided in a service level agreement (“SLA”). The SLA may also provide the resource utilization condition, the quality condition, and the workload condition.
The description of the example method for maintaining a service on a cloud network of FIG. 1 may be applied to the description of the example method for maintaining a service on a cloud network of FIG. 2, and vice versa. For brevity, the complete description of one is not repeated in the description of the other and vice versa. Further details of the example methods are provided in the descriptions of FIGS. 3-6.
In block 202, the set of status data may be derived from the metrics provided by a monitor that may analyze the cloud network. The scale rule may provide a threshold or a set of thresholds to monitor. For example, the monitor may send an event notice when the metric surpasses a threshold of the scale rule. The set of thresholds may include a resource utilization threshold associated with a utilization level of a set of resources of the service, a quality threshold associated with a performance level of the service, and a workload threshold associated with a workload volume of the service. The monitor may monitor individual resources and/or groups of resources during runtime. Thresholds are discussed in more detail in the description associated with FIG. 4.
The set of status data may provide information associated with a group of servers providing the service or a group of servers to provide the service. The group of servers may be grouped based on a characteristic or a relationship. For example, a set of servers may be grouped based on a database service, a web service, and an application service into a database server group, a web server group, and an application group. The scale rule may be associated with a plurality of servers of a server group and may base the set of conditions on the server group. For example, at least one of the set of conditions may include a number of servers of the plurality of servers to satisfy the one of the set of conditions. The set of status data and the scale rule may be in the form of an aggregated amount of individual servers or a percentage of the servers of the group. For example, the scale rule may describe adding an additional web server to a group when 60% of the web server group has a high resource utilization, 20% of the web server group may have high quality of service, and the web server request workload volume per server is high on 70% of the web server group.
In block 204, applying a scale rule to a set of status data may include comparing a metric to an associated condition. For example, a set of status data may include a set of workload volume data and the set of workload volume data may be compared to the workload condition.
The set of conditions may be described by a domain specific language and may be associated with threshold values used by the monitors in sending event notices, including the set of status data. For example, a domain specific language may have a variable designated for “high” and that variable may be associated with a percentage of a cloud unit at that metric, such as 90% resource utilization. The range of a condition of the scale rule may be defined by a domain specific language where the conditions are determined beforehand or during runtime of the service. For example, the domain specific language may allow the service to scale if the resource utilization is “high,” and the term “high” may be defined by a predetermined condition range. Alternatively, the condition ranges may be based on a machine learning method or otherwise dynamically adaptive during runtime. A machine learning method may be any appropriate method that analyzes data associated with the service on the network that is adaptable to use with a scale rule. The domain specific language may use primitives describing any appropriate value or range of a metric associated with a cloud unit, including resources, servers, server groups, or clouds, that may meet a condition. For example, the primitive may be any appropriate amount of cloud units from none to all and may be described any appropriate manner, such as a percentage, ratio, or a range of values.
In block 206, a duration rule may be applied to an observation duration. The duration rule may include an evaluation window and a set of conditions associated with the time that metrics and/or the set of status data satisfy conditions of the scale rule.
The observation duration may be related to an accumulated length of time the set of status data satisfies the scale rule within the evaluation window. For example, if a high workload condition initiates the scale action, the duration rule may set that the workload condition be satisfied for an accumulated one minute of a five minute window before a scale action may occur. The duration rule may include an observation duration to assist against high frequency scaling for temporary conditions. For example, the scale rule may be met for thirty seconds as employees automatically log in to the service at the beginning of the work day, but the customer may not want to scale for this reason. The evaluation window may be a time frame to evaluate the observation duration. For example, the evaluation window may be a sliding five minute window or a static time from six o'clock to eleven o'clock in which the time the scale rule is satisfied may be stored, aggregated, or otherwise accumulated to be the observed duration.
The duration rule may assist scaling that may not occur because of the false negatives produced from using a continuous duration. For example, if a system without the duration rule scaled when a workload condition is satisfied for a continuous, one minute time length before initiating a scale action and the workload is high for thirty seconds every two minutes, a scale action may not perform because the scale rule may not be satisfied continually; however, the customer may want to scale the cloud because of the consistency of the requests. If a system using the duration rule is placed in a similar scenario where the observed duration is one minute and the evaluation window is five minutes, the duration rule may be satisfied and the scale action may occur.
The evaluation window and the observed duration may change dynamically. For example, during peak hours, the evaluation window may be a time frame of two minutes and the observed duration may be thirty seconds and, during non-peak hours, the evaluation window may change to a time frame of thirty minutes and the observed duration may change to be five minutes. The evaluation window and the observed duration may change dynamically based on a machine-learning method. A machine-learning method is discussed in more detail in the description associated with FIG. 4.
In block 208, the scale action may be performed on a cloud network and may include adding an entire cloud to the service. For example, a cloud network may be executing service privately and may assign service requests to a public cloud for peak service times. The scale rule may consider varying costs in cloud units, such as resources, of different forms of clouds or cloud networks. For example, performing the scale action in a private cloud may be cheaper than the performing the scale action in a public cloud.
A scale action may execute using a network device and/or programming instructions. For example, a server device may instantiate virtual machines to add a cloud unit and may terminate a virtual machine to remove a cloud unit. The scale action may also initiate another action related to maintenance of the cloud. For example after a scale action, a load balancer may distribute service requests across a server group according to a balancing method by reassigning the service or a service request from a first cloud unit to a second cloud unit. Hardware and programming instructions to implement the example methods of FIGS. 1 and 2 are discussed in more detail in the description associated with FIGS. 4 and 5.
FIG. 3 depicts example operations for maintaining a service on a cloud network. The example operations described herein may be examples of a method and/or a system 300 for maintaining a service on a cloud network. The service may be performed on a cloud network 320. The cloud network 320 may include a resource 322, a server device 324, a server group 326, a cloud 328, and/or any additional cloud units, such as cloud units 322, 324, 326, and 328. The service may execute on any appropriate combination of cloud units.
The resource 322 may be any appropriate resource used in computing, such as a hard disk, a random-access memory (“RAM”), a CPU, or virtual representation thereof. The server device 324 may contain multiple resources, including virtual resources. For example, a server device 324 may execute a plurality of virtual machines having a set of resources 322. The server group 326 may be a group of server devices 324 that provide a related service. For example, a plurality of database servers may be grouped into a database server group. The cloud 328 may be a type of cloud in the cloud network 320 or assigned to a particular service. For example, the cloud network 320 may represent a hybrid cloud that contains a public cloud 328 and a private server group 326. The cloud 328 may include a plurality of server groups 326, server devices 324, and/or resources 322.
The monitor operator 302 may monitor a status of the cloud network 320 and/or a cloud unit of the cloud network 320. For example, a monitor operator 302 may track resource utilization associated with a resource 322 and workload volume associated with a server group 326. The monitor operator 302 may retrieve a set of status data based on the status of the cloud network 320 and/or a cloud unit of the cloud network 320.
The monitor operator 302 may be coupled to a data store 310. The data store 310 may store a set of status data. The monitor operator 302 may work in conjunction with other operators to store a set of status data in the data store 310. For example, the monitor operator 302 may manage a utilization operator 312 for monitoring resource utilization of a service, a quality operator 314 for monitoring the quality of a service, a workload operator 316 for monitory the workload volume of a service, and a cost operator 318 for retrieving and/or verifying the cost of a service. The operators 302, 312, 314, 316, and 318 may place data in the data store 310 based on a set of thresholds. The set of thresholds may be based on the set of conditions of the scale rule. The data store 310 may be any appropriate storage medium capable of containing the data retrieved from operators 302, 312, 314, 316, and 318, such as the set of status data.
The rule operator 304 may apply a scale rule to the set of status data. The rule operator 304 may receive a scale rule, including a set of conditions, from the data store 310 or an application programming interface (“API”) accessible by the customer or the cloud service provider. The rule operator 304 may receive the set of status data from the data store 310. The set of status data may be applied to the set of conditions and/or set of thresholds associated with the data monitored and/or retrieved by monitor operator 302. For example, the scale rule may include a resource utilization condition, a quality condition, a workload condition, and a budget condition and a set of status data may be retrieved by the utilization operator 312, the quality operator 314, the workload operator 316, and the cost operator 318 to compare the set of status data to the set of conditions.
The scale operator 306 may perform a scale action on the cloud network 320 if the scale rule is satisfied. The scale action may add a cloud unit to the service or remove a cloud unit to the service. For example, if the scale rule for scaling out the cloud network 320 to service a high amount of requests is satisfied, then the scale operator 306 may add a public cloud 328 to assist completion of the requests. Once the amount of requests reduces, a second scale rule may scale in the cloud network 320 by removing the public cloud 328 from the service. As mentioned above, the cloud unit may be virtual. Therefore, a server device 324 or a server group 326 may dynamically instantiate or terminate a virtual resource 322, a virtual machine (for example a server device 324), or any other virtualization of a cloud unit.
The duration operator 308 may apply a duration rule to the set of status data. The duration operator 308 may ensure that the scale action occurs when the set of status data satisfies the set of conditions of the scale rule for the observed duration within an evaluation window. For example, the scale operator 306 may not perform a scale action on the cloud network 320 if the duration rule is not satisfied, even if the scale rule is satisfied. As discussed herein, the observed duration may be an accumulated length of time the set of status data have met the set of conditions of the scale rule.
In general, the operators 302, 304, 306, and 308 of FIG. 3 described above may represent operations, processes, interactions, and/or other actions performed by or in connection with the engines 402, 404, 406, and 408 of FIG. 4.
FIGS. 4 and 5 are block diagrams depicting example systems for maintaining a service on a cloud network. Referring to FIG. 4, an example system 400 for maintaining a service on a cloud network may generally include a monitor engine 402, a rule engine 404, and a scale engine 406. The example system 400 may also include a duration engine 408 and a data store 410. The data storage 410 may be the same as the data store 310 of FIG. 3 and the description associated with the data store 310 may be applied to the data store 410, and vice versa, as appropriate.
The monitor engine 402 may represent any appropriate combination of hardware and programming configured to maintain a set of status data associated with a resource utilization of the service, a quality of the service, and a workload volume of the service. The monitor engine 402 may maintain the set of status data continuously, dynamically, or based on a schedule.
The monitor engine 402 may receive a set of thresholds from the rule engine 404 based on a set of conditions of a scale rule. The set of thresholds may include a resource utilization threshold associated with a set of resources utilized by the service, a quality threshold associated with a performance level of the service, and a workload threshold associated with the workload volume of the service. The resource utilization threshold may include a utilization level of the set of resources of the service, a percentage of a resource used by the service, or other metric related to resource utilization. The quality threshold may include a request completion rate, a packet loss rate, latency, or other metrics related to performance quality. The workload threshold may include service request per time period or other metric related to a rate of workload, or workload volume.
The set of status data may include data related to the set of thresholds. For example, the set of status data may include a set of resource utilization data associated with a set of resources utilized by the service, a set of quality data associated with a performance level of the service, and a set of workload data associated with the workload volume of the service.
The monitor engine 402 may be coupled to a data store 410. The monitor engine 402 may maintain the set of status data by adding, modifying, updating, and/or deleting a set of status data contained in a data store 410. The data store 410 may be a storage medium capable of containing data associated with the service, such as a set of resource utilization data, a set of quality data, a set of workload data, and a set of cost data.
The cost of the service and the budget of the customer may be made available to the monitor engine 402 by the rule engine 404 described below. The set of status data may include the cost of the service and the budget condition associated with the cost of the service. The cost of the service may be attainable from a public server, a website, an API an SLA, or any other document or accessible medium. The cost of the service may include the cloud unit cost associated with one of the resource, the server, the server group, and the cloud.
The monitor engine 402 may monitor the cloud network based on individual cloud units, service transactions, or a cloud unit group. The cloud unit group may include a plurality of cloud units utilized by the service where each of the plurality of cloud units is related by type or another appropriate relationship for group the cloud unit. For example, each of the plurality of cloud units may have a service relationship to each of the plurality of cloud units, such as a plurality of web servers.
The monitor engine 402 may aggregate the set of status data of the cloud unit group. The set of conditions may be based on a ratio of a portion of the plurality of the cloud units that satisfy the set of conditions to the plurality of cloud units. For example, if there are ten cloud units and the scale action will be performed if the set of conditions of the scale rule is satisfied by 70% of the cloud units, then the monitor engine 402 may send out a set of status data when seven of the ten cloud units satisfy the set of conditions.
The monitor engine 402 may monitor the cloud network and/or maintain the data store 410 continuously, dynamically, or based on a schedule. For example, the monitor engine 402 may obtain status updates every minute. The monitor engine 402 may send an event notification containing the set of status data to the rule engine 404 continuously, dynamically, or based on a schedule. For example, the monitor engine 402 may send an event notification to the rule engine 404 when a threshold is achieved. The rule engine 404 may receive the event notification and compare the set of status data to the scale rule.
The rule engine 404 may represent any appropriate combination of hardware and programming configured to apply a scale rule to the set of status data. The scale rule may include a set of conditions to be satisfied by the set of status data. For example, the set of conditions may include a resource utilization condition to compare to the resource utilization of the service, a quality condition to compare to the quality of the service, a workload condition to compare to the workload volume of the service, and a budget condition to compare to a cost of the service.
The rule engine 404 may receive the set of conditions based on a domain specific language and may use the set of conditions to determine a set of thresholds to be monitored by the monitor engine 402. For example, a domain specific language may provide primitives of high, medium, and low, that may be determined by the rule engine 404 to monitor a set of thresholds that are greater than 71%, between 70% and 41%, and less than 40% respectively. The domain specific language may allow the rule engine 404 to scale based on maximum values, minimum values, ratios, and/or percentages of a cloud unit group.
The rule engine 404 may receive the set of status data from the monitor engine 402. The set of status data may be the level of the metrics monitored by the monitor engine 402. For example, the set of status data may include the level of resource utilization, the level of performance, and the level of workload volume. The set of status data may be a primitive of the domain specific language associated with the metrics monitored by the monitor engine 402. For example, the monitor engine 402 may send an event to the rule engine 404 when the web server resource utilization is at 80% and the set of status data may be the high primitive; the rule engine 404 may then compare the set of conditions of the scale rule to the high primitive.
At least one of the resource utilization condition, the quality condition, the workload condition, the budget condition, and the observed duration condition (discussed with the duration engine 408 below) may be maintained dynamically based on a machine-learning method. The machine-learning method may observe the service and learn the environment and times when the set of conditions are satisfied or times when the cloud network scales. The machine-learning method may apply the knowledge learned an update the set of conditions and/or the set of thresholds, accordingly.
The scale engine 406 may represent any appropriate combination of hardware and programming configured to maintain a cloud unit of a cloud network based on a scale rule and a set of status data. The scale engine 406 may maintain a service by performing a scale action on the cloud network. The scale engine 406 may perform a scale action based on the scale rule applied to the set of status data. For example, based on the scale rule, the scale engine 406 may be configured to at least one of add the cloud unit to the service and remove the cloud unit from the service.
The scale engine 406 may be configured to receive a scale notification from the rule engine 404. For example, the rule engine 404 may send a scale notification to the scale engine 406 when the set of conditions of the scale rule have been satisfied. The scale notification may include the dimension(s) to scale and the cloud unit(s) to scale. For example, the scale notification may contain a request to add a server group to the service.
The scale engine 406 may request a cloud network action in conjunction with the scale actions. For example, a load balancing action may balance the service load when a cloud unit is added to or removed from the service.
The duration engine 408 may represent any appropriate combination of hardware and programming configured to apply a duration rule to an accumulated length of time the set of status data satisfies the scale rule within an evaluated window. The duration engine 408 may track the accumulated length of time the set of status data satisfies the scale rule within an evaluation window. For example, the duration engine 408 may analyze the set of status data in the data store 410 based on an evaluation window to compare whether the accumulated length of time within the evaluation window achieves an observed duration threshold to satisfy an observed duration condition. The duration engine 408 may work in conjunction with the monitor engine 402 and the rule engine 404 to increase a time variable, such as an observed duration, as the set of status data satisfies the scale rule, and decrease the time variable as the evaluation window changes. The scale engine 406 may maintain the cloud unit if the accumulated length of time satisfies an observed duration condition.
The duration engine 408 may be configured to send a duration notification to the scale engine 406 and/or the rule engine 404. For example, the scale engine 406 may wait to perform a scale action until both a scale notification and a duration notification are received. For another example, the rule engine 404 may wait to send a scale notification until a duration notification is received.
FIG. 5 depicts that the example system 500 for maintaining a service on a cloud network may be implemented on a memory resource 530 operatively coupled to a processor resource 532. The processor resource 532 may be operatively coupled to the data store 510. The data store 510 may be the same as the data store 310 of FIG. 3 and/or the data store 410 of FIG. 4 and the description associated with the data store 310 and/or the data store 410 may be applied to the data store 510, and vice versa, as appropriate.
Referring to FIG. 5, the memory resource 530 may contain a set of instructions that may be executable by the processor resource 532. The set of instructions may implement the system 500 when executed by the processor resource 532. The set of instructions stored on the memory resource 530 may be represented as a monitor module 502, a rule module 504, a scale module 506, and a duration module 508. The processor resource 532 may carry out the set of instructions to execute the monitor module 502, the rule module 504, the scale module 506, the duration module 508, and/or any appropriate operations between or otherwise associated with the modules of the system 500. For example, the processor resource 532 may carry out a set of instructions to maintain a set of status data associated with a resource utilization of the service, apply the set of status data to a scale rule, and maintain a cloud unit based on the scale rule. The monitor module 502 may represent program instructions that when executed function as the monitor engine 402 of FIG. 4. The rule module 504 may represent program instructions that when executed function as the rule engine 404 of FIG. 4. The scale module 506 may represent program instructions that when executed function as the scale engine 406 of FIG. 4. The duration module 508 may represent program instructions that when executed as the duration engine 408 of FIG. 4.
The processor resource 532 may be one or multiple CPUs capable of retrieving instructions from the memory resource 530 and executing those instructions. The processor resource 532 may process the instructions serially, concurrently, or in partial concurrence, unless described otherwise herein.
The memory resource 530 and the data store 510 may represent a medium to store data utilized by the system 500. The medium may be any non-transitory medium or combination of non-transitory mediums able to electronically store data and/or capable of storing the modules of the system 500 and/or data used by the system 500. A storage medium, such as memory resource 530, may be distinct from a transmission medium, such as a signal. The medium may be machine-readable, such as computer-readable. The data of the data store 510 may include representations of resource utilization, QoS, workload volume, cost, and/or other data or information mentioned herein, such as a duration values and/or time values.
In the discussion herein, engines 402, 404, 406, and 408 and the module 502, 504, 506, and 508 have been described as combinations of hardware and programming. Such components may be implemented in a number of fashions. Looking at FIG. 5, the programming may be processor executable instructions stored on the memory resource 530, which is a tangible, non-transitory computer readable storage medium, and the hardware may include processor resource 532 for executing those instructions. The processor resource 532, for example, may include one or multiple processors. Such multiple processors may be integrated in a single device or distributed across devices. For example in reference to FIG. 6, the processor resource 532 may be distributed across any combination of server devices 692 and client devices 694. The memory resource 530 may be said to store program instructions that when executed by processor resource 532 implements the system 500 in FIG. 5. The memory resource 530 may be integrated in the same device as processor resource 532 or it may be separate but accessible to that device and processor resource 532. The memory resource 530 may be distributed across devices, such as server devices 692 and client devices 694 of FIG. 6. The memory resource 530 and the data store 510 may represent the same physical medium unless otherwise described herein.
In one example, the program instructions may be part of an installation package that when installed may be executed by processor resource 532 to implement the system 500. In this case, memory resource 530 may be a portable medium such as a CD, DVD, or flash drive or memory maintained by a server device, such as server device 692 of FIG. 6, from which the installation package may be downloaded and installed. In another example, the program instructions may be part of an application or applications already installed. Here, the memory resource 530 may include integrated memory such as a hard drive, solid state drive, or the like.
FIG. 6 depicts example environments in which various examples for maintaining a service on a cloud network may be implemented. The environment 690 is shown to include a cloud scale system 600 for maintaining a service on a cloud network. The system 600 described herein with respect to FIGS. 4 and 5, may represent generally any combination of hardware and programming configured to maintain a service on a cloud network. The system 600 may be integrated into a server device 692 or a client device 694. The system 600 may be distributed across server devices 692, client devices 694, or a combination of server devices 692 and client devices 694. The environment 690 may be a cloud computing environment. For example, any combination of the system 600, server device 692, and client devices 694 may be a virtual instance and/or may reside and/or execute on a cloud. The environment 690 may include a cloud computing environment, such as a cloud network 620. The cloud network 620 may be the same as the cloud network 320 of FIG. 3 and the description associated with the cloud network 320 may be applied to the cloud network 620, and vice versa, as appropriate.
In the example of FIG. 6, a server device 692 may access a client device 694. The server devices 692 may represent generally any computing devices configured to respond to a network request received from the client device 694. For example, a server device 692 may be a cloud unit of the cloud network 620 providing the service and the client device 694 may be a computing device configured to access the cloud network 620 and receive and/or communicate with the service. A server device 692 may include a web server, an application server, or a data server. The client devices 694 may represent generally any computing devices configured with a browser or other application to communicate such requests and receive and/or process the corresponding responses. A link 696 may represent generally one or any combination of a cable, wireless, fiber optic, or remote connections via a telecommunications link, an infrared link, a radio frequency link, or any other connectors of systems that provide electronic communication. The link 696 may include, at least in part, intranet, the Internet, or a combination of both. The link 696 may also include intermediate proxies, routers, switches, load balancers, and the like.
Referring to FIGS. 4-6, the engines 402, 404, 406, and 408 and/or the module 502, 504, 506, and 508 may be distributed across any combination of server devices, client devices, and storage mediums. The engines 402, 404, 406, and 408 and/or the module 502, 504, 506, and 508 may complete or assist completion of operations performed in describing another engine 402, 404, 406 and 408 and/or the module 502, 504, 506, and 508. The engines 402, 404, 406, and 408 and/or the module 502, 504, 506, and 508 may perform the example operations of operators 302, 304, 306, and 308 and/or the example methods described in connection with FIGS. 1 and 2.
Although the flow diagrams of FIGS. 1 and 2 illustrate specific orders of execution, the order of execution may differ from that which is illustrated. For example, the order of execution of the blocks may be scrambled relative to the order shown. Also, the blocks shown in succession may be executed concurrently or with partial concurrence. All such variations are within the scope of the present invention.
The present description has been shown and described with reference to the foregoing examples. It is understood, however, that other forms, details, and examples may be made without departing from the spirit and scope of the invention that is defined in the following claims.

Claims (20)

What is claimed is:
1. A method for maintaining a service on a cloud network comprising:
receiving a set of status data associated with the service, the set of status data including a set of utilization data related to a set of resources utilized by the service, a set of quality data related to a performance level of the service, and a set of workload data related to a workload volume of the service;
performing a scale action on the cloud network based on a scale rule applied to the set of status data, the scale rule including a set of conditions, the set of conditions including a utilization condition associated with the set of resources utilized by the service, a quality condition associated with the performance level of the service, a workload condition associated with the workload volume of the service, and a budget condition associated with a cost of the service; and
applying a duration rule to the set of status data, wherein the duration rule comprises an observation duration for the set of status data to satisfy the scale rule, wherein the observation duration comprises an accumulated length of time that the set of status data satisfies the scale rule within an evaluation window, wherein the evaluation window comprises a time frame to evaluate the observation duration.
2. The method of claim 1, wherein the scale action includes at least one of:
add a first cloud unit to the service, wherein the first cloud unit is at least one of a resource, a server, a server group, and a cloud; and
remove a second cloud unit from the service, wherein the second cloud unit is at least one of the resource, the server, the server group, and the cloud.
3. The method of claim 1, wherein the set of workload data includes the rate of workload requests per second and the workload condition includes a range of rates that satisfy the scale rule.
4. The method of claim 1, wherein at least one of the set of conditions is associated with a plurality of servers and includes a number of servers of the plurality of servers to satisfy the one of the set of conditions.
5. The method of claim 1, wherein the scale rule is associated with a plurality of servers of a cloud unit in the cloud network that are grouped based on a characteristic.
6. The method of claim 1, wherein the set of status data is aggregated from a plurality of cloud units and the set of conditions is based on a ratio of a portion of the plurality of cloud units that satisfy the set of conditions.
7. The method of claim 1, wherein the scale action comprises adding an entire cloud service.
8. A system for maintaining a service on a cloud network comprising:
a processor; and
a non-transitory computer readable storage medium having a set of instructions executable by the processor, the set of instructions comprising:
instructions to
maintain a set of status data associated with a resource utilization of the service, a quality of the service, and a workload volume of the service;
instructions to
apply a scale rule to the set of status data, the scale rule including a set of conditions, the set of conditions including a resource utilization condition to compare to the resource utilization of the service, a quality condition to compare to the quality of the service, a workload condition to compare to the workload volume of the service, and a budget condition to compare to a cost of the service;
instructions to
maintain a cloud unit based on the scale rule and the set of status data, wherein the cloud unit is at least one of a resource, a server, a server group, and a cloud; and
instructions to apply a duration rule to the set of status data, wherein the duration rule comprises an observation duration condition for the set of status data to satisfy the scale rule, wherein the observation duration condition comprises an accumulated length of time that the set of status data satisfies the scale rule within an evaluation window, wherein the evaluation window comprises a time frame to evaluate the observation duration condition.
9. The system of claim 8, wherein the the instructions to maintain a cloud unit based on the scale rule comprises at least one of:
instructions to add the cloud unit to the service; and
instructions to remove the cloud unit from the service.
10. The system of claim 8, wherein the instructions to maintain the cloud unit comprises instructions to maintain the cloud unit if the accumulated length of time satisfies the observation duration condition.
11. The system of claim 8, wherein at least one of the resource utilization condition, the quality condition, the workload condition, the budget condition, and the observed duration condition, is maintained dynamically based on a machine-learning method.
12. The system of claim 8, wherein the set of status data includes the cost of the service and the budget condition associated with the cost of the service, the cost of the service including the cloud unit cost associated with the one of the resource, the server, the server group, and the cloud.
13. The system of claim 8, wherein the scale rule is associated with a plurality of servers of a cloud unit in the cloud network that are grouped based on a characteristic.
14. The system of claim 8, wherein the set of status data is aggregated from a plurality of cloud units and the set of conditions is based on a ratio of a portion of the plurality of cloud units that satisfy the set of conditions.
15. A non-transitory computer readable storage medium having instructions stored thereon executable by a processor, the instructions including:
instructions to maintain a set of status data associated with a resource utilization of the service, a quality of the service, and a workload volume of the service;
instructions to apply the set of status data to a scale rule, the scale rule including a set of conditions, the set of conditions including a resource utilization condition to compare to the resource utilization of the service, a quality condition to compare to the quality of the service, a workload condition to compare to the workload volume of the service and a budget condition to compare to a cost of the service;
instructions to maintain a cloud unit based on the scale rule, wherein the cloud unit is at least one of a resource, a server, a server group, and a cloud; and
instructions to apply a duration rule to the set of status data, wherein the duration rule comprises an observation duration for the set of status data to satisfy the scale rule, wherein the observation duration comprises an accumulated length of time that the set of status data satisfies the scale rule within an evaluation window, wherein the evaluation window comprises a time frame to evaluate the observation duration.
16. The non-transitory computer readable storage medium of claim 15, wherein the instructions to maintain a cloud unit based on the scale rule comprises at least one of:
instructions to add the cloud unit to the service; and
instructions to remove the cloud unit from the service.
17. The non-transitory computer readable storage medium of claim 15, further comprising:
instructions to receive, based on a domain specific language, the set of conditions associated with a cloud unit group, the cloud unit group including a plurality of cloud units utilized by the service, each of the plurality of cloud units having a relationship to each of the plurality of cloud units.
18. The non-transitory computer readable storage medium of claim 17, wherein the the instructions to maintain the set of status data associated with a resource utilization of the service comprises instructions to aggregate the set of status data of the cloud unit group, the set of conditions based on a ratio of a portion of the plurality of the cloud units that satisfy the set of conditions to the plurality of cloud units.
19. The non-transitory computer readable storage medium of claim 15, wherein the scale rule is associated with a plurality of servers of a cloud unit in the cloud network that are grouped based on a characteristic.
20. The non-transitory computer readable storage medium of claim 15, wherein the set of status data is aggregated from a plurality of cloud units and the set of conditions is based on a ratio of a portion of the plurality of cloud units that satisfy the set of conditions.
US14/909,584 2013-08-30 2013-08-30 Maintain a service on a cloud network based on a scale rule Active 2034-03-07 US10057139B2 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2013/057738 WO2015030828A1 (en) 2013-08-30 2013-08-30 Maintain a service on a cloud network based on a scale rule

Publications (2)

Publication Number Publication Date
US20160191343A1 US20160191343A1 (en) 2016-06-30
US10057139B2 true US10057139B2 (en) 2018-08-21

Family

ID=52587170

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/909,584 Active 2034-03-07 US10057139B2 (en) 2013-08-30 2013-08-30 Maintain a service on a cloud network based on a scale rule

Country Status (4)

Country Link
US (1) US10057139B2 (en)
EP (1) EP3039560A1 (en)
CN (1) CN105493066B (en)
WO (1) WO2015030828A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11416296B2 (en) 2019-11-26 2022-08-16 International Business Machines Corporation Selecting an optimal combination of cloud resources within budget constraints
US11531536B2 (en) 2019-11-20 2022-12-20 Red Hat, Inc. Analyzing performance impacts of source code changes
US11635995B2 (en) 2019-07-16 2023-04-25 Cisco Technology, Inc. Systems and methods for orchestrating microservice containers interconnected via a service mesh in a multi-cloud environment based on a reinforcement learning policy
US20230231904A1 (en) * 2022-01-14 2023-07-20 Dell Products, L.P. Cloud provisioning readiness verification

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10430219B2 (en) * 2014-06-06 2019-10-01 Yokogawa Electric Corporation Configuring virtual machines in a cloud computing platform
US10817544B2 (en) * 2015-04-20 2020-10-27 Splunk Inc. Scaling available storage based on counting generated events
US10282455B2 (en) 2015-04-20 2019-05-07 Splunk Inc. Display of data ingestion information based on counting generated events
US10635423B2 (en) * 2015-06-30 2020-04-28 Vmware, Inc. Methods and apparatus for software lifecycle management of a virtual computing environment
CN106453457B (en) * 2015-08-10 2019-12-10 微软技术许可有限责任公司 Multi-priority service instance allocation within a cloud computing platform
US10616370B2 (en) 2015-08-21 2020-04-07 Hewlett Packard Enterprise Development Lp Adjusting cloud-based execution environment by neural network
US10432644B2 (en) * 2015-09-28 2019-10-01 Box, Inc. Access control system for enterprise cloud storage
US11082499B2 (en) * 2015-10-19 2021-08-03 Citrix Systems, Inc. Browser server session transfer
US10069869B2 (en) * 2016-05-17 2018-09-04 Amazon Technologies, Inc. Versatile autoscaling
CN106169096B (en) * 2016-06-24 2018-07-24 山西大学 A kind of appraisal procedure of machine learning system learning performance
US10230582B2 (en) * 2016-08-29 2019-03-12 International Business Machines Corporation Identifying resources for purging in a cloud based on inter-dependency graph analysis
US10178045B2 (en) * 2016-09-07 2019-01-08 Sap Se Dynamic discovery and management of microservices for multi-cluster computing platforms
US10355945B2 (en) * 2016-09-21 2019-07-16 International Business Machines Corporation Service level management of a workload defined environment
US10572310B2 (en) 2016-09-21 2020-02-25 International Business Machines Corporation Deploying and utilizing a software library and corresponding field programmable device binary
US10417012B2 (en) 2016-09-21 2019-09-17 International Business Machines Corporation Reprogramming a field programmable device on-demand
US10599479B2 (en) 2016-09-21 2020-03-24 International Business Machines Corporation Resource sharing management of a field programmable device
US10409642B1 (en) 2016-11-22 2019-09-10 Amazon Technologies, Inc. Customer resource monitoring for versatile scaling service scaling policy recommendations
US10318287B2 (en) * 2016-12-21 2019-06-11 Hewlett Packard Enterprise Development Lp Deploying documents to a server in a specific environment
US10936215B2 (en) * 2018-04-30 2021-03-02 EMC IP Holding Company LLC Automated data quality servicing framework for efficient utilization of information technology resources
US10901721B2 (en) 2018-09-20 2021-01-26 Vmware, Inc. Methods and apparatus for version aliasing mechanisms and cumulative upgrades for software lifecycle management
US10817345B2 (en) * 2019-01-25 2020-10-27 Virtustream Ip Holding Company Llc Distributed ledger for monitoring quality of services provided by cloud service providers
CN112929293A (en) * 2019-12-05 2021-06-08 金色熊猫有限公司 Task execution method, device, equipment, platform and medium
US12041122B2 (en) * 2021-03-04 2024-07-16 Electronics And Telecommunications Research Institute Apparatus and method for autoscaling service shared in cloud
CN113515382B (en) * 2021-07-22 2024-04-09 中移(杭州)信息技术有限公司 Cloud resource allocation method and device, electronic equipment and storage medium

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1728167A (en) 2004-07-29 2006-02-01 国际商业机器公司 Method and system for allocating resources in a utility services environment
CN1751294A (en) 2003-03-10 2006-03-22 国际商业机器公司 Methods and apparatus for managing computing deployment in presence of variable workload
US20100050172A1 (en) 2008-08-22 2010-02-25 James Michael Ferris Methods and systems for optimizing resource usage for cloud-based networks
US20100185961A1 (en) 2009-01-20 2010-07-22 Microsoft Corporation Flexible visualization for services
US20110153824A1 (en) 2009-12-17 2011-06-23 Chikando Eric N Data Processing Workload Administration In A Cloud Computing Environment
US20110258621A1 (en) 2010-04-14 2011-10-20 International Business Machines Corporation Autonomic Scaling Of Virtual Machines In A Cloud Computing Environment
US20120204187A1 (en) 2011-02-08 2012-08-09 International Business Machines Corporation Hybrid Cloud Workload Management
US20120233328A1 (en) 2011-03-07 2012-09-13 Gravitant, Inc Accurately predicting capacity requirements for information technology resources in physical, virtual and hybrid cloud environments
EP2568383A1 (en) 2011-09-07 2013-03-13 Accenture Global Services Limited Cloud service monitoring system
US20130117157A1 (en) 2011-11-09 2013-05-09 Gravitant, Inc. Optimally sourcing services in hybrid cloud environments
US20130179895A1 (en) * 2012-01-09 2013-07-11 Microsoft Corporation Paas hierarchial scheduling and auto-scaling
US20130179894A1 (en) * 2012-01-09 2013-07-11 Microsoft Corporation Platform as a service job scheduling
US20130179881A1 (en) * 2012-01-09 2013-07-11 Microsoft Corporation Decoupling paas resources, jobs, and scheduling
US20140244311A1 (en) * 2013-02-25 2014-08-28 International Business Machines Corporation Protecting against data loss in a networked computing environment
US20140282217A1 (en) * 2013-03-12 2014-09-18 Oracle International Corporation Window grouping and management across applications and devices
US9183189B1 (en) * 2012-02-01 2015-11-10 Amazon Technologies, Inc. Network site hosting in a managed environment
US9363154B2 (en) * 2012-09-26 2016-06-07 International Business Machines Corporaion Prediction-based provisioning planning for cloud environments

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1751294A (en) 2003-03-10 2006-03-22 国际商业机器公司 Methods and apparatus for managing computing deployment in presence of variable workload
CN1728167A (en) 2004-07-29 2006-02-01 国际商业机器公司 Method and system for allocating resources in a utility services environment
US20100050172A1 (en) 2008-08-22 2010-02-25 James Michael Ferris Methods and systems for optimizing resource usage for cloud-based networks
US20100185961A1 (en) 2009-01-20 2010-07-22 Microsoft Corporation Flexible visualization for services
US20110153824A1 (en) 2009-12-17 2011-06-23 Chikando Eric N Data Processing Workload Administration In A Cloud Computing Environment
US20110258621A1 (en) 2010-04-14 2011-10-20 International Business Machines Corporation Autonomic Scaling Of Virtual Machines In A Cloud Computing Environment
US20120204187A1 (en) 2011-02-08 2012-08-09 International Business Machines Corporation Hybrid Cloud Workload Management
US20120233328A1 (en) 2011-03-07 2012-09-13 Gravitant, Inc Accurately predicting capacity requirements for information technology resources in physical, virtual and hybrid cloud environments
EP2568383A1 (en) 2011-09-07 2013-03-13 Accenture Global Services Limited Cloud service monitoring system
US20130117157A1 (en) 2011-11-09 2013-05-09 Gravitant, Inc. Optimally sourcing services in hybrid cloud environments
US20130179895A1 (en) * 2012-01-09 2013-07-11 Microsoft Corporation Paas hierarchial scheduling and auto-scaling
US20130179894A1 (en) * 2012-01-09 2013-07-11 Microsoft Corporation Platform as a service job scheduling
US20130179881A1 (en) * 2012-01-09 2013-07-11 Microsoft Corporation Decoupling paas resources, jobs, and scheduling
US9183189B1 (en) * 2012-02-01 2015-11-10 Amazon Technologies, Inc. Network site hosting in a managed environment
US9363154B2 (en) * 2012-09-26 2016-06-07 International Business Machines Corporaion Prediction-based provisioning planning for cloud environments
US20140244311A1 (en) * 2013-02-25 2014-08-28 International Business Machines Corporation Protecting against data loss in a networked computing environment
US20140282217A1 (en) * 2013-03-12 2014-09-18 Oracle International Corporation Window grouping and management across applications and devices

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
Bunch, C. et al., Neptune: A Domain Specific Language for Deploying HPC Software on Cloud Platforms, Jun. 8, 2011, 10 pages, <http://www.cs.ucsb.edu/˜ckrintz/papers/sciencecloud11.pdf>.
Cloud Taxonomy, Gravitant, OpenCrod, 2010, 1 page, <http://cloudtaxonomy.opencrowd.com/taxonomy/infrastructure-as-a-service/cloud-broker/gravitant/ >.
Duncan McGreggor, "Autoscaling API for Heat," Mar. 20, 2013, pp. 1-3 [online], Canonical Ltd., Retrieved from the Internet on Nov. 15, 2017 at URL: <https://web.archive.org/web/20130527194022/https:/blueprints.launchpad.net/heat/+spec/heat-autoscaling >.
Han, R. et al., Lightweight Resource Scaling for Cloud Applications, Jun. 18, 2012, 9 pages, <http://www.doc.ic.ac.uk/˜rh1910/papers/Lightweight_Resource_Scaling_for_Cloud_Applications.pdf>.
Hewlett-Packard Development Company, L.P., "Dynamic Workload Management for Private Cloud," Business White Paper, Apr. 2012, pp. 1-8.
International Search Report and Written Opinion, PCT Patent Application No. PCT/US2013/057738, Jun. 2, 2014, 12 pages.
Trieu C. Chieu et al., Dynamic Scaling of Web Applications in a Virtualized Cloud Computing Environment. IEEE International Conference on e-Business Engineering, 2009, pp. 28.

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11635995B2 (en) 2019-07-16 2023-04-25 Cisco Technology, Inc. Systems and methods for orchestrating microservice containers interconnected via a service mesh in a multi-cloud environment based on a reinforcement learning policy
US11531536B2 (en) 2019-11-20 2022-12-20 Red Hat, Inc. Analyzing performance impacts of source code changes
US11416296B2 (en) 2019-11-26 2022-08-16 International Business Machines Corporation Selecting an optimal combination of cloud resources within budget constraints
US20230231904A1 (en) * 2022-01-14 2023-07-20 Dell Products, L.P. Cloud provisioning readiness verification
US11979456B2 (en) * 2022-01-14 2024-05-07 Dell Products, L.P. Cloud provisioning readiness verification

Also Published As

Publication number Publication date
WO2015030828A1 (en) 2015-03-05
US20160191343A1 (en) 2016-06-30
CN105493066B (en) 2019-02-05
CN105493066A (en) 2016-04-13
EP3039560A1 (en) 2016-07-06

Similar Documents

Publication Publication Date Title
US10057139B2 (en) Maintain a service on a cloud network based on a scale rule
Rahman et al. Load balancer as a service in cloud computing
US10594562B1 (en) Intelligent autoscale of services
US10616370B2 (en) Adjusting cloud-based execution environment by neural network
US8880679B2 (en) Techniques for web server management
US10387208B2 (en) Distributed cloud computing elasticity
US20140122698A1 (en) Cdn traffic management in the cloud
Al-Haidari et al. Evaluation of the impact of EDoS attacks against cloud computing services
US20220224694A1 (en) Resource appropriation in a multi-tenant environment using risk and value modeling systems and methods
US11265389B2 (en) System and method for computation of user experience score for virtual apps and desktop users
US20160048407A1 (en) Flow migration between virtual network appliances in a cloud computing network
US20210117280A1 (en) Method, device, and computer program product for scheduling backup jobs
Han et al. Analysing virtual machine usage in cloud computing
Singh Study of response time in cloud computing
Jodayree et al. A predictive workload balancing algorithm in cloud services
US8725868B2 (en) Interactive service management
Rankothge et al. On the scaling of virtualized network functions
Shifrin et al. Optimal control of VNF deployment and scheduling
Xiong et al. Ensuring cloud service guarantees via service level agreement (SLA)-based resource allocation
CN115269193A (en) Method and device for realizing distributed load balance in automatic test
US10771392B1 (en) Increasing efficiency to rate limiting
US9537742B2 (en) Automatic adjustment of application launch endpoints
Oral et al. Supporting performance isolation in software as a service systems with rich clients
CN110798496A (en) CDN scheduling system, method and device
Demirci et al. Optimal placement of virtual security functions to minimize energy consumption

Legal Events

Date Code Title Description
AS Assignment

Owner name: HEWLETT PACKARD ENTERPRISE DEVELOPMENT LP, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P.;REEL/FRAME:038536/0001

Effective date: 20151027

AS Assignment

Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DONG, JING;GREEN, JOHN M;GALLAGHER, PHYLLIS;AND OTHERS;SIGNING DATES FROM 20130829 TO 20130912;REEL/FRAME:040900/0355

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4