WO2015047302A1 - Processing a hybrid flow associated with a service class - Google Patents

Processing a hybrid flow associated with a service class Download PDF

Info

Publication number
WO2015047302A1
WO2015047302A1 PCT/US2013/062201 US2013062201W WO2015047302A1 WO 2015047302 A1 WO2015047302 A1 WO 2015047302A1 US 2013062201 W US2013062201 W US 2013062201W WO 2015047302 A1 WO2015047302 A1 WO 2015047302A1
Authority
WO
WIPO (PCT)
Prior art keywords
execution
service class
flow
service
sub
Prior art date
Application number
PCT/US2013/062201
Other languages
French (fr)
Inventor
Alkiviadis Simitsis
William Kevin Wilkinson
Matteo LISSANDRINI
Original Assignee
Hewlett-Packard Development Company, L.P.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hewlett-Packard Development Company, L.P. filed Critical Hewlett-Packard Development Company, L.P.
Priority to EP13894923.5A priority Critical patent/EP3049959A1/en
Priority to PCT/US2013/062201 priority patent/WO2015047302A1/en
Priority to US14/917,059 priority patent/US20160203026A1/en
Priority to CN201380079890.3A priority patent/CN105580001A/en
Publication of WO2015047302A1 publication Critical patent/WO2015047302A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/25Integrating or interfacing systems involving database management systems
    • G06F16/254Extract, transform and load [ETL] procedures, e.g. ETL data flows in data warehouses

Definitions

  • PCT/US2013/035080 filed on April 3, 2013 and entitled "Modifying a Flow of Operations to he Executed in a Plurality of Execution Environments", which is hereby incorporated by reference.
  • Operations on data can be performed by different types of executio environments.
  • one execution environment can be a database management system (DBMS) environment, in which data is stored in relational fables and subject to database-based operations.
  • DBMS database management system
  • a execution environment can include a MapReduce environment, which performs operations using map tasks and reduce tasks.
  • MapReduce environment which performs operations using map tasks and reduce tasks.
  • FiG, 1 illustrates a method of processing a hybrid flow associated with a service class, according to an example.
  • FSG. 2 illustrates a method of managing resources of a plurality of execution environments, according to an example.
  • [COOS] FSG. 3 illustrates a method of validating a sub-flow, according to an example.
  • FIG, 4 illustrates a hybrid flow management system that supports management of hybrid flows associated with multiple service classes, according to an example.
  • FSG. 5 illustrates a computer- readable medium for processing a hybrid flow associated with a service class, according to an example.
  • a single analytic computation can be modeled as a directed graph in which starting nodes are data sources, ending nodes are data targets, intermediate nodes are data operations, and arcs represent data flow.
  • Such a computation can be referred to as an analytic data flow (or simply a "flow"), in other examples, a flow can have a different representation (other than a directed graph).
  • An analytic data f ow that utilizes more than one data repository or execution environment is referred to as a hybrid analytic data flow (or simply a "hybrid flow").
  • a collection of analytic data flows that is managed as a unit for some objective (e.g., the flows should complete before a deadline, or an average response time of individual fiows should not exceed a threshold, etc) is referred to as a workload.
  • a service class workload manager may be used to manage a collection of flows associated with a respective service class.
  • a hybrid flow can include collections of operations that are performed in different execution environments.
  • a collection of operations of the hybrid flow that is performed in a respective execution environment can be referred to as a sub- flow.
  • a single hybrid flow may include three sub-flows directed to a first execution environment and two sub-flows directed to a second execution environment,
  • Examples of different types of execution environments include at least some of the following: database management system (DBMS) environment,
  • Each execution environment can include an execution engine and a respective storage repository of data.
  • An execution engine can include one or multiple execution stages for applying respective operators on data, where the operators can transform or perform some other action with respect to data.
  • a storage repository refers to one or multiple collections of data.
  • An execution environmeni can be available in a public cloud or pubiic network, in which case the execution environment can be referred to as a public cloud execution environment.
  • an execution environment that is available in a private network can be referred to as a private execution
  • a DBMS environment stores data in relational tables and applies database operators (e.g. join operators, update operators, merge operators, and so forth) on data in the relational tables.
  • a MapReduce environment includes map tasks and reduce tasks that can apply a map function and a reduce function, respectively.
  • a map task processes input data to produce intermediate results, based on the respective map function that defines the processing to be performed by the map task.
  • a reduce task takes as input partitions of the intermediate results from the map task to produce an output, based on the corresponding reduce function that defines the processing to be performed by the reduce tasks.
  • Another example execution environment includes an ETL environment, which extracts data from a source (or multiple sources), transforms the data, and loads the transformed data Into a destination.
  • a mathematical/statistical analysis environment may perform mathematical computation on arrays or vectors of data.
  • An event stream processing environment may perform computations on event streams such as sensor data,
  • a service class can be associated with a target performance objective, which identifies one or multiple goals that are to be met by the execution of a workload associated with the service class,
  • a performance objective can also be referred to as a service level objective.
  • An example performance objective relates to an execution time (e.g., a time duration for executing the workload or a target deadline by which the workload is to be completed).
  • performance objective is a resource usage objective, which can specify that usage of resources, such as computing resources, storage resources, or communication resources, should not exceed a target level.
  • resources such as computing resources, storage resources, or communication resources
  • other performance objectives can be employed,
  • An execution plan for a hybrid flow specifies where (i.e., target execution environments) the sub-flows of the hybrid flow are to execute, and can specify other details associated with execution of the sub-flows (such as order),
  • a single hybrid flow can have many alternative execution plans due to overlap in
  • a management system can deplo the sub-flows of the hybrid flow in the target execution environments, and can orchestrate the execution of the hybrid flow.
  • the state of a computing infrastructure of at least one execution environment may have changed between the time the execution plan was produced and the time the execution plan is executed. For example, the execution environment may have become overloaded (such that there is contention for resources) or the computing infrastructure may have experienced a fault.
  • the hybrid flow is associated with a performance objective that has to be met. In some cases, penalties may be specified for not meeting performance objectives.
  • the management system should ensure that target performance objectives are achieved,
  • a workload manager may exist within an individual execution environment, and this workload manager can adjust a priority of a task, the number of concurrently running tasks, and so forth, to increase the likelihood that a workload within the individual execution environment meets a respective target objective.
  • a workload manager within a single execution environment can only make decisions optimal for that environment.
  • Such a workload manager has no knowledge of the state of other execution environments. So, for a workload that has flows to be executed across a number of different types of execution environments, workload management becomes more challenging.
  • a hybrid flow management system to apply a policy-based hybrid flow management for hybrid flows in a plurality of service classes.
  • the system may include multiple service class workload managers (e.g., one for each service class) to facilitate
  • a hybrid flow associated with one of the plurality of service classes may be received.
  • the hybrid flow may include multiple sub-flows directed to multiple execution environments.
  • a schedule may be generated to run the sub-flows on the multiple execution environments based on criteria.
  • the criteria may include one or more objectives associated with the hybrid flow's 83598076
  • One or more sub-flows may be dispatched for execution in the execution environments. Execution of the dispatched sub-flows may be monitored to generate statistics. A remedial action may be taken according to a policy associated with the service class if the statistics differ from the objective by a threshold. Accordingly, a hybrid flow associated with a service class may be managed to achieve performance objectives. Furthermore, multiple service classes may be supported, such that a hybrid flow associated with any of the service classes may be appropriately managed to achieve performance objectives associated with the respective service class. Additional examples, advantages, features, modifications and the like are described below with reference to the drawings.
  • FIG. 1 illustrates a method for processing a hybrid flow associated with a service class, according to an example.
  • Method 100 may be performed by a computing device, system, or computer, such as hybrid flow management system 400 or computing system 500.
  • Computer-readable instructions for implementing method 100 may be stored on a computer readable storage medium. These instructions as stored on the medium are referred to herein as "modules" and may be executed by a computer.
  • System 400 may include and/or be
  • the computers may be server computers, workstation computers, desktop computers, laptops, mobile devices, or the like, and may be part of a distributed system.
  • the computers may include one or more controllers and one or more machine-readable storage media.
  • controller may include a processor and a memory for implementing machine readable instructions.
  • the processor may include at least one central processing unit (CPU), at least one semiconductor-based microprocessor, at least 83598076
  • DSP digital signal processor
  • the processor can include single or multiple cores on a chip, multiple cores across multiple chips, multiple cores across multiple devices, or combinations thereof.
  • the processor may fetch, decode, and execute instructions from memory to perform various functions.
  • the processor may include at least one integrated circuit (IC), other control logic, other electronic circuits, or combinations thereof that include a number of electronic components for performing various tasks or functions,
  • the controller may include memory, such as a machine-readable storage medium.
  • the machine-readable storage medium may be any electronic, magnetic, optical, or other physical storage device that contains or stores executable instructions.
  • the machine-readable storage medium may comprise, for example, various Random Access Memory (RAM), Read Only Memory (ROM), flash memory, and combinations thereof.
  • RAM Random Access Memory
  • ROM Read Only Memory
  • flash memory and combinations thereof.
  • the machine-readabie medium may include a Non-Volatile Random Access Memory (NVRAM), an NVRAM, an Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel® 845555B Intel®
  • system 400 may include one or more machine-readable storage media separate from the one or more controllers.
  • Hybrid flow management system 400 may include a number of
  • system 400 may include an interface 410 and optimizer 420.
  • the interface 410 can receive a flow definition for creating or modifying a flow.
  • the interface 410 can present a graphical user interface (GUI) to allow for users to interactively create or modify a flow.
  • GUI graphical user interface
  • a flow can be written in a declarative language and imported through the interface.
  • Th flow that is created or modified using the interface 410 can be a hybrid flow.
  • optimizer 420 generates mu!tip!e candidate execution plans for each flow.
  • the optimizer 420 is able to consider alternative candidate execution plans for a given flow, and can estimate the respective costs of the candidate execution plans.
  • An execution plan (which can be an optimal execution plan) from among the candidate execution plans can be selected for execution, where an optimal execution plan can refer to an execution plan that is associated with a lowest cost or that satisfies some other criterion,
  • System 400 may also include an executor 430 and may be connected to execution environments 490 via a network.
  • the network may be any type of communications network, including, but not limited to, wire-based networks (e.g., cable), wireless networks (e.g., cellular, satellite), cellular telecommunications networks), and IP-based telecommunications network(s) (e.g., Voice over Internet Protocol networks).
  • the network may also include traditional landline or a public switched telephone network (PSTN), or combinations of the foregoing.
  • PSTN public switched telephone network
  • the optimizer 420 can provide the selected execution plan to the executor 430,
  • the selected execution plan is itself a flow, and is a hybrid flow if it includes sub-flows directed to two or more of execution environments 490, Executor 430 will now be described in more detail with reference to methods 100-300,
  • Method 100 may begin at 110, where a hybrid flow is received.
  • executor 430 may receive a selected execution plan from the optimizer 420.
  • the hybrid flow may be associated with one of a plurality of service classes supported by the system 400.
  • Service classes are an approach for providing differentiated service for requests.
  • service class may be determined and/or assigned to the hybrid flow in various ways. For example, a user may have signed in to interface 410 to create a flow. The user may be associated with a particular service class by user id by virtue of the user's role in the enterprise. For example, if the user id corresponds to a CEO of th enterprise, any generated flow may be assigned to a service class a 83598076
  • a user id corresponding to a marketing department or engineering department may be assigned to a service class having a lower priority.
  • the service class may also be dictated by the size of the flow, the type of flow, the number of execution environments required by the flow, etc.
  • a service class may be determined by the geographical location and time of the request origination. Various other methods of determining and/or assigning a service class to a flow may be used.
  • a schedule may be generated to run the sub-flows of the hybrid flow on the execution environments.
  • scheduler 450 may be configured to generate the schedule for execution of the sub-flows.
  • scheduler 450 may generate a separate schedule for execution of hybrid flows associated with different service classes, such that each service class has its own schedule for execution of flows associated with that service class.
  • Various scheduling techniques may be used, such as first fit or bi packing. Additionally, the schedule may be generated based on various criteria. For example, the schedule may be generated based on one o more objectives associated with the service class, allocation of resources of the multiple execution environments to the service class, resource availability, and parameters and/or constraints associated with the hybrid flow.
  • the objectives may be any of various specified service level objectives for the service class (e.g., 90% of flows must complete in fewer than one minute, the entire set of flows must complete by a specified time, etc).
  • the objectives for a given service class may be stored in computer-readable medium 486 of the corresponding service class workload manager 480 as objectives 487.
  • the scheduler 450 may thus operate under the constraint that the generated schedule meets objectives 487.
  • each service class may be allocated a set of resources from each execution environment. For instance, a service class may be allocated a number 83598076
  • the ailocaiion of resources may be specific to each execution environment 490, For instance, a service class may be ai located 50% of the processing cores of E but only 25% of the processing cores of E2- [0034]
  • Resource manager 440 may be used to manage the allocatio of resources to the various service classes by assigning particular allocations to the service class workload managers 480. In one example, the allocation of resources among all the service classes should not exceed the total number of resources in the execution environments. In another example, the initial ailocaiion of resources may exceed the total number of resources, but logic in th resource manager 440 may ensure that as the total number in use of any one resource is being
  • the allocated amount of that resource to each service class workload manager 480 is modified so that there isn't resource contention
  • Resource manager 440 may perform method 200.
  • resource manager 440 may monitor resource availability. This may be done by querying the execution environments 490 for their status via an application programming interface.
  • resource manager 440 may manage the allocation of resources to the service class workload managers 480. As discussed above, the allocation of resources may be dictated by the service classes. For example, each service class may have a set of resource requirements which are communicated to the resource manager 440. As resource availability changes over time (such as due to additional resources being made available in an execution environment, resources failing in an execution environment, etc.), resource manager 440 ma modify the resource allocation. Additionally, as discussed later, resource manager 440 may modify the resource allocation in response to requests from the service class workload managers 480. Resource manager 440 may also maintain other parameters and statistics associated with the execution environments, such as a multi-programming level allowed for each environment.
  • the criteria by which the schedule is generated may include parameters and/or constraints associated with the hybrid 83598076
  • the hybrid flow may be associated with an arrival time (e.g., the time when the flow first arrived at optimizer 420 or executor 430) and a requested execution time (e.g., a requested time input by a user via interface 410).
  • arrival time e.g., the time when the flow first arrived at optimizer 420 or executor 430
  • requested execution time e.g., a requested time input by a user via interface 410.
  • the parameters may also relate to the number of sub-flows in the hybrid flow, the minimum amount of time needed to fully execute the hybrid flow, the number and type of execution environments needed for execution of the flow, etc. Other parameters and constraints may be considered as well.
  • Scheduler 450 may take all of these criteria into consideration to generate a schedule for execution of the hybrid flow. Furthermore, the scheduler 450 may- generate this schedule for a plurality of hybrid flows associated with the given service class. For example, there may be five independent hybrid flows associated with a given service class that are being managed at the same time by system 400. The scheduler 450 may consider these criteria for all five hybrid flows to generate a schedule for execution of the sub-flows of these five hybrid flows on execution environments 490 such that objectives and constraints associated with the flows are met. Additionally, scheduler 450 may recompute the schedule if a new flow is received, if resources are reallocated, etc,
  • one or more of the sub-flows may be dispatched to the execution environments for execution based on the schedule.
  • dispatcher 470 may be configured to dispatch the sub-flows to execution environments 490.
  • Validator 460 may perform method 300.
  • the execution requirements of a sub-flow to be dispatched may be verified.
  • validator 460 may verify via resource monitor 440 that all resources required for execution of the sub-flow are available and appropriately allocated to the service class. It may also be verified that objectives 487 will be met, that all 83598076
  • an action may be taken. For example, in the event of failure, validator 460 may inform the service class workload manager 480 that the verification fa tied. Action module 484 of service class workload manager 480 may then take a remedial acfion(s) in accordance with one or more policies 488. Policies 488 may be specified for each service class to deaf with plan snvaiidation. There may be separate policies for the various circumstances of plan invalidation and the policies may result in different actions being taken.
  • the service class workload manager 480 may request additional resources from one or more of the execution environments 490. Such a request may entail removing resources allocated to another service class and reallocating those resources to the service class associated with the invalidated sub-flow.
  • the service class workload manager 480 may request instantiation of an additional execution environment.
  • the service class workload manager 480 may request that the given sub-flo be sent to a different execution environment for execution thereon.
  • the service class workload manager 480 may request that the sub-flow be decomposed into smaller fragments.
  • the service class workload manager 480 may request that scheduler 450 generate a new schedule, that optimizer 420 send a new execution plan appropriate to the current state of the computing environments (e.g., from a list of previously generated execution plans), or that optimizer 420 generate a ne execution plan altogether.
  • execution of the dispatched sub-flows may be monitored.
  • monitoring module 482of the service class workload manager 480 may monitor execution of the sub-flows associated with the service class.
  • Statistics 489 may be generated based on the monitoring, such as average execution time of the sub-flows, proportion of sub-flows executed, projected total execution time of the hybrid flow, etc. If the statistics differ from one or more 83598076
  • action module 484 may take an action according to one or more of policies 488, such as described above,
  • FIG. 5 illustrates a computing system for processing a hybrid flow associated with a service class, according to an example.
  • Computing system 500 may include and/or be implemented by one or more computers.
  • the computers may be server computers, workstation computers, desktop computers, laptops, mobile devices, or the like, and may be part of a distributed system.
  • the computers may include one or more controllers and one or more machine-readable storage media, as described with respect to hybrid flow management system 400, for example.
  • users of computing system 500 may interact with computing system 500 through one or more other computers, which may or may not be considered part of computing system 500.
  • a user may interact with system 500 via a computer application residing on system 500 or on another computer, such as a desktop computer, workstation computer, tablet computer, or the like.
  • the computer application can include a user interface (e.g., touch interface, mouse, keyboard, gesture input device).
  • Computing system 500 may perform methods 100-300, and variations thereof, and components 510-540 may be configured to perform various portions of methods 100-300, and variations thereof. Additionally, the functionality
  • components 510-540 may be part of a larger software platform, system, application, or the like.
  • these components may be part of a data analysis system.
  • Computers 510 may have access to database 540,
  • the database may include one or more computers, and may include one or more controllers and machine-readable storage mediums, as described herein.
  • the computer may be connected to the database via a network.
  • the network may be any type of communications network, including, but not limited to, wire-based networks (e.g., cable), wireless networks (e.g., cellular, satellite), cellular telecommunications 83598076
  • nefwork(s), and IP-based telecommunications network(s) e.g., Voice over internet Protocol networks
  • the network may also include traditional landline or a public switched telephone network (PSTN), or combinations of the foregoing.
  • PSTN public switched telephone network
  • Processor 520 may be at least one central processing unit (CPU), at least one semiconductor-based microprocessor, other hardware devices or processing elements suitable to retrieve and execute instructions stored in machine-readable storage medium 530, or combinations thereof.
  • Processor 520 can include single or multiple cores on a chip, multiple cores across multiple chips, multiple cores across multiple devices, or combinations thereof.
  • Processor 520 may fetch, decode, and execute instructions 532-538 among others, to implement various processing.
  • processor 520 may include at ieast one integrated circuit (JC), other control logic, other electronic circuits, or combinations thereof that include a number of electronic components for performing the functionality of instructions 532-538. Accordingly, processor 520 may be implemented across multiple processing units and instructions 532-538 may be implemented by different processing units in different areas of computer 510.
  • JC integrated circuit
  • Machine-readable storage medium 530 may be any electronic, magnetic, optical, or other physical storage device that contains or stores executable instructions.
  • the machine-readable storage medium may comprise, for example, various Random Access Memory (RAM), Read Only Memory (ROM), flash memory, and combinations thereof.
  • the machine-readabl medium may include a Non-Volatile Random Access Memory (NVRAM), an
  • machine-readable storage medium 530 can be computer-readable and non-transitory.
  • Machine-readable storage medium 530 may be encoded with a series of executable instructions for managing processing elements.
  • the instructions 532-538 when executed by processor 520 can cause processor 520 to perform processes, for example, methods 100-300, and/or variations and portions thereof.
  • service class instructions 532 may cause processor 520 to determine a service class associated with each of a plurality of hybrid flows.
  • Scheduling instructions 534 may cause processor 520 to generate a separate execution schedule for the hybrid flows in each service class based o parameters and constraints associated with each hybrid flow, objectives and resource allocation associated with the respective service class, and resource availability.
  • Monitoring instructions 538 may cause processor 510 to monitor the execution of each hybrid flow to generate statistics.
  • Action instructions 538 may cause processor 520 to take a remedial action if the statistics for a given hybrid flow differ from objectives associated with the given hybrid flow's service class by a threshold. The remedial action may be dictated by a policy associated with the given hybrid flow's service ciass.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Debugging And Monitoring (AREA)

Abstract

Described herein are techniques for processing hybrid flows. A hybrid flow may be associated with one of a plurality of service classes and may include sub-flows directed to multiple execution environments. An execution schedule may be generated based on an objective associated with the service class and resource allocation or availability. An action may be taken according to a policy associated with the service class if execution of the hybrid flow fails to meet the objective.

Description

83598076
PROCESSING A HYBRID FLOW ASSOCIATED WITH A SERVICE CLASS
REFERENCE TO RELATED APPLICATIONS
|0001] This application is related to International Patent Application No.
PCT/US2013/035080, filed on April 3, 2013 and entitled "Modifying a Flow of Operations to he Executed in a Plurality of Execution Environments", which is hereby incorporated by reference.
BACKGROUND
[0002] Operations on data can be performed by different types of executio environments. For example, one execution environment can be a database management system (DBMS) environment, in which data is stored in relational fables and subject to database-based operations. As another example, a execution environment can include a MapReduce environment, which performs operations using map tasks and reduce tasks. There can also be other types of execution environments,
BRIEF DESCRIPTION OF DRAWINGS
[0003] The following detailed description refers to the drawings, wherein:
[00043 FiG, 1 illustrates a method of processing a hybrid flow associated with a service class, according to an example. 83598076
[0005] FSG. 2 illustrates a method of managing resources of a plurality of execution environments, according to an example.
[COOS] FSG. 3 illustrates a method of validating a sub-flow, according to an example.
[00073 FIG, 4 illustrates a hybrid flow management system that supports management of hybrid flows associated with multiple service classes, according to an example.
[0008] FSG. 5 illustrates a computer- readable medium for processing a hybrid flow associated with a service class, according to an example.
DETAILED DESCRIPTION
[00093 Enterprises (e.g. business concerns, educational organizations, government agencies, etc.) can depend on reports and analyses (generall referred to as "computations") that integrate data from a diverse collection of data repositories and that operate on the data using a variety of execution
environments. In some examples, a single analytic computation can be modeled as a directed graph in which starting nodes are data sources, ending nodes are data targets, intermediate nodes are data operations, and arcs represent data flow. Such a computation can be referred to as an analytic data flow (or simply a "flow"), in other examples, a flow can have a different representation (other than a directed graph). An analytic data f ow that utilizes more than one data repository or execution environment is referred to as a hybrid analytic data flow (or simply a "hybrid flow"). A collection of analytic data flows that is managed as a unit for some objective (e.g., the flows should complete before a deadline, or an average response time of individual fiows should not exceed a threshold, etc) is referred to as a workload. A service class workload manager may be used to manage a collection of flows associated with a respective service class. 83598076
[0010] A hybrid flow can include collections of operations that are performed in different execution environments. A collection of operations of the hybrid flow that is performed in a respective execution environment can be referred to as a sub- flow. There may be multiple sub-flows directed to a single execution environment in a hybrid flow. For example, a single hybrid flow may include three sub-flows directed to a first execution environment and two sub-flows directed to a second execution environment,
[0011] Examples of different types of execution environments include at least some of the following: database management system (DBMS) environment,
apRe uce environment, an extract, transform, and load (ETL) environment, a mathematical and statistical analysis environment, an event stream processing environment or other execution environments. Each execution environment can include an execution engine and a respective storage repository of data. An execution engine can include one or multiple execution stages for applying respective operators on data, where the operators can transform or perform some other action with respect to data. A storage repository refers to one or multiple collections of data. An execution environmeni can be available in a public cloud or pubiic network, in which case the execution environment can be referred to as a public cloud execution environment. Alternatively, an execution environment that is available in a private network can be referred to as a private execution
environment.
[00123 A DBMS environment stores data in relational tables and applies database operators (e.g. join operators, update operators, merge operators, and so forth) on data in the relational tables. A MapReduce environment includes map tasks and reduce tasks that can apply a map function and a reduce function, respectively. A map task processes input data to produce intermediate results, based on the respective map function that defines the processing to be performed by the map task. A reduce task takes as input partitions of the intermediate results from the map task to produce an output, based on the corresponding reduce function that defines the processing to be performed by the reduce tasks. 83598076
[0013] Another example execution environment includes an ETL environment, which extracts data from a source (or multiple sources), transforms the data, and loads the transformed data Into a destination. A mathematical/statistical analysis environment may perform mathematical computation on arrays or vectors of data. An event stream processing environment may perform computations on event streams such as sensor data,
[00143 Although specific types of different execution environments are listed above, it is noted that in other examples, other types of execution environments can be used to perform operations on data.
[0015] A service class can be associated with a target performance objective, which identifies one or multiple goals that are to be met by the execution of a workload associated with the service class, A performance objective can also be referred to as a service level objective. An example performance objective relates to an execution time (e.g., a time duration for executing the workload or a target deadline by which the workload is to be completed). Another example
performance objective is a resource usage objective, which can specify that usage of resources, such as computing resources, storage resources, or communication resources, should not exceed a target level. In other examples, other performance objectives can be employed,
[0016] An execution plan for a hybrid flow specifies where (i.e., target execution environments) the sub-flows of the hybrid flow are to execute, and can specify other details associated with execution of the sub-flows (such as order), A single hybrid flow can have many alternative execution plans due to overlap in
functionality among the execution environments, multiple implementatio details for operations, objectives for th execution plans (e.g. objectives relating to fault- tolerance, latency, etc.), and so forth. Based on an execution plan for a hybrid flow, a management system can deplo the sub-flows of the hybrid flow in the target execution environments, and can orchestrate the execution of the hybrid flow. 83598076
[0017] There can be several issues associated with deploying an execution plan in the target execution environments. First, the state of a computing infrastructure of at least one execution environment may have changed between the time the execution plan was produced and the time the execution plan is executed. For example, the execution environment may have become overloaded (such that there is contention for resources) or the computing infrastructure may have experienced a fault. Second, the hybrid flow is associated with a performance objective that has to be met. In some cases, penalties may be specified for not meeting performance objectives. Thus, the management system should ensure that target performance objectives are achieved,
[0018] In some examples, a workload manager may exist within an individual execution environment, and this workload manager can adjust a priority of a task, the number of concurrently running tasks, and so forth, to increase the likelihood that a workload within the individual execution environment meets a respective target objective. However, a workload manager within a single execution environment can only make decisions optimal for that environment. Such a workload manager has no knowledge of the state of other execution environments. So, for a workload that has flows to be executed across a number of different types of execution environments, workload management becomes more challenging.
[0019] in accordance with some implementations, a hybrid flow management system is provided to apply a policy-based hybrid flow management for hybrid flows in a plurality of service classes. The system may include multiple service class workload managers (e.g., one for each service class) to facilitate
management of the hybrid flows according to service class.
[0020] In an example method, a hybrid flow associated with one of the plurality of service classes may be received. The hybrid flow may include multiple sub-flows directed to multiple execution environments. A schedule may be generated to run the sub-flows on the multiple execution environments based on criteria. The criteria may include one or more objectives associated with the hybrid flow's 83598076
service class, allocation of resources of the multiple execution environments to the service class, resource availability, and parameters or constraints associated with the hybrid flow. One or more sub-flows may be dispatched for execution in the execution environments. Execution of the dispatched sub-flows may be monitored to generate statistics. A remedial action may be taken according to a policy associated with the service class if the statistics differ from the objective by a threshold. Accordingly, a hybrid flow associated with a service class may be managed to achieve performance objectives. Furthermore, multiple service classes may be supported, such that a hybrid flow associated with any of the service classes may be appropriately managed to achieve performance objectives associated with the respective service class. Additional examples, advantages, features, modifications and the like are described below with reference to the drawings.
[0021 FIG. 1 illustrates a method for processing a hybrid flow associated with a service class, according to an example. Method 100 may be performed by a computing device, system, or computer, such as hybrid flow management system 400 or computing system 500. Computer-readable instructions for implementing method 100 may be stored on a computer readable storage medium. These instructions as stored on the medium are referred to herein as "modules" and may be executed by a computer.
[0022] Methods 100-300 wilt be described here relative to hybrid flow
management system 400 of FIG, 4. System 400 may include and/or be
implemented by one or more computers. For example, the computers may be server computers, workstation computers, desktop computers, laptops, mobile devices, or the like, and may be part of a distributed system. The computers may include one or more controllers and one or more machine-readable storage media.
[00233 controller may include a processor and a memory for implementing machine readable instructions. The processor may include at least one central processing unit (CPU), at least one semiconductor-based microprocessor, at least 83598076
one digital signal processor (DSP) such as a digital image processing unit, other hardware devices or processing elements suitable to retrieve and execute instructions stored in memory, or combinations thereof. The processor can include single or multiple cores on a chip, multiple cores across multiple chips, multiple cores across multiple devices, or combinations thereof. The processor may fetch, decode, and execute instructions from memory to perform various functions. As an alternative or in addition to retrieving and executing instructions, the processor may include at least one integrated circuit (IC), other control logic, other electronic circuits, or combinations thereof that include a number of electronic components for performing various tasks or functions,
[0024] The controller may include memory, such as a machine-readable storage medium. The machine-readable storage medium may be any electronic, magnetic, optical, or other physical storage device that contains or stores executable instructions. Thus, the machine-readable storage medium may comprise, for example, various Random Access Memory (RAM), Read Only Memory (ROM), flash memory, and combinations thereof. For example, the machine-readabie medium may Include a Non-Volatile Random Access Memory (NVRAM), an
Electrically Erasable Programmable Read-Only Memory (EEPROM), a storage drive, a NAND flash memory, and the like. Further, the machine-readable storage medium can be computer-readable and non-transitory. Additionally, system 400 may include one or more machine-readable storage media separate from the one or more controllers.
[0025]: Hybrid flow management system 400 may include a number of
components. For example, system 400 may include an interface 410 and optimizer 420. The interface 410 can receive a flow definition for creating or modifying a flow. As an example, the interface 410 can present a graphical user interface (GUI) to allow for users to interactively create or modify a flow. Alternatively, a flow can be written in a declarative language and imported through the interface. Th flow that is created or modified using the interface 410 can be a hybrid flow. The 83598076
optimizer 420 generates mu!tip!e candidate execution plans for each flow. The optimizer 420 is able to consider alternative candidate execution plans for a given flow, and can estimate the respective costs of the candidate execution plans.
Examples of costs can include processing resource usage cost, storage resource usage cost, communication cost, input/output (I/O) cost, and so forth. An execution plan (which can be an optimal execution plan) from among the candidate execution plans can be selected for execution, where an optimal execution plan can refer to an execution plan that is associated with a lowest cost or that satisfies some other criterion,
|0026] System 400 may also include an executor 430 and may be connected to execution environments 490 via a network. The network may be any type of communications network, including, but not limited to, wire-based networks (e.g., cable), wireless networks (e.g., cellular, satellite), cellular telecommunications networks), and IP-based telecommunications network(s) (e.g., Voice over Internet Protocol networks). The network may also include traditional landline or a public switched telephone network (PSTN), or combinations of the foregoing.
[00273 The optimizer 420 can provide the selected execution plan to the executor 430, The selected execution plan is itself a flow, and is a hybrid flow if it includes sub-flows directed to two or more of execution environments 490, Executor 430 will now be described in more detail with reference to methods 100-300,
[0028] Method 100 may begin at 110, where a hybrid flow is received. For example, executor 430 may receive a selected execution plan from the optimizer 420. The hybrid flow may be associated with one of a plurality of service classes supported by the system 400. Service classes are an approach for providing differentiated service for requests.
[0029] service class may be determined and/or assigned to the hybrid flow in various ways. For example, a user may have signed in to interface 410 to create a flow. The user may be associated with a particular service class by user id by virtue of the user's role in the enterprise. For example, if the user id corresponds to a CEO of th enterprise, any generated flow may be assigned to a service class a 83598076
guaranteeing the highest priority (e.g., fastest execution, largest number of allocated resources, etc.). In contrast, a user id corresponding to a marketing department or engineering department may be assigned to a service class having a lower priority. The service class may also be dictated by the size of the flow, the type of flow, the number of execution environments required by the flow, etc. A service class may be determined by the geographical location and time of the request origination. Various other methods of determining and/or assigning a service class to a flow may be used.
10030] At 120, a schedule may be generated to run the sub-flows of the hybrid flow on the execution environments. For example, scheduler 450 may be configured to generate the schedule for execution of the sub-flows. Additionally, scheduler 450 may generate a separate schedule for execution of hybrid flows associated with different service classes, such that each service class has its own schedule for execution of flows associated with that service class.
[0031] Various scheduling techniques may be used, such as first fit or bi packing. Additionally, the schedule may be generated based on various criteria. For example, the schedule may be generated based on one o more objectives associated with the service class, allocation of resources of the multiple execution environments to the service class, resource availability, and parameters and/or constraints associated with the hybrid flow.
[0032] The objectives may be any of various specified service level objectives for the service class (e.g., 90% of flows must complete in fewer than one minute, the entire set of flows must complete by a specified time, etc). The objectives for a given service class may be stored in computer-readable medium 486 of the corresponding service class workload manager 480 as objectives 487. The scheduler 450 may thus operate under the constraint that the generated schedule meets objectives 487.
[0033] The allocation of resources may also be dictated by the service class. For example, each service class may be allocated a set of resources from each execution environment. For instance, a service class may be allocated a number 83598076
of processing cores, a minimum or maximum amount of memory, a fraction of network bandwidth, etc. Additionally, the ailocaiion of resources may be specific to each execution environment 490, For instance, a service class may be ai located 50% of the processing cores of E but only 25% of the processing cores of E2- [0034] Resource manager 440 may be used to manage the allocatio of resources to the various service classes by assigning particular allocations to the service class workload managers 480. In one example, the allocation of resources among all the service classes should not exceed the total number of resources in the execution environments. In another example, the initial ailocaiion of resources may exceed the total number of resources, but logic in th resource manager 440 may ensure that as the total number in use of any one resource is being
approached, the allocated amount of that resource to each service class workload manager 480 is modified so that there isn't resource contention,
[0035] Resource manager 440 may perform method 200. At 210, resource manager 440 may monitor resource availability. This may be done by querying the execution environments 490 for their status via an application programming interface. At 220, resource manager 440 may manage the allocation of resources to the service class workload managers 480. As discussed above, the allocation of resources may be dictated by the service classes. For example, each service class may have a set of resource requirements which are communicated to the resource manager 440. As resource availability changes over time (such as due to additional resources being made available in an execution environment, resources failing in an execution environment, etc.), resource manager 440 ma modify the resource allocation. Additionally, as discussed later, resource manager 440 may modify the resource allocation in response to requests from the service class workload managers 480. Resource manager 440 may also maintain other parameters and statistics associated with the execution environments, such as a multi-programming level allowed for each environment.
£0036] Finally, returning to method 100, the criteria by which the schedule is generated may include parameters and/or constraints associated with the hybrid 83598076
flow. For example, various sub-f!ows in the hybrid flow may have dependencies on one another, such that certain sub-flows should be executed concurrently, certain sub-flows should be executed before or after other sub-flows, etc. Also, the hybrid flow may be associated with an arrival time (e.g., the time when the flow first arrived at optimizer 420 or executor 430) and a requested execution time (e.g., a requested time input by a user via interface 410). The parameters may also relate to the number of sub-flows in the hybrid flow, the minimum amount of time needed to fully execute the hybrid flow, the number and type of execution environments needed for execution of the flow, etc. Other parameters and constraints may be considered as well.
[0037] Scheduler 450 may take all of these criteria into consideration to generate a schedule for execution of the hybrid flow. Furthermore, the scheduler 450 may- generate this schedule for a plurality of hybrid flows associated with the given service class. For example, there may be five independent hybrid flows associated with a given service class that are being managed at the same time by system 400. The scheduler 450 may consider these criteria for all five hybrid flows to generate a schedule for execution of the sub-flows of these five hybrid flows on execution environments 490 such that objectives and constraints associated with the flows are met. Additionally, scheduler 450 may recompute the schedule if a new flow is received, if resources are reallocated, etc,
[0038] At 130, one or more of the sub-flows may be dispatched to the execution environments for execution based on the schedule. For example, dispatcher 470 may be configured to dispatch the sub-flows to execution environments 490.
However, before dispatching the sub-flows, a validation process may be performed according to method 300 of FIG. 3. Validator 460 may perform method 300. At 310, the execution requirements of a sub-flow to be dispatched may be verified. For example, validator 460 may verify via resource monitor 440 that all resources required for execution of the sub-flow are available and appropriately allocated to the service class. It may also be verified that objectives 487 will be met, that all 83598076
dependencies have been appropriately satisfied, etc. If the verification succeeds, the sub-flow ma be dispatched,
[0039] At 320, if the verification fails, an action may be taken. For example, in the event of failure, validator 460 may inform the service class workload manager 480 that the verification fa tied. Action module 484 of service class workload manager 480 may then take a remedial acfion(s) in accordance with one or more policies 488. Policies 488 may be specified for each service class to deaf with plan snvaiidation. There may be separate policies for the various circumstances of plan invalidation and the policies may result in different actions being taken.
[0040] The following is a non-exhaustive list of remedial actions that may be taken. For example, the service class workload manager 480 may request additional resources from one or more of the execution environments 490. Such a request may entail removing resources allocated to another service class and reallocating those resources to the service class associated with the invalidated sub-flow. The service class workload manager 480 may request instantiation of an additional execution environment. The service class workload manager 480 may request that the given sub-flo be sent to a different execution environment for execution thereon. The service class workload manager 480 may request that the sub-flow be decomposed into smaller fragments. The service class workload manager 480 may request that scheduler 450 generate a new schedule, that optimizer 420 send a new execution plan appropriate to the current state of the computing environments (e.g., from a list of previously generated execution plans), or that optimizer 420 generate a ne execution plan altogether.
[0041]: Returning to method 100, execution of the dispatched sub-flows may be monitored. For example, monitoring module 482of the service class workload manager 480 may monitor execution of the sub-flows associated with the service class. Statistics 489 may be generated based on the monitoring, such as average execution time of the sub-flows, proportion of sub-flows executed, projected total execution time of the hybrid flow, etc. If the statistics differ from one or more 83598076
objectives 487 by a threshold, action module 484 may take an action according to one or more of policies 488, such as described above,
[0042] FIG. 5 illustrates a computing system for processing a hybrid flow associated with a service class, according to an example. Computing system 500 may include and/or be implemented by one or more computers. For example, the computers may be server computers, workstation computers, desktop computers, laptops, mobile devices, or the like, and may be part of a distributed system. The computers may include one or more controllers and one or more machine-readable storage media, as described with respect to hybrid flow management system 400, for example.
[0043] In addition, users of computing system 500 may interact with computing system 500 through one or more other computers, which may or may not be considered part of computing system 500. As an example, a user may interact with system 500 via a computer application residing on system 500 or on another computer, such as a desktop computer, workstation computer, tablet computer, or the like. The computer application can include a user interface (e.g., touch interface, mouse, keyboard, gesture input device).
[0044] Computing system 500 may perform methods 100-300, and variations thereof, and components 510-540 may be configured to perform various portions of methods 100-300, and variations thereof. Additionally, the functionality
implemented by components 510-540 may be part of a larger software platform, system, application, or the like. For example, these components may be part of a data analysis system.
[0045] Computers) 510 may have access to database 540, The database may include one or more computers, and may include one or more controllers and machine-readable storage mediums, as described herein. The computer may be connected to the database via a network. The network may be any type of communications network, including, but not limited to, wire-based networks (e.g., cable), wireless networks (e.g., cellular, satellite), cellular telecommunications 83598076
nefwork(s), and IP-based telecommunications network(s) (e.g., Voice over internet Protocol networks). The network may also include traditional landline or a public switched telephone network (PSTN), or combinations of the foregoing.
[0046] Processor 520 may be at least one central processing unit (CPU), at least one semiconductor-based microprocessor, other hardware devices or processing elements suitable to retrieve and execute instructions stored in machine-readable storage medium 530, or combinations thereof. Processor 520 can include single or multiple cores on a chip, multiple cores across multiple chips, multiple cores across multiple devices, or combinations thereof. Processor 520 may fetch, decode, and execute instructions 532-538 among others, to implement various processing. As an alternative or in addition to retrieving and executing instructions, processor 520 may include at ieast one integrated circuit (JC), other control logic, other electronic circuits, or combinations thereof that include a number of electronic components for performing the functionality of instructions 532-538. Accordingly, processor 520 may be implemented across multiple processing units and instructions 532-538 may be implemented by different processing units in different areas of computer 510.
[0047] Machine-readable storage medium 530 may be any electronic, magnetic, optical, or other physical storage device that contains or stores executable instructions. Thus, the machine-readable storage medium may comprise, for example, various Random Access Memory (RAM), Read Only Memory (ROM), flash memory, and combinations thereof. For example, the machine-readabl medium may include a Non-Volatile Random Access Memory (NVRAM), an
Electrically Erasable Programmable Read-Only Memory (EEPROM), a storage drive, a NAND flash memory, and the like. Further, the machine-readable storage medium 530 can be computer-readable and non-transitory. Machine-readable storage medium 530 may be encoded with a series of executable instructions for managing processing elements. 83598076
[0048] The instructions 532-538 when executed by processor 520 (e.g., via one processing element or multiple processing elements of the processor) can cause processor 520 to perform processes, for example, methods 100-300, and/or variations and portions thereof.
[00493 For example, service class instructions 532 may cause processor 520 to determine a service class associated with each of a plurality of hybrid flows.
Scheduling instructions 534 may cause processor 520 to generate a separate execution schedule for the hybrid flows in each service class based o parameters and constraints associated with each hybrid flow, objectives and resource allocation associated with the respective service class, and resource availability. Monitoring instructions 538 may cause processor 510 to monitor the execution of each hybrid flow to generate statistics. Action instructions 538 may cause processor 520 to take a remedial action if the statistics for a given hybrid flow differ from objectives associated with the given hybrid flow's service class by a threshold. The remedial action may be dictated by a policy associated with the given hybrid flow's service ciass.
[0050] In the foregoing description, numerous details are set forth to provide an understanding of the subject matter disclosed herein. However, implementations may be practiced without some o all of these details. Other implementations may include modifications and variations from the details discussed above. It is intended that the appended claims cover such modifications and variations.

Claims

83598076 CLAIMS What is claimed is;
1. A method comprising, by a hybrid flow management system:
receiving a hybrid flow comprising multiple sub-flows directed to at least two execution environments, the hybrid flow being associated with a first service class of a plurality of service classes;
generating a schedule to run the hybrid flow's sub-flows on the at least two execution environments based on criteria including an objective associated with the first service class and resource availability of the at least two execution
environments;
dispatching a sub-flow of the multiple sub-flows to one of the execution environments for execution based on the schedule;
monitoring the execution of the dispatched sub-flow to generate statistics; and
taking an action according to a policy associated with the first service class if the statistics differ from the objective by a threshold.
2. The method of claim 1 , further comprising;
before dispatching the sub-flow to one of the execution environments;
verifying that all resources required for execution of the sub-flow are available; and
taking an action according to a policy associated with the first service class if all resources required for execution of the sub-flow are not available.
3. The method of claim 1 , further comprising
monitoring resource availability of the at least two execution environments; and 83598076
managing allocation of resources of the at least two execution environments to the plurality of service classes.
4. The method of claim 1 , wherein taking an action according to a service policy associated with the first service class comprises requesting additional resources from one or more of the at least two execution environments or requesting an additional execution environment.
5. The method of claim 1 , wherein taking an action according to a service policy associated with the first service class comprises decomposing the sub-flow into multiple parts.
6. The method of claim 1 , wherein taking an action according to a service policy associated with the first service class comprises selecting a different schedule to run the hybrid flow's sub-flows or generating a different schedule to run the hybrid flow's sub-flows.
7. The method of claim 1 , wherein taking an action according to a service policy associated with the first service class comprises sending the sub-flow to a different execution environment for execution.
8. The method of claim 1 , comprising:
receiving a second hybrid flow comprising second multiple sub-flows directed to the at least two execution environments, the second hybrid flow being associated with a second service class of the plurality of service classes;
generating a second schedule to run the second hybrid flow's sub-flows on the at least two execution environments based on criteria including a second objective associated with the second service class and resource availability of the at least two execution environments, 83598076
wherein the resource availability for the first service class differs from the resource availability for the second service class due to differing allocations of resources of the at least two execution environments to the first and second service classes,
9. A system comprising:
a resource manager to manage allocation of resources of a plurality of execution environments to a plurality of service classes, the plurality of service classes including a first service class;
a scheduler to generate a schedule for execution of a hybrid flow associated with the first service class based on service level objectives associated with the first service class and an allocation of resources to the first service class; and
a service class workload manager associated with the first service class comprising:
a monitoring module to monitor execution of the hybrid flow in the plurality of execution environments; and
an action module to take an action according to a policy associated with the first service class if execution of the hybrid flow fails to meet one or more of the service level objectives associated with the first service class.
10. The system of claim 9, comprising a dispatcher to dispatch sub-flows of the hybrid flow to the plurality of execution environments for execution according to the schedule.
11. The system of claim 10, comprising a validator to validate a sub-flow before dispatch by the dispatcher by verifying that resources required for execution of the sub-flow are available and that the objectives are expected to be met. 83598076
12 The system of claim 9, wherein the policy causes the action module to request additional resources from the resource manager.
13. The system of claim 9, wherein the policy causes the action module to request that a sub-flow of the hybhd flow be sent to another execution environment for execution.
14. The system of claim 9, comprising:
a second service class workload manager associated with a second service ciass of the plurality of service classes, the second service ciass workload manager comprising:
a second monitoring module to monitor execution of hybrid flows associated with the second service class in the plurality of executio environments; and
a second action module to fake an action according to a policy associated with the second service class if execution of one of the hybrid flows associated with the second service class fails to meet one or more service level objectives associated with the second service class.
15. A non-transitory computer-readable storage medium storing instructions for execution by a computer, the instructions when executed causing the computer to: determine a service class associated with each of a plurality of hybrid flows; generate a separate execution schedule for the hybrid flows in each service class based on parameters and constraints associated with each hybrid flow, objectives and resource allocation associated with the respective service ciass, and resource availability;
monitor the execution of each hybrid flow to generate statistics; and take a remedial action if the statistics for a given hybrid flow differ from objectives associated with the given hybrid flow's service ciass by a threshold, the 83598076
remediaS action being dictated by a policy associated with the given hybrid flow's service class.
PCT/US2013/062201 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class WO2015047302A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP13894923.5A EP3049959A1 (en) 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class
PCT/US2013/062201 WO2015047302A1 (en) 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class
US14/917,059 US20160203026A1 (en) 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class
CN201380079890.3A CN105580001A (en) 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2013/062201 WO2015047302A1 (en) 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class

Publications (1)

Publication Number Publication Date
WO2015047302A1 true WO2015047302A1 (en) 2015-04-02

Family

ID=52744205

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2013/062201 WO2015047302A1 (en) 2013-09-27 2013-09-27 Processing a hybrid flow associated with a service class

Country Status (4)

Country Link
US (1) US20160203026A1 (en)
EP (1) EP3049959A1 (en)
CN (1) CN105580001A (en)
WO (1) WO2015047302A1 (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015094269A1 (en) * 2013-12-19 2015-06-25 Hewlett-Packard Development Company, L.P. Hybrid flows containing a continuous flow
US10375161B1 (en) * 2014-04-09 2019-08-06 VCE IP Holding Company LLC Distributed computing task management system and method
US10623481B2 (en) * 2015-04-27 2020-04-14 Microsoft Technology Licensing, Llc Balancing resources in distributed computing environments
US10715526B2 (en) 2016-12-14 2020-07-14 Microsoft Technology Licensing, Llc Multiple cores with hierarchy of trust
US10402273B2 (en) 2016-12-14 2019-09-03 Microsoft Technology Licensing, Llc IoT device update failure recovery
US10416991B2 (en) 2016-12-14 2019-09-17 Microsoft Technology Licensing, Llc Secure IoT device update
CN109445931A (en) * 2018-08-31 2019-03-08 安徽四创电子股份有限公司 A kind of big data resource scheduling system and method
CN114268800B (en) * 2020-09-14 2023-06-16 腾讯科技(深圳)有限公司 Live broadcast resource scheduling method, device, equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070214171A1 (en) * 2006-03-10 2007-09-13 International Business Machines Corporation Data flow system and method for heterogeneous data integration environments
US20090241117A1 (en) * 2008-03-20 2009-09-24 International Business Machines Corporation Method for integrating flow orchestration and scheduling for a batch of workflows

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7596618B2 (en) * 2004-12-07 2009-09-29 Hewlett-Packard Development Company, L.P. Splitting a workload of a node
US7958507B2 (en) * 2005-06-16 2011-06-07 Hewlett-Packard Development Company, L.P. Job scheduling system and method
US8131302B2 (en) * 2007-07-20 2012-03-06 Broadcom Corporation Method and system for dynamically splitting jobs across multiple agnostic processors in wireless system
US8284205B2 (en) * 2007-10-24 2012-10-09 Apple Inc. Methods and apparatuses for load balancing between multiple processing units
US20140298349A1 (en) * 2008-04-21 2014-10-02 Adaptive Computing Enterprises, Inc. System and Method for Managing Energy Consumption in a Compute Environment
US8990812B2 (en) * 2008-07-07 2015-03-24 Infosys Limited Task decomposition with throttled message processing in a heterogeneous environment
US8606762B2 (en) * 2008-09-16 2013-12-10 Sap Ag Data quality administration framework
US8266477B2 (en) * 2009-01-09 2012-09-11 Ca, Inc. System and method for modifying execution of scripts for a job scheduler using deontic logic
US8453156B2 (en) * 2009-03-30 2013-05-28 Intel Corporation Method and system to perform load balancing of a task-based multi-threaded application
US8479216B2 (en) * 2009-08-18 2013-07-02 International Business Machines Corporation Method for decentralized load distribution in an event-driven system using localized migration between physically connected nodes and load exchange protocol preventing simultaneous migration of plurality of tasks to or from a same node
US8854381B2 (en) * 2009-09-03 2014-10-07 Advanced Micro Devices, Inc. Processing unit that enables asynchronous task dispatch
CN102043673B (en) * 2009-10-21 2015-06-03 Sap欧洲公司 Calibration of resource allocation during parallel processing
US8307198B2 (en) * 2009-11-24 2012-11-06 Advanced Micro Devices, Inc. Distributed multi-core memory initialization
EP2609502A4 (en) * 2010-08-24 2017-03-29 Jay Moorthi Method and apparatus for clearing cloud compute demand
US8510284B2 (en) * 2010-12-20 2013-08-13 Microsoft Corporation Large-scale event evaluation using realtime processors
US9223631B2 (en) * 2011-03-17 2015-12-29 Apple Inc. Performing an operation using multiple services
US9063790B2 (en) * 2011-06-13 2015-06-23 Accenture Global Services Limited System and method for performing distributed parallel processing tasks in a spot market
US9201690B2 (en) * 2011-10-21 2015-12-01 International Business Machines Corporation Resource aware scheduling in a distributed computing environment
CN103197952B (en) * 2012-01-09 2017-08-25 华为技术有限公司 The management system and method disposed for application system maintenance based on cloud infrastructure
US9477708B2 (en) * 2012-11-19 2016-10-25 Nec Corporation System for multi-store analytics execution environments with storage constraints
US9031531B2 (en) * 2012-11-20 2015-05-12 At&T Intellectual Property I, L.P. Distributed computing task costing with a mobile device
US9262492B2 (en) * 2012-11-30 2016-02-16 Hewlett Packard Enterprise Development Lp Dividing and combining operations
US10489416B2 (en) * 2013-02-28 2019-11-26 Micro Focus Llc Optimizing and managing execution of hybrid flows
US10102039B2 (en) * 2013-05-17 2018-10-16 Entit Software Llc Converting a hybrid flow
US9602423B2 (en) * 2013-06-28 2017-03-21 Pepperdata, Inc. Systems, methods, and devices for dynamic resource monitoring and allocation in a cluster system
JP6451149B2 (en) * 2014-09-01 2019-01-16 富士通株式会社 Management device, migration control program, information processing system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070214171A1 (en) * 2006-03-10 2007-09-13 International Business Machines Corporation Data flow system and method for heterogeneous data integration environments
US20090241117A1 (en) * 2008-03-20 2009-09-24 International Business Machines Corporation Method for integrating flow orchestration and scheduling for a batch of workflows

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
BOEHM, MATTHIAS: "Cost-Based Optimization of Integration Flows", PH.D. THESIS, 2011, XP055292039, Retrieved from the Internet <URL:https://wwwdb.inf.tu-dresden.de/misc/team/boehm/pubs/diss_final.pdf> *
KLLAPI, HERALD ET AL.: "Schedule Optimization for Data Processing Flows on the Cloud", ACM SIGMOD INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA 2011, 12 June 2011 (2011-06-12), pages 289 - 300, XP055292040, Retrieved from the Internet <URL:http://www.sigmod2011.org/index.shtml> *
SIMITSIS, ALKIS ET AL.: "HFMS: Managing the lifecycle and complexity of hybrid analytic data flows", IEEE 29TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE) 2013, 8 April 2013 (2013-04-08), pages 1174 - 1185, XP055292038, Retrieved from the Internet <URL:http://ieeexplore.ieee.org/xpl/mostRecentIssue.jsp?punumber=6530811> *

Also Published As

Publication number Publication date
EP3049959A1 (en) 2016-08-03
CN105580001A (en) 2016-05-11
US20160203026A1 (en) 2016-07-14

Similar Documents

Publication Publication Date Title
Abdulhamid et al. Fault tolerance aware scheduling technique for cloud computing environment using dynamic clustering algorithm
US10423457B2 (en) Outcome-based software-defined infrastructure
US20160203026A1 (en) Processing a hybrid flow associated with a service class
Tumanov et al. TetriSched: global rescheduling with adaptive plan-ahead in dynamic heterogeneous clusters
US10623269B2 (en) Operator fusion management in a stream computing environment
US10885127B2 (en) Machine-learning to alarm or pre-empt query execution
AU2017236024B2 (en) Flow engine for building automated flows within a cloud based development platform
US11467874B2 (en) System and method for resource management
US20200082316A1 (en) Cognitive handling of workload requests
US10620928B2 (en) Global cloud applications management
Rajavel et al. Achieving service level agreement in cloud environment using job prioritization in hierarchical scheduling
US10635492B2 (en) Leveraging shared work to enhance job performance across analytics platforms
US9658893B2 (en) Multilayered resource scheduling
Incerto et al. Symbolic performance adaptation
Tuli et al. Optimizing the performance of fog computing environments using ai and co-simulation
US20230266997A1 (en) Distributed scheduling in container orchestration engines
Loganathan et al. Job scheduling with efficient resource monitoring in cloud datacenter
Mosa et al. Towards a cloud native big data platform using micado
Islam et al. Resource management and scheduling for big data applications in cloud computing environments
Fernández-Cerero et al. Quality of cloud services determined by the dynamic management of scheduling models for complex heterogeneous workloads
Hamad An overview of Hadoop scheduler algorithms
Liao et al. Efficient proactive resource allocation for multi-stage cloud-native microservices
Vaidehi et al. An Efficient Job Classification Technique to Enhance Scheduling in Cloud to Accelerate the Performance
Chu et al. Applying Software Defined Methodologies to Software Computing
US10310899B2 (en) Database and a development project system, method, and recording medium

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201380079890.3

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13894923

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 14917059

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2013894923

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2013894923

Country of ref document: EP