US9841998B2 - Processor power optimization with response time assurance - Google Patents
Processor power optimization with response time assurance Download PDFInfo
- Publication number
- US9841998B2 US9841998B2 US15/070,792 US201615070792A US9841998B2 US 9841998 B2 US9841998 B2 US 9841998B2 US 201615070792 A US201615070792 A US 201615070792A US 9841998 B2 US9841998 B2 US 9841998B2
- Authority
- US
- United States
- Prior art keywords
- processor
- tasks
- allocating
- computer
- response time
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Program initiating; Program switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/4881—Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
- G06F9/4893—Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues taking into account power or heat criteria
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/329—Power saving characterised by the action undertaken by task scheduling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5011—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5027—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5094—Allocation of resources, e.g. of the central processing unit [CPU] where the allocation takes into account power or heat criteria
-
- Y02B60/142—
-
- Y02B60/144—
-
- Y02B60/167—
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- the present invention relates generally to the field of computing, and more particularly to processor power management.
- Processor power management is an increasingly important strategy with the proliferation of computers. Reducing energy usage by processors provides many advantages, such as reducing costs, increasing component longevity and reducing heat production. For example, in a datacenter containing servers used in cloud computing, the energy required to power processors and attendant cooling systems may represent a significant operating expense. To reduce processor energy consumption, processor power usage mechanisms have been developed to dynamically control processor power usage.
- Examples of some processor power usage mechanisms may include: dynamic voltage and frequency scaling (DVFS), processor core nap, and processor folding.
- Dynamic voltage scaling is a power usage mechanism where the voltage used in a processor can be increased or decreased dynamically.
- Dynamic frequency scaling is a power usage mechanism where the processor frequency can be increased or decreased dynamically. Lower voltage or lower frequency yields less power consumption by the processor.
- Processor core nap can be used to clock off most of the circuits inside a processor core when there is no work to be done by the core to save energy. This causes the core to go into a low-power idle state.
- Processor folding is a power usage mechanism where the number of available processor cores can be increased or decreased dynamically. Tasks are reallocated to the available cores and the other cores remain in low-power idle states.
- processor power usage mechanisms must be utilized to facilitate desired energy reduction goals. If, however, processor power usage mechanisms are aggressively used to reduce energy consumption, processor throughput may drop too low resulting in unacceptable task response time for a user.
- a method for managing processor power optimization may include receiving a plurality of tasks for processing by a processor environment.
- the method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
- a computer system for managing processor power optimization may include one or more processors, one or more computer-readable memories, one or more computer-readable tangible storage devices, and program instructions stored on at least one of the one or more storage devices for execution by at least one of the one or more processors via at least one of the one or more memories, wherein the computer system is capable of performing a method.
- the method may include receiving a plurality of tasks for processing by a processor environment.
- the method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
- a computer program product for managing processor power optimization.
- the computer program product may include one or more computer-readable storage devices and program instructions stored on at least one of the one or more tangible storage devices, the program instructions executable by a processor.
- the computer program product may include receiving a plurality of tasks for processing by a processor environment.
- the method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
- FIG. 1 is a block diagram of the processor power optimization with response time assurance system according to one embodiment
- FIG. 2 is an operational flow chart illustrating a process for task classification according to at least one embodiment
- FIG. 3 is an operational flowchart illustrating a process for class allocation according to at least one embodiment
- FIG. 4 is an operational flowchart illustrating a process for energy management according to at least one embodiment.
- FIG. 5 is a block diagram of internal and external components of computers and servers.
- Embodiments of the present invention relate generally to the field of computing, and more particularly to processor power management.
- the following described exemplary embodiments provide a system, method and program product for processor power optimization with response time assurance.
- the present invention may be a system, a method, and/or a computer program product.
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the block may occur out of the order noted in the figures.
- two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- the following described exemplary embodiments provide a system, method and program product for processor power optimization with response time assurance.
- the processor power optimization with response time assurance system 100 may include a task classifier process unit 102 , a class allocation process unit 104 and an energy manager process unit 106 .
- the processor power optimization with response time assurance system 100 receives tasks 108 having execution time and response time characteristics.
- the task classifier process unit 102 may then assign each of the tasks 108 to one of class queues 110 a - 110 c based on each tasks 108 execution time and response time characteristics.
- the task classifier process unit 102 may monitor header information associated with the received tasks 108 .
- the task classifier process unit 102 may utilize the header information to indicate the task-type (e.g., content streaming, searches, payment transaction, etc.).
- the task classifier process unit 102 may correlate each indicated task-type to one of the class queues 110 a - 110 c .
- received task A 1 may be determined to be a content streaming task-type.
- the lookup table may accordingly assign class queue 110 b to the content streaming task-type, whereby class queue 110 b stores or buffers task-types having the same or similar execution time and response time characteristics to that of the content streaming task designated by A 1 .
- the class allocation process unit 104 then allocates a portion of a compute resource 112 to each active class queue 110 for processing a task 108 within the desired response time.
- the compute resource 112 may be the total number of threads of execution (threads) supported by one or more processor cores contained within a processor environment 114 .
- the processor environment 114 may be, for example, a single device or multiple interconnected devices where each device may contain one or more single core or multi-core processors. Examples of such devices may include a server, a desktop computer, a laptop, a telephone, a PDA, a tablet computer, or any other device containing one or more processors.
- the energy manager process unit 106 monitors the amount of compute resource utilization and increases or decreases available compute resource 112 as necessary through processor power usage mechanisms such as DVFS, processor core nap and processor folding to ensure compute resource utilization is within a target range.
- Task classifier process 200 may be executed within the task classifier process unit 102 ( FIG. 1 ).
- the task classifier process 200 begins at 202 by defining N different task classes. A class is created corresponding to each unique combination of execution time and response time characteristics for every task that the processor power optimization with response time assurance system 100 ( FIG. 1 ) will process.
- a class queue 110 ( FIG. 1 ) data structure may be created for each class defined.
- N class queues 110 ( FIG. 1 ) may be created.
- the class queues 110 ( FIG. 1 ) created may, according to one embodiment, operate under a first in first out (FIFO) policy.
- the class queues 110 ( FIG. 1 ) may be implemented, for example, in software by using an array or linked list, or in hardware by using a buffer or registers.
- the task classifier process 200 checks for an arriving task 108 ( FIG. 1 ) to process. If there is no new task 108 ( FIG. 1 ) found, the task classifier process unit 102 may wait a short amount of time and check again for the arrival a new task 108 ( FIG. 1 ).
- the new task 108 ( FIG. 1 ) is classified based on the task 108 ( FIG. 1 ) execution time and response time characteristics. Then, according to one embodiment, the task 108 ( FIG. 1 ) is placed at the end of one the designated class queues 110 a - 110 b ( FIG. 1 ) corresponding to the class the task 108 ( FIG. 1 ) was assigned. The task classifier process unit 102 then returns to 206 to check for arrival of any new tasks 108 ( FIG. 1 ) to process.
- Class allocation process 300 may be executed within the class allocation process unit 104 ( FIG. 1 ).
- the class allocation process 300 specifies a normalized delay factor D for each class i associated with the received tasks 108 ( FIG. 1 ).
- each class i is assigned based on its respective response time characteristic. Classes having faster response time characteristics (e.g., streaming video) are assigned numerically higher delay factors D i , while classes having slower response times (e.g., payment transaction) are assigned numerically lower delay factors D i .
- D 2 5
- the class allocation process 300 specifies a normalized execution time E for each class i associated with the received tasks 108 ( FIG. 1 ).
- each class i is assigned based on its respective execution time characteristic.
- Classes requiring more time to execute e.g., payment transaction
- classes requiring less time to execute are assigned numerically lower execution times E i .
- may require a shorter execution time and be assigned a normalized delay factor E 2 1.
- the compute resource 112 ( FIG. 1 ) available for all task classes at the current time t is specified as P(t), where P(t) is less than or equal to the maximum processing capability of the compute resource 112 ( FIG. 1 ).
- the value P(t) comprises, for example, the total number of threads that can be processed by all of the active cores.
- a server may have (15) processors, where each processor includes (20) cores each supporting (4) threads.
- the class allocation process 300 calculates the amount of compute resource 112 ( FIG. 1 ) each class will be allocated in proportion to the delay factor D and execution time E of each class i.
- the compute resource allocation factor A for each class i may be determined by the following formula:
- a i D i ⁇ E i ( D 1 ⁇ E 1 ) + ( D 2 ⁇ E 2 ) + ... + ( D N ⁇ E N )
- E in the compute resource allocation factor formula is replaced with 0 for classes that have no tasks currently within the class queue 110 ( FIG. 1 ) (i.e., the class is inactive) to ensure that no compute resources 112 ( FIG. 1 ) are allocated to an inactive class.
- the resulting value of A i is less than or equal to 1 and greater than 0 when there is at least one class defined.
- the sum of all A i equals 1, representing 100%, thereby ensuring that all of the available compute resource will be apportioned across all of the active classes.
- the compute resource allocation factor formula is computed three times, once for every active class.
- the resource allocation factor is computed for class 1 as follows:
- the resource allocation factor is then computed for class 2 and class 3 using the same formula.
- the compute resource 112 ( FIG. 1 ) is allocated to each class corresponding to the proportion determined at 308 .
- the compute resource 112 ( FIG. 1 ) processes all active classes simultaneously but each class may be processed at different rates.
- the amount of total compute resource available P(t) (e.g., total number of available threads) when multiplied by the resource allocation factor A i results in the amount of compute resource R i number of threads) that may be used by class i to process tasks 108 ( FIG. 1 ).
- having a class 1 with a compute resource allocation factor A 1 0.645
- 774 threads (0.645 ⁇ 1200 774) allocated to class 1 to process tasks 108 ( FIG. 1 ).
- all remaining active classes will be allocated compute resource 112 ( FIG. 1 ) using the same calculation.
- the class allocation process 300 checks if the energy manager process unit 106 ( FIG. 1 ) has sent a notification to reapportion the compute resource 112 ( FIG. 1 ). If a notification to reapportion compute resources is received, the class allocation process will go to 306 . However, if no notification is received, the class allocation process 300 may recheck until a notification is received. For example, if one task (e.g., task of class 1) completes, the energy manager process within the energy manager process unit 106 ( FIG. 1 ) initiates a reapportioning of the compute resources by sending a notification to the class allocation process 300 . In addition to sending the notification, the energy manager process may also reduce the total available compute resource in accordance with the number of tasks that are actively being processed.
- one task e.g., task of class 1
- the energy manager process may also reduce the total available compute resource in accordance with the number of tasks that are actively being processed.
- a target compute resource utilization range is defined by designating two limits.
- the target compute resource utilization range extends between a lower limit U L representing the minimum target resource utilization and an upper limit U H representing the maximum target resource utilization.
- the compute resource utilization level is initially set to full utilization. This may be done, for example, by making all of the processors and processor cores active and running at full speed in the processor environment 114 ( FIG. 1 ). Full compute resource utilization is made available to ensure that the initial set of tasks 108 ( FIG. 1 ) are supported until compute resource utilization level can be reliably measured.
- the compute resource utilization level U is measured at the current time t.
- the current utilization U(t) is compared with the lower limit U L . If the current utilization U(t) is lower than the lower limit U L , then there is too much active compute resource not being utilized that may in turn be consuming unnecessary energy.
- current utilization U(t) is not lower than the lower limit U L , then at 410 , the current utilization U(t) is compared with the upper limit U H . If current utilization U(t) is higher than the upper limit U H , then there is not enough active or available compute resources. However, if current utilization U(t) is not higher than the upper limit U H , then current utilization U(t) is within the target compute resource utilization range and there is no need to change the available compute resource and the energy manager process 400 returns to 406 .
- the energy manager process 400 at 408 , checks if U(t) (93%) is less than U L (90%). In this example, U(t) is not lower than U L and the process 400 continues to 410 to check if U(t) (93%) is higher than U H (95%).
- U(t) is not higher than U H , thus the energy manager process 400 has determined that U(t) falls within the target resource utilization range.
- the energy manager process 400 responds to the condition where current utilization U(t) is lower than the lower limit U L (“yes” at 408 ) by reducing available compute resources through processor power management mechanisms.
- processor power management mechanisms may include utilizing DVFS, processor core nap and processor folding to scale down processor core frequency or voltage, or put processor cores into low-power idle states to reduce available compute resource 112 ( FIG. 1 ) and also reduce energy usage.
- One example for triggering a reduction in computer resource may occur when one or more tasks, as indicated by T 1A and T 1B ( FIG. 1 ), in the class queue (e.g., FIG. 1 : queue 110 a ) are completed.
- the energy manager process 400 responds to the condition where current utilization U(t) is higher than the upper limit U H (“yes” at 410 ) by increasing available compute resources through processor power management mechanisms.
- processor power management mechanisms may include utilizing DVFS, processor core nap and processor folding to reactivate processor cores from low-power idle states or scale processor core frequency or voltage up to increase available compute resource 112 ( FIG. 1 ) for processing tasks 108 ( FIG. 1 ).
- One example for triggering an increase in computer resource may occur when one or more tasks, as indicated by T 1A and T 1B ( FIG. 1 ), are added to the class queue (e.g., FIG. 1 : queue 110 a ).
- a notification is sent to the class allocation process 300 associated with class allocation process unit 104 ( FIG. 1 ) to reapportion the compute resource because the amount of compute resource available has changed.
- the notification system may be implemented in a variety of ways, such as, for example, a function call or as a flag variable set to a predefined value.
- FIG. 5 shows a block diagram of the exemplary components of a data processing system 800 , 900 , such as the processor power optimization with response time assurance system 100 ( FIG. 1 ) in accordance with an illustrative embodiment of the present invention. It should be appreciated that FIG. 5 provides only an illustration of one implementation and does not imply any limitations with regard to the environments in which different embodiments may be implemented. Many modifications to the depicted environments may be made based on design and implementation requirements.
- Data processing system 800 , 900 is representative of any electronic device capable of executing machine-readable program instructions.
- Data processing system 800 , 900 may be representative of a smart phone, a computer system, PDA, or other electronic devices.
- Examples of computing systems, environments, and/or configurations that may represented by data processing system 800 , 900 include, but are not limited to, personal computer systems, server computer systems, thin clients, thick clients, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, network PCs, minicomputer systems, and distributed cloud computing environments that include any of the above systems or devices.
- the processor power optimization with response time assurance system 100 may include internal component 800 and external component 900 illustrated in FIG. 5 .
- Each internal component 800 includes one or more processors 820 , one or more computer-readable RAMs 822 and one or more computer-readable ROMs 824 on one or more buses 826 , and one or more operating systems 828 and one or more computer-readable tangible storage devices 830 .
- the one or more operating systems 828 and programs such as a processor power optimization with response time assurance program (i.e., PRA Program) given by processes 200 ( FIG. 2 ), 300 ( FIG. 3 ), and 400 ( FIG.
- PRA Program processor power optimization with response time assurance program
- each of the computer-readable tangible storage devices 830 is a magnetic disk storage device of an internal hard drive.
- each of the computer-readable tangible storage devices 830 is a semiconductor storage device such as ROM 824 , EPROM, flash memory or any other computer-readable tangible storage device that can store a computer program and digital information.
- Each internal component 800 also includes a R/W drive or interface 832 to read from and write to one or more portable computer-readable tangible storage devices 936 such as a CD-ROM, DVD, memory stick, magnetic tape, magnetic disk, optical disk or semiconductor storage device.
- the PRA Program associated with the processor power optimization with response time assurance system 100 can be stored on one or more of the respective portable computer-readable tangible storage devices 936 , read via the respective R/W drive or interface 832 and loaded into the respective hard drive 830 .
- Each internal component 800 may also include network adapters (or switch port cards) or interfaces 836 such as a TCP/IP adapter cards, wireless wi-fi interface cards, or 3G or 4G wireless interface cards or other wired or wireless communication links.
- the PRA Program associated with the processor power optimization with response time assurance system 100 can be downloaded from an external computer (e.g., server) via a network (for example, the Internet, a local area network or other, wide area network) and respective network adapters or interfaces 836 . From the network adapters (or switch port adaptors) or interfaces 836 , the PRA Program of the processor power optimization with response time assurance system 100 ( FIG. 1 ) is loaded into the respective hard drive 830 .
- the network may comprise copper wires, optical fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- Each external component 900 can include a computer display monitor 920 , a keyboard 930 , and a computer mouse 934 .
- External component 900 can also include touch screens, virtual keyboards, touch pads, pointing devices, and other human interface devices.
- Each internal component 800 also includes device drivers 840 to interface to computer display monitor 920 , keyboard 930 and computer mouse 934 .
- the device drivers 840 , R/W drive or interface 832 and network adapter or interface 836 comprise hardware and software (stored in storage device 830 and/or ROM 824 ).
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Power Sources (AREA)
Abstract
A method for managing processor power optimization is provided. The method may include receiving a plurality of tasks for processing by a processor environment. The method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
Description
This application is a continuation of U.S. patent application Ser. No. 14/185,974 filed on Feb. 21, 2014, which application is incorporated by reference herein.
Field of the Invention
The present invention relates generally to the field of computing, and more particularly to processor power management.
Background of the Related Art
Processor power management is an increasingly important strategy with the proliferation of computers. Reducing energy usage by processors provides many advantages, such as reducing costs, increasing component longevity and reducing heat production. For example, in a datacenter containing servers used in cloud computing, the energy required to power processors and attendant cooling systems may represent a significant operating expense. To reduce processor energy consumption, processor power usage mechanisms have been developed to dynamically control processor power usage.
Examples of some processor power usage mechanisms may include: dynamic voltage and frequency scaling (DVFS), processor core nap, and processor folding. Dynamic voltage scaling is a power usage mechanism where the voltage used in a processor can be increased or decreased dynamically. Dynamic frequency scaling is a power usage mechanism where the processor frequency can be increased or decreased dynamically. Lower voltage or lower frequency yields less power consumption by the processor. Processor core nap can be used to clock off most of the circuits inside a processor core when there is no work to be done by the core to save energy. This causes the core to go into a low-power idle state. Processor folding is a power usage mechanism where the number of available processor cores can be increased or decreased dynamically. Tasks are reallocated to the available cores and the other cores remain in low-power idle states.
These processor power usage mechanisms must be utilized to facilitate desired energy reduction goals. If, however, processor power usage mechanisms are aggressively used to reduce energy consumption, processor throughput may drop too low resulting in unacceptable task response time for a user.
According to one embodiment, a method for managing processor power optimization is provided. The method may include receiving a plurality of tasks for processing by a processor environment. The method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
According to another embodiment, a computer system for managing processor power optimization is provided. The computer system may include one or more processors, one or more computer-readable memories, one or more computer-readable tangible storage devices, and program instructions stored on at least one of the one or more storage devices for execution by at least one of the one or more processors via at least one of the one or more memories, wherein the computer system is capable of performing a method. The method may include receiving a plurality of tasks for processing by a processor environment. The method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
According to a further embodiment, a computer program product for managing processor power optimization is provided. The computer program product may include one or more computer-readable storage devices and program instructions stored on at least one of the one or more tangible storage devices, the program instructions executable by a processor. The computer program product may include receiving a plurality of tasks for processing by a processor environment. The method may also include allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
These and other objects, features and advantages of the present invention will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings. The various features of the drawings are not to scale as the illustrations are for clarity in facilitating one skilled in the art in understanding the invention in conjunction with the detailed description. In the drawings:
Detailed embodiments of the claimed structures and methods are disclosed herein; however, it can be understood that the disclosed embodiments are merely illustrative of the claimed structures and methods that may be embodied in various forms. This invention may, however, be embodied in many different forms and should not be construed as limited to the exemplary embodiments set forth herein. Rather, these exemplary embodiments are provided so that this disclosure will be thorough and complete and will fully convey the scope of this invention to those skilled in the art. In the description, details of well-known features and techniques may be omitted to avoid unnecessarily obscuring the presented embodiments.
Embodiments of the present invention relate generally to the field of computing, and more particularly to processor power management. The following described exemplary embodiments provide a system, method and program product for processor power optimization with response time assurance.
The present invention may be a system, a method, and/or a computer program product. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.)
These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
The following described exemplary embodiments provide a system, method and program product for processor power optimization with response time assurance.
Referring now to FIG. 1 , a system block diagram of a processor power optimization with response time assurance system 100 according to at least one embodiment is depicted. The processor power optimization with response time assurance system 100 may include a task classifier process unit 102, a class allocation process unit 104 and an energy manager process unit 106.
As depicted, according to one embodiment, the processor power optimization with response time assurance system 100 receives tasks 108 having execution time and response time characteristics. The task classifier process unit 102 may then assign each of the tasks 108 to one of class queues 110 a-110 c based on each tasks 108 execution time and response time characteristics.
According to one implementation, the task classifier process unit 102 may monitor header information associated with the received tasks 108. The task classifier process unit 102 may utilize the header information to indicate the task-type (e.g., content streaming, searches, payment transaction, etc.). Based on, for example, a lookup table entry, the task classifier process unit 102 may correlate each indicated task-type to one of the class queues 110 a-110 c. For example, received task A1 may be determined to be a content streaming task-type. The lookup table may accordingly assign class queue 110 b to the content streaming task-type, whereby class queue 110 b stores or buffers task-types having the same or similar execution time and response time characteristics to that of the content streaming task designated by A1.
The class allocation process unit 104 then allocates a portion of a compute resource 112 to each active class queue 110 for processing a task 108 within the desired response time. The compute resource 112 may be the total number of threads of execution (threads) supported by one or more processor cores contained within a processor environment 114. The processor environment 114 may be, for example, a single device or multiple interconnected devices where each device may contain one or more single core or multi-core processors. Examples of such devices may include a server, a desktop computer, a laptop, a telephone, a PDA, a tablet computer, or any other device containing one or more processors. The energy manager process unit 106 monitors the amount of compute resource utilization and increases or decreases available compute resource 112 as necessary through processor power usage mechanisms such as DVFS, processor core nap and processor folding to ensure compute resource utilization is within a target range.
Referring now to FIG. 2 , an operational flowchart illustrating the exemplary task classifier process 200 performed the task classifier process unit 102 (FIG. 1 ) according to at least one embodiment is depicted. Task classifier process 200 may be executed within the task classifier process unit 102 (FIG. 1 ). The task classifier process 200 begins at 202 by defining N different task classes. A class is created corresponding to each unique combination of execution time and response time characteristics for every task that the processor power optimization with response time assurance system 100 (FIG. 1 ) will process.
Next, at 204, a class queue 110 (FIG. 1 ) data structure may be created for each class defined. Where N classes were defined at 202, N class queues 110 (FIG. 1 ) may be created. The class queues 110 (FIG. 1 ) created may, according to one embodiment, operate under a first in first out (FIFO) policy. The class queues 110 (FIG. 1 ) may be implemented, for example, in software by using an array or linked list, or in hardware by using a buffer or registers.
Then, at 206, the task classifier process 200 checks for an arriving task 108 (FIG. 1 ) to process. If there is no new task 108 (FIG. 1 ) found, the task classifier process unit 102 may wait a short amount of time and check again for the arrival a new task 108 (FIG. 1 ).
However, if there is a new task 108 (FIG. 1 ) found, at 208, the new task 108 (FIG. 1 ) is classified based on the task 108 (FIG. 1 ) execution time and response time characteristics. Then, according to one embodiment, the task 108 (FIG. 1 ) is placed at the end of one the designated class queues 110 a-110 b (FIG. 1 ) corresponding to the class the task 108 (FIG. 1 ) was assigned. The task classifier process unit 102 then returns to 206 to check for arrival of any new tasks 108 (FIG. 1 ) to process.
With respect to FIG. 3 , an operational flowchart illustrating the exemplary class allocation process 300 performed by the class allocation process unit 104 (FIG. 1 ) according to at least one embodiment is depicted. Class allocation process 300 may be executed within the class allocation process unit 104 (FIG. 1 ). At 302, the class allocation process 300 specifies a normalized delay factor D for each class i associated with the received tasks 108 (FIG. 1 ). As previously indicated, each class i is assigned based on its respective response time characteristic. Classes having faster response time characteristics (e.g., streaming video) are assigned numerically higher delay factors Di, while classes having slower response times (e.g., payment transaction) are assigned numerically lower delay factors Di. For example, class 1 (i.e., i=1) may require a faster response time and be assigned a normalized delay factor D1=25. Alternatively, class 2 (i.e., i=2) may tolerate a slower response time and be assigned a normalized delay factor D2=5.
At 304, the class allocation process 300 specifies a normalized execution time E for each class i associated with the received tasks 108 (FIG. 1 ). As previously indicated, each class i is assigned based on its respective execution time characteristic. Classes requiring more time to execute (e.g., payment transaction) are assigned numerically higher execution times Ei, while classes requiring less time to execute (e.g., streaming video) are assigned numerically lower execution times Ei. For example, class 1 (i.e., i=1) may require a longer execution time and be assigned a normalized execution time E1=4. Alternatively, class 2 (i.e., i=2) may require a shorter execution time and be assigned a normalized delay factor E2=1.
Next, at 306, the compute resource 112 (FIG. 1 ) available for all task classes at the current time t is specified as P(t), where P(t) is less than or equal to the maximum processing capability of the compute resource 112 (FIG. 1 ). The value P(t) comprises, for example, the total number of threads that can be processed by all of the active cores. For example, a server may have (15) processors, where each processor includes (20) cores each supporting (4) threads. Thus, the total number of threads for the server is P(t)=15×20×4=1200.
At 308, the class allocation process 300 calculates the amount of compute resource 112 (FIG. 1 ) each class will be allocated in proportion to the delay factor D and execution time E of each class i. According to one embodiment, the compute resource allocation factor A for each class i may be determined by the following formula:
The value of E in the compute resource allocation factor formula is replaced with 0 for classes that have no tasks currently within the class queue 110 (
An illustrative example of the process at 308, for a server having three active classes, the compute resource allocation factor formula is computed three times, once for every active class. The three active classes in the example server may be: class 1 corresponding to search tasks where D1=5 and E1=8, class 2 corresponding to payment transaction tasks where D2=3 and E2=4, and class 3 corresponding to content streaming tasks where D3=10 and E3=1. The resource allocation factor is computed for class 1 as follows:
The resource allocation factor is then computed for
Then, at 310, the compute resource 112 (FIG. 1 ) is allocated to each class corresponding to the proportion determined at 308. The compute resource 112 (FIG. 1 ) processes all active classes simultaneously but each class may be processed at different rates. According to one implementation, tasks from each class i are allocated a portion of the compute resource R using the following formula:
R i =A i ×P(t)
In other words, the amount of total compute resource available P(t) (e.g., total number of available threads) when multiplied by the resource allocation factor Ai results in the amount of compute resource Ri number of threads) that may be used by class i to process tasks 108 (FIG. 1 ). For example, a server capable of supporting 1200 threads at the current time (i.e., P(t)=1200), having a class 1 with a compute resource allocation factor A1=0.645, will result in 774 threads (0.645×1200=774) allocated to class 1 to process tasks 108 (FIG. 1 ). Similarly, all remaining active classes will be allocated compute resource 112 (FIG. 1 ) using the same calculation.
R i =A i ×P(t)
In other words, the amount of total compute resource available P(t) (e.g., total number of available threads) when multiplied by the resource allocation factor Ai results in the amount of compute resource Ri number of threads) that may be used by class i to process tasks 108 (
At 312, according to one implementation, the class allocation process 300 checks if the energy manager process unit 106 (FIG. 1 ) has sent a notification to reapportion the compute resource 112 (FIG. 1 ). If a notification to reapportion compute resources is received, the class allocation process will go to 306. However, if no notification is received, the class allocation process 300 may recheck until a notification is received. For example, if one task (e.g., task of class 1) completes, the energy manager process within the energy manager process unit 106 (FIG. 1 ) initiates a reapportioning of the compute resources by sending a notification to the class allocation process 300. In addition to sending the notification, the energy manager process may also reduce the total available compute resource in accordance with the number of tasks that are actively being processed.
With respect to FIG. 4 , an operational flowchart illustrating the exemplary energy manager process 400 performed by the energy manager process unit 106 (FIG. 1 ) according to at least one embodiment is depicted. Energy manager process 400 may be executed within the energy manager process unit 106 (FIG. 1 ). At 402, a target compute resource utilization range is defined by designating two limits. The target compute resource utilization range extends between a lower limit UL representing the minimum target resource utilization and an upper limit UH representing the maximum target resource utilization. An example target utilization range may be defined by a lower limit UL=90% and an upper limit UH=95%.
Next, at 404, according to one embodiment, the compute resource utilization level is initially set to full utilization. This may be done, for example, by making all of the processors and processor cores active and running at full speed in the processor environment 114 (FIG. 1 ). Full compute resource utilization is made available to ensure that the initial set of tasks 108 (FIG. 1 ) are supported until compute resource utilization level can be reliably measured.
Then, at 406, the compute resource utilization level U is measured at the current time t. Next, at 408, the current utilization U(t) is compared with the lower limit UL. If the current utilization U(t) is lower than the lower limit UL, then there is too much active compute resource not being utilized that may in turn be consuming unnecessary energy.
If current utilization U(t) is not lower than the lower limit UL, then at 410, the current utilization U(t) is compared with the upper limit UH. If current utilization U(t) is higher than the upper limit UH, then there is not enough active or available compute resources. However, if current utilization U(t) is not higher than the upper limit UH, then current utilization U(t) is within the target compute resource utilization range and there is no need to change the available compute resource and the energy manager process 400 returns to 406.
For example, an energy manager process 400 at 408 and 410 may have a measured current utilization U(t)=93% and a target resource utilization range defined by a lower limit UL=90% and an upper limit UH=95%. The energy manager process 400, at 408, checks if U(t) (93%) is less than UL (90%). In this example, U(t) is not lower than UL and the process 400 continues to 410 to check if U(t) (93%) is higher than UH (95%). Here, U(t) is not higher than UH, thus the energy manager process 400 has determined that U(t) falls within the target resource utilization range.
At 412, the energy manager process 400 responds to the condition where current utilization U(t) is lower than the lower limit UL (“yes” at 408) by reducing available compute resources through processor power management mechanisms. For example, such mechanisms may include utilizing DVFS, processor core nap and processor folding to scale down processor core frequency or voltage, or put processor cores into low-power idle states to reduce available compute resource 112 (FIG. 1 ) and also reduce energy usage. One example for triggering a reduction in computer resource may occur when one or more tasks, as indicated by T1A and T1B (FIG. 1 ), in the class queue (e.g., FIG. 1 : queue 110 a) are completed.
At 414, the energy manager process 400 responds to the condition where current utilization U(t) is higher than the upper limit UH (“yes” at 410) by increasing available compute resources through processor power management mechanisms. For example, such mechanisms may include utilizing DVFS, processor core nap and processor folding to reactivate processor cores from low-power idle states or scale processor core frequency or voltage up to increase available compute resource 112 (FIG. 1 ) for processing tasks 108 (FIG. 1 ). One example for triggering an increase in computer resource may occur when one or more tasks, as indicated by T1A and T1B (FIG. 1 ), are added to the class queue (e.g., FIG. 1 : queue 110 a).
Next, at 416, according to one implementation, a notification is sent to the class allocation process 300 associated with class allocation process unit 104 (FIG. 1 ) to reapportion the compute resource because the amount of compute resource available has changed. The notification system may be implemented in a variety of ways, such as, for example, a function call or as a flag variable set to a predefined value. Once the notification is sent, the energy manager process 400 returns to 406.
The processor power optimization with response time assurance system 100 (FIG. 1 ) may include internal component 800 and external component 900 illustrated in FIG. 5 . Each internal component 800 includes one or more processors 820, one or more computer-readable RAMs 822 and one or more computer-readable ROMs 824 on one or more buses 826, and one or more operating systems 828 and one or more computer-readable tangible storage devices 830. The one or more operating systems 828 and programs such as a processor power optimization with response time assurance program (i.e., PRA Program) given by processes 200 (FIG. 2 ), 300 (FIG. 3 ), and 400 (FIG. 4 ) may be stored on one or more computer-readable tangible storage devices 830 for execution by one or more processors 820 via one or more RAMs 822 (which typically include cache memory). In the embodiment illustrated in FIG. 5 , each of the computer-readable tangible storage devices 830 is a magnetic disk storage device of an internal hard drive. Alternatively, each of the computer-readable tangible storage devices 830 is a semiconductor storage device such as ROM 824, EPROM, flash memory or any other computer-readable tangible storage device that can store a computer program and digital information.
Each internal component 800 also includes a R/W drive or interface 832 to read from and write to one or more portable computer-readable tangible storage devices 936 such as a CD-ROM, DVD, memory stick, magnetic tape, magnetic disk, optical disk or semiconductor storage device. The PRA Program associated with the processor power optimization with response time assurance system 100 (FIG. 1 ) can be stored on one or more of the respective portable computer-readable tangible storage devices 936, read via the respective R/W drive or interface 832 and loaded into the respective hard drive 830.
Each internal component 800 may also include network adapters (or switch port cards) or interfaces 836 such as a TCP/IP adapter cards, wireless wi-fi interface cards, or 3G or 4G wireless interface cards or other wired or wireless communication links. The PRA Program associated with the processor power optimization with response time assurance system 100 (FIG. 1 ), can be downloaded from an external computer (e.g., server) via a network (for example, the Internet, a local area network or other, wide area network) and respective network adapters or interfaces 836. From the network adapters (or switch port adaptors) or interfaces 836, the PRA Program of the processor power optimization with response time assurance system 100 (FIG. 1 ) is loaded into the respective hard drive 830. The network may comprise copper wires, optical fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
Each external component 900 can include a computer display monitor 920, a keyboard 930, and a computer mouse 934. External component 900 can also include touch screens, virtual keyboards, touch pads, pointing devices, and other human interface devices. Each internal component 800 also includes device drivers 840 to interface to computer display monitor 920, keyboard 930 and computer mouse 934. The device drivers 840, R/W drive or interface 832 and network adapter or interface 836 comprise hardware and software (stored in storage device 830 and/or ROM 824).
The descriptions of the various embodiments of the present invention have been presented for purposes of illustration, but are not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement over technologies found in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.
Claims (20)
1. A method for managing processor power optimization, the method comprising:
receiving a plurality of tasks for processing by a processor environment; and
allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
2. The method of claim 1 , further comprising:
classifying the received plurality of tasks into a plurality of classes, wherein each of the plurality of classes includes a corresponding resource allocation factor that is determined based on the execution time and the response time associated with each of the received plurality of tasks.
3. The method of claim 2 , wherein the allocating of the portion of the compute resource corresponding to the processor environment comprises allocating the corresponding resource allocation factor to each of the plurality of classes.
4. The method of claim 1 , wherein the processor environment comprises one of a single processor multiple core device, a single processor single core device, a multiprocessor multiple core device, and a multiprocessor single core device.
5. The method of claim 1 , wherein the processor environment comprises a distributed computer environment.
6. The method of claim 1 , wherein the execution time corresponds to a length of time required for completing a received task.
7. The method of claim 1 , wherein the response time corresponds to a delay sensitivity of a received task.
8. The method of claim 1 , wherein the received plurality of tasks comprise a search process, a payment transaction, multi-media data streaming, and a file transfer process.
9. A computer system for managing processor power optimization, comprising:
a memory;
a processor in communication with the memory, the processor comprising an instruction fetching unit for fetching instructions from memory and one or more execution units for executing fetched instructions;
wherein said computer system is capable of performing a method comprising:
receiving a plurality of tasks for processing by a processor environment; and
allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
10. The computer system of claim 9 , further comprising:
classifying the received plurality of tasks into a plurality of classes, wherein each of the plurality of classes includes a corresponding resource allocation factor that is determined based on the execution time and the response time associated with each of the received plurality of tasks.
11. The computer system of claim 10 , wherein the allocating of the portion of the compute resource corresponding to the processor environment comprises allocating the corresponding resource allocation factor to each of the plurality of classes.
12. The computer system of claim 9 , wherein the processor environment comprises one of a single processor multiple core device, a single processor single core device, a multiprocessor multiple core device, and a multiprocessor single core device.
13. The computer system of claim 9 , wherein the processor environment comprises a distributed computer environment.
14. The computer system of claim 9 , wherein the execution time corresponds to a length of time required for completing a received task.
15. The computer system of claim 9 , wherein the response time corresponds to a delay sensitivity of a received task.
16. The computer system of claim 9 , wherein the received plurality of tasks comprise a search process, a payment transaction, multi-media data streaming, and a file transfer process.
17. A computer program product for managing processor power optimization, comprising:
one or more computer-readable storage devices and program instructions stored on at least one of the one or more tangible storage devices, the program instructions executable by a processor, the program instructions comprising:
receiving a plurality of tasks for processing by a processor environment; and
allocating a portion of a compute resource corresponding to the processor environment to each of the received plurality of tasks, the allocating of the portion being based on both an execution time and a response time associated with each of the received plurality of tasks.
18. The computer program product of claim 17 , further comprising:
classifying the received plurality of tasks into a plurality of classes, wherein each of the plurality of classes includes a corresponding resource allocation factor that is determined based on the execution time and the response time associated with each of the received plurality of tasks.
19. The computer program product of claim 18 , wherein the allocating of the portion of the compute resource corresponding to the processor environment comprises allocating the corresponding resource allocation factor to each of the plurality of classes.
20. The computer program product of claim 17 , wherein the processor environment comprises one of a single processor multiple core device, a single processor single core device, a multiprocessor multiple core device, and a multiprocessor single core device.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/070,792 US9841998B2 (en) | 2014-02-21 | 2016-03-15 | Processor power optimization with response time assurance |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/185,974 US9323574B2 (en) | 2014-02-21 | 2014-02-21 | Processor power optimization with response time assurance |
US15/070,792 US9841998B2 (en) | 2014-02-21 | 2016-03-15 | Processor power optimization with response time assurance |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/185,974 Continuation US9323574B2 (en) | 2014-02-21 | 2014-02-21 | Processor power optimization with response time assurance |
Publications (2)
Publication Number | Publication Date |
---|---|
US20160196165A1 US20160196165A1 (en) | 2016-07-07 |
US9841998B2 true US9841998B2 (en) | 2017-12-12 |
Family
ID=53882297
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/185,974 Active 2034-08-24 US9323574B2 (en) | 2014-02-21 | 2014-02-21 | Processor power optimization with response time assurance |
US15/070,792 Active US9841998B2 (en) | 2014-02-21 | 2016-03-15 | Processor power optimization with response time assurance |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/185,974 Active 2034-08-24 US9323574B2 (en) | 2014-02-21 | 2014-02-21 | Processor power optimization with response time assurance |
Country Status (1)
Country | Link |
---|---|
US (2) | US9323574B2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11782870B2 (en) * | 2019-09-09 | 2023-10-10 | Shanghai Denglin Technologies Co., Ltd. | Configurable heterogeneous AI processor with distributed task queues allowing parallel task execution |
US11789895B2 (en) * | 2019-09-09 | 2023-10-17 | Shanghai Denglin Technologies Co., Ltd. | On-chip heterogeneous AI processor with distributed tasks queues allowing for parallel task execution |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9436608B1 (en) * | 2015-02-12 | 2016-09-06 | International Business Machines Corporation | Memory nest efficiency with cache demand generation |
US10699211B2 (en) | 2016-02-29 | 2020-06-30 | Oracle International Corporation | Supervised method for classifying seasonal patterns |
US10331802B2 (en) | 2016-02-29 | 2019-06-25 | Oracle International Corporation | System for detecting and characterizing seasons |
US10867421B2 (en) | 2016-02-29 | 2020-12-15 | Oracle International Corporation | Seasonal aware method for forecasting and capacity planning |
US10885461B2 (en) | 2016-02-29 | 2021-01-05 | Oracle International Corporation | Unsupervised method for classifying seasonal patterns |
US10296067B2 (en) | 2016-04-08 | 2019-05-21 | Qualcomm Incorporated | Enhanced dynamic clock and voltage scaling (DCVS) scheme |
US10198339B2 (en) | 2016-05-16 | 2019-02-05 | Oracle International Corporation | Correlation-based analytic for time-series data |
US11082439B2 (en) | 2016-08-04 | 2021-08-03 | Oracle International Corporation | Unsupervised method for baselining and anomaly detection in time-series data for enterprise systems |
US10635563B2 (en) | 2016-08-04 | 2020-04-28 | Oracle International Corporation | Unsupervised method for baselining and anomaly detection in time-series data for enterprise systems |
US11461513B2 (en) * | 2016-08-18 | 2022-10-04 | Cato | Data center power scenario simulation |
US10915830B2 (en) | 2017-02-24 | 2021-02-09 | Oracle International Corporation | Multiscale method for predictive alerting |
US10949436B2 (en) | 2017-02-24 | 2021-03-16 | Oracle International Corporation | Optimization for scalable analytics using time series models |
US10817803B2 (en) | 2017-06-02 | 2020-10-27 | Oracle International Corporation | Data driven methods and systems for what if analysis |
US10621005B2 (en) * | 2017-08-31 | 2020-04-14 | Oracle International Corporation | Systems and methods for providing zero down time and scalability in orchestration cloud services |
CN109918192B (en) * | 2017-12-25 | 2022-02-18 | 郑州云海信息技术有限公司 | BMC resource allocation method for server |
CN110543148B (en) * | 2018-05-28 | 2021-04-09 | 华为技术有限公司 | Task scheduling method and device |
US10997517B2 (en) | 2018-06-05 | 2021-05-04 | Oracle International Corporation | Methods and systems for aggregating distribution approximations |
US10963346B2 (en) | 2018-06-05 | 2021-03-30 | Oracle International Corporation | Scalable methods and systems for approximating statistical distributions |
US11138090B2 (en) | 2018-10-23 | 2021-10-05 | Oracle International Corporation | Systems and methods for forecasting time series with variable seasonality |
US12001926B2 (en) | 2018-10-23 | 2024-06-04 | Oracle International Corporation | Systems and methods for detecting long term seasons |
US10855548B2 (en) | 2019-02-15 | 2020-12-01 | Oracle International Corporation | Systems and methods for automatically detecting, summarizing, and responding to anomalies |
US11533326B2 (en) | 2019-05-01 | 2022-12-20 | Oracle International Corporation | Systems and methods for multivariate anomaly detection in software monitoring |
US11537940B2 (en) | 2019-05-13 | 2022-12-27 | Oracle International Corporation | Systems and methods for unsupervised anomaly detection using non-parametric tolerance intervals over a sliding window of t-digests |
US11887015B2 (en) | 2019-09-13 | 2024-01-30 | Oracle International Corporation | Automatically-generated labels for time series data and numerical lists to use in analytic and machine learning systems |
US11886224B2 (en) * | 2020-06-26 | 2024-01-30 | Advanced Micro Devices, Inc. | Core selection based on usage policy and core constraints |
Citations (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5381546A (en) * | 1987-04-13 | 1995-01-10 | Gte Laboratories Incorporated | Control process for allocating services in communications systems |
US6105051A (en) * | 1997-10-23 | 2000-08-15 | International Business Machines Corporation | Apparatus and method to guarantee forward progress in execution of threads in a multithreaded processor |
US6289488B1 (en) | 1997-02-24 | 2001-09-11 | Lucent Technologies Inc. | Hardware-software co-synthesis of hierarchical heterogeneous distributed embedded systems |
US6567839B1 (en) * | 1997-10-23 | 2003-05-20 | International Business Machines Corporation | Thread switch control in a multithreaded processor system |
EP1365312A1 (en) | 2002-05-20 | 2003-11-26 | Texas Instruments Incorporated | Energy-aware scheduling of application execution |
US20040111506A1 (en) | 2002-12-10 | 2004-06-10 | International Business Machines Corporation | System and method for managing web utility services |
US7487319B2 (en) | 2004-11-18 | 2009-02-03 | International Business Machines Corporation | Resource allocation unit queue |
US7721286B2 (en) * | 1993-09-21 | 2010-05-18 | Microsoft Corporation | Preemptive multi-tasking with cooperative groups of tasks |
US7930565B2 (en) * | 2006-09-11 | 2011-04-19 | Texas Instruments Incorporated | Power-optimizing memory analyzer, method of operating the analyzer and system employing the same |
US8046605B2 (en) | 2006-11-08 | 2011-10-25 | International Business Machines Corporation | Computer system management and throughput maximization in the presence of power constraints |
US8140371B2 (en) * | 2005-02-18 | 2012-03-20 | International Business Machines Corporation | Providing computing service to users in a heterogeneous distributed computing environment |
US8161482B1 (en) * | 2007-04-13 | 2012-04-17 | Marvell International Ltd. | Power optimization for multi-core devices |
US8214829B2 (en) | 2009-01-15 | 2012-07-03 | International Business Machines Corporation | Techniques for placing applications in heterogeneous virtualized systems while minimizing power and migration cost |
US20120180061A1 (en) | 2011-01-10 | 2012-07-12 | International Business Machines Corporation | Organizing Task Placement Based On Workload Characterizations |
US8381004B2 (en) | 2010-05-26 | 2013-02-19 | International Business Machines Corporation | Optimizing energy consumption and application performance in a multi-core multi-threaded processor system |
US8458720B2 (en) * | 2007-08-17 | 2013-06-04 | International Business Machines Corporation | Methods and systems for assigning non-continual jobs to candidate processing nodes in a stream-oriented computer system |
US8484492B2 (en) | 2009-11-09 | 2013-07-09 | Pegatron Corporation | Electronic device and power control module for controlling a starting process of the electronic device utilizing a delay latch circuit and an inverse logic unit |
US8484495B2 (en) | 2010-03-25 | 2013-07-09 | International Business Machines Corporation | Power management in a multi-processor computer system |
US8510741B2 (en) | 2007-03-28 | 2013-08-13 | Massachusetts Institute Of Technology | Computing the processor desires of jobs in an adaptively parallel scheduling environment |
US8527994B2 (en) * | 2011-02-10 | 2013-09-03 | International Business Machines Corporation | Guarded, multi-metric resource control for safe and efficient microprocessor management |
US20130283289A1 (en) | 2012-04-19 | 2013-10-24 | International Business Machines Corporation | Environmentally aware load-balancing |
US8578386B1 (en) | 2009-04-01 | 2013-11-05 | Netapp, Inc. | Processor scheduling method and system using domains |
US8589932B2 (en) * | 2010-07-02 | 2013-11-19 | International Business Machines Corporation | Data processing workload control |
US8615763B2 (en) | 2012-02-01 | 2013-12-24 | Texas Instruments Incorporated | System and method of task allocation in multiprocessing environment based on minimum density or maximum harmonicity that results in lower clock frequency |
US8621472B2 (en) * | 2009-11-03 | 2013-12-31 | International Business Machines Corporation | Job scheduling with optimization of power consumption |
US8635625B2 (en) | 2008-04-04 | 2014-01-21 | International Business Machines Corporation | Power-aware workload allocation in performance-managed computing environments |
US8661443B2 (en) * | 2009-04-17 | 2014-02-25 | Empire Technology Development, Llc | Scheduling and/or organizing task execution for a target computing platform |
US8745629B2 (en) | 2010-01-11 | 2014-06-03 | Qualcomm Incorporated | System and method of controlling power in an electronic device |
US8782653B2 (en) | 2010-03-26 | 2014-07-15 | Virtualmetrix, Inc. | Fine grain performance resource management of computer systems |
US8869157B2 (en) | 2012-06-21 | 2014-10-21 | Breakingpoint Systems, Inc. | Systems and methods for distributing tasks and/or processing recources in a system |
US8887165B2 (en) * | 2010-02-19 | 2014-11-11 | Nec Corporation | Real time system task configuration optimization system for multi-core processors, and method and program |
US8954487B2 (en) | 2009-06-23 | 2015-02-10 | Samsung Electronics Co., Ltd. | Management server and method for providing cloud computing service |
US8959370B2 (en) * | 2008-10-03 | 2015-02-17 | University Of Sydney | Scheduling an application for performance on a heterogeneous computing system |
-
2014
- 2014-02-21 US US14/185,974 patent/US9323574B2/en active Active
-
2016
- 2016-03-15 US US15/070,792 patent/US9841998B2/en active Active
Patent Citations (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5381546A (en) * | 1987-04-13 | 1995-01-10 | Gte Laboratories Incorporated | Control process for allocating services in communications systems |
US7721286B2 (en) * | 1993-09-21 | 2010-05-18 | Microsoft Corporation | Preemptive multi-tasking with cooperative groups of tasks |
US6289488B1 (en) | 1997-02-24 | 2001-09-11 | Lucent Technologies Inc. | Hardware-software co-synthesis of hierarchical heterogeneous distributed embedded systems |
US6105051A (en) * | 1997-10-23 | 2000-08-15 | International Business Machines Corporation | Apparatus and method to guarantee forward progress in execution of threads in a multithreaded processor |
US6567839B1 (en) * | 1997-10-23 | 2003-05-20 | International Business Machines Corporation | Thread switch control in a multithreaded processor system |
US8032891B2 (en) * | 2002-05-20 | 2011-10-04 | Texas Instruments Incorporated | Energy-aware scheduling of application execution |
EP1365312A1 (en) | 2002-05-20 | 2003-11-26 | Texas Instruments Incorporated | Energy-aware scheduling of application execution |
US20040111506A1 (en) | 2002-12-10 | 2004-06-10 | International Business Machines Corporation | System and method for managing web utility services |
US7487319B2 (en) | 2004-11-18 | 2009-02-03 | International Business Machines Corporation | Resource allocation unit queue |
US8140371B2 (en) * | 2005-02-18 | 2012-03-20 | International Business Machines Corporation | Providing computing service to users in a heterogeneous distributed computing environment |
US7930565B2 (en) * | 2006-09-11 | 2011-04-19 | Texas Instruments Incorporated | Power-optimizing memory analyzer, method of operating the analyzer and system employing the same |
US8046605B2 (en) | 2006-11-08 | 2011-10-25 | International Business Machines Corporation | Computer system management and throughput maximization in the presence of power constraints |
US8510741B2 (en) | 2007-03-28 | 2013-08-13 | Massachusetts Institute Of Technology | Computing the processor desires of jobs in an adaptively parallel scheduling environment |
US8161482B1 (en) * | 2007-04-13 | 2012-04-17 | Marvell International Ltd. | Power optimization for multi-core devices |
US8458720B2 (en) * | 2007-08-17 | 2013-06-04 | International Business Machines Corporation | Methods and systems for assigning non-continual jobs to candidate processing nodes in a stream-oriented computer system |
US8635625B2 (en) | 2008-04-04 | 2014-01-21 | International Business Machines Corporation | Power-aware workload allocation in performance-managed computing environments |
US8959370B2 (en) * | 2008-10-03 | 2015-02-17 | University Of Sydney | Scheduling an application for performance on a heterogeneous computing system |
US8214829B2 (en) | 2009-01-15 | 2012-07-03 | International Business Machines Corporation | Techniques for placing applications in heterogeneous virtualized systems while minimizing power and migration cost |
US8578386B1 (en) | 2009-04-01 | 2013-11-05 | Netapp, Inc. | Processor scheduling method and system using domains |
US8661443B2 (en) * | 2009-04-17 | 2014-02-25 | Empire Technology Development, Llc | Scheduling and/or organizing task execution for a target computing platform |
US8954487B2 (en) | 2009-06-23 | 2015-02-10 | Samsung Electronics Co., Ltd. | Management server and method for providing cloud computing service |
US8621472B2 (en) * | 2009-11-03 | 2013-12-31 | International Business Machines Corporation | Job scheduling with optimization of power consumption |
US8484492B2 (en) | 2009-11-09 | 2013-07-09 | Pegatron Corporation | Electronic device and power control module for controlling a starting process of the electronic device utilizing a delay latch circuit and an inverse logic unit |
US8745629B2 (en) | 2010-01-11 | 2014-06-03 | Qualcomm Incorporated | System and method of controlling power in an electronic device |
US8887165B2 (en) * | 2010-02-19 | 2014-11-11 | Nec Corporation | Real time system task configuration optimization system for multi-core processors, and method and program |
US8484495B2 (en) | 2010-03-25 | 2013-07-09 | International Business Machines Corporation | Power management in a multi-processor computer system |
US8782653B2 (en) | 2010-03-26 | 2014-07-15 | Virtualmetrix, Inc. | Fine grain performance resource management of computer systems |
US8381004B2 (en) | 2010-05-26 | 2013-02-19 | International Business Machines Corporation | Optimizing energy consumption and application performance in a multi-core multi-threaded processor system |
US8589932B2 (en) * | 2010-07-02 | 2013-11-19 | International Business Machines Corporation | Data processing workload control |
US20120180061A1 (en) | 2011-01-10 | 2012-07-12 | International Business Machines Corporation | Organizing Task Placement Based On Workload Characterizations |
US8527994B2 (en) * | 2011-02-10 | 2013-09-03 | International Business Machines Corporation | Guarded, multi-metric resource control for safe and efficient microprocessor management |
US8615763B2 (en) | 2012-02-01 | 2013-12-24 | Texas Instruments Incorporated | System and method of task allocation in multiprocessing environment based on minimum density or maximum harmonicity that results in lower clock frequency |
US20130283289A1 (en) | 2012-04-19 | 2013-10-24 | International Business Machines Corporation | Environmentally aware load-balancing |
US8869157B2 (en) | 2012-06-21 | 2014-10-21 | Breakingpoint Systems, Inc. | Systems and methods for distributing tasks and/or processing recources in a system |
Non-Patent Citations (19)
Title |
---|
"USPTO-Notice of Allowance" U.S. Appl. No. 14/185,974, dated Feb. 29, 2016; 12 pages. |
"USPTO-Office Action" U.S. Appl. No. 14/185,974, dated Oct. 22, 2015; 16 pages. |
"USPTO—Notice of Allowance" U.S. Appl. No. 14/185,974, dated Feb. 29, 2016; 12 pages. |
"USPTO—Office Action" U.S. Appl. No. 14/185,974, dated Oct. 22, 2015; 16 pages. |
Brylski et al, "Optimization of the FPGA Parallel Image Processor", ACM, pp. 183-188, 2012. * |
Chatziparaskevas et al, "An FPGA-based Parallel Processor for Black-Scholes Option Pricing Using Finite Differences Schemes", ACM, pp. 1-6, 2012. (This document only identifies the year of publication (without the month), but the year of publication is sufficiently earlier than the effective U.S. filing date and any foreign priority date so that the particular month of publication is not in issue. See MPEP 609.04(a)). |
Corbalan et ai, "Performance-Driven Processor Allocation", ACM, pp. 1-13, 2000. (This document only identifies the year of publication (without the month), but the year of publication is sufficiently earlier than the effective U.S. filing date and any foreign priority date so that the particular month of publication is not in issue. See MPEP 609.04(a)). |
Eswari et al. "Expected Completion Time based Scheduling Algorithm for Heterogeneous Processors", 2011 International Conference on Information Communication and Management IPCSIT vol. 16 (2011) IACSIT Press Singapore. pp. 72-77. (This document only identifies the year of publication (without the month), but the year of publication is sufficiently earlier than the effective U.S. filing date and any foreign priority date so that the particular month of publication is not in issue. See MPEP 609.04(a)). |
Ghasemzadeh et al, "Energy-Efficient Signal Processing in Wearable Embedded Systems: An Optimal Feature Selection Approach", ACM, pp. 357-362, 2012. * |
Gorjiara et al, "Ultra-Fast and Efficient Algorithm for Energy Optimization by Gradient-Based Stochastic Voltage and Task Scheduling", ACM Transactions on Design Automation of Electronic Systems, vol. 12, No. 4, Article 39, pp. 1-20, 2007. * |
Hanumaiah et ai, "Throughput Optimal Task Allocation under Thermal Constraints for Multi-core Processors" ACM, pp. 776-781, Jul. 26-31, 2009. |
Holmbacka et al, "A Run-Time Power Manager Exploiting Software Parallelism", ACM, pp. 471-472, 2014. * |
Ishfaq Ahmad et al., "Using Game Therory for Scheduling Tasks on Multi-Core Processors for Simultaneous Optimization of Performance and Energy", 2008 IEEE. 6 pages. (This document only identifies the year of publication (without the month), but the year of publication is sufficiently earlier than the effective U.S. filing date and any foreign priority date so that the particular month of publication is not in issue. See MPEP 609.04(a)). |
Papaefstathiou et ai, Software processing performance in network processors, IEEE, pp. 1-6, 2004. (This document only identifies the year of publication (without the month), but the year of publication is sufficiently earlier than the effective U.S. filing date and any foreign priority date so that the particular month of publication is not in issue. See MPEP 609.04(a)). |
Shih et ai, "A performance goal oriented processor allocation technique for centralized heterogeneous multi-cluster environments", IEEE, pp. 614-615, 2011 (This document only identifies the year of publication (without the month), but the year of publication is sufficiently earlier than the effective U.S. filing date and any foreign priority date so that the particular month of publication is not in issue. See MPEP 609.04(a)). |
Shin et al, "Leakage Power Minimization for the Synthesis of Parallel Multiplier Circuits", ACM, pp. 166-169, 2004. * |
Topcuoglu, H. et al., "Performance-effective and low-complexity task scheduling for heterogeneous computing", 2002 IEEE. Transactions on Parallel and Distributed Systems, vol. 13, No. 3, Mar. 2002, pp. 260-274. |
Wang et ai, "Optimizing Software Cache Performance of Packet Processing Applications ", ACM, pp. 227-235, Jun. 13-15, 2007. |
Zhou et al, "Power Signal Processing: A New Perspective for Power Analysis and Optimization", ACM, pp. 165-170, 2007. * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11782870B2 (en) * | 2019-09-09 | 2023-10-10 | Shanghai Denglin Technologies Co., Ltd. | Configurable heterogeneous AI processor with distributed task queues allowing parallel task execution |
US11789895B2 (en) * | 2019-09-09 | 2023-10-17 | Shanghai Denglin Technologies Co., Ltd. | On-chip heterogeneous AI processor with distributed tasks queues allowing for parallel task execution |
Also Published As
Publication number | Publication date |
---|---|
US20160196165A1 (en) | 2016-07-07 |
US9323574B2 (en) | 2016-04-26 |
US20150242243A1 (en) | 2015-08-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9841998B2 (en) | Processor power optimization with response time assurance | |
US10241843B2 (en) | Application processing allocation in a computing system | |
US9733965B2 (en) | Closed-loop feedback mechanism for achieving optimum performance in a consolidated workload environment | |
US9342376B2 (en) | Method, system, and device for dynamic energy efficient job scheduling in a cloud computing environment | |
US9760392B1 (en) | Adaptive throttling in hybrid storage environments | |
CN111247515A (en) | Apparatus and method for providing a performance-based packet scheduler | |
US9372725B2 (en) | Dynamically adjusting wait periods according to system performance | |
EP3961395A1 (en) | Method and apparatus for scheduling memory access request, device and storage medium | |
US9753851B2 (en) | Multi-section garbage collection system including real-time garbage collection scheduling | |
US10579416B2 (en) | Thread interrupt offload re-prioritization | |
US9817696B2 (en) | Low latency scheduling on simultaneous multi-threading cores | |
US9514072B1 (en) | Management of allocation for alias devices | |
US9471387B2 (en) | Scheduling in job execution | |
US20120102499A1 (en) | OPTIMIZING THE PERFORMANCE OF HYBRID CPU SYSTEMS BASED UPON THE THREAD TYPE OF APPLICATIONS TO BE RUN ON THE CPUs | |
US20160019089A1 (en) | Method and system for scheduling computing | |
CN110851276A (en) | Service request processing method, device, server and storage medium | |
US8862786B2 (en) | Program execution with improved power efficiency | |
US20220214965A1 (en) | System and method for storage class memory tiering | |
US20140281322A1 (en) | Temporal Hierarchical Tiered Data Storage | |
US10198262B2 (en) | Adaptive core grouping | |
US8543687B2 (en) | Moving deployment of images between computers | |
CN113204382B (en) | Data processing method, device, electronic equipment and storage medium | |
US9176910B2 (en) | Sending a next request to a resource before a completion interrupt for a previous request | |
US20240202159A1 (en) | Quality of service (qos) control of processor applications | |
Kang et al. | Mcredit2: Enhanced High‐Performance Xen Scheduler via Dynamic Weight Allocation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: LENOVO GLOBAL TECHNOLOGIES INTERNATIONAL LTD, HONG Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE LTD.;REEL/FRAME:050308/0781 Effective date: 20180101 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |