US10698472B2 - Instruction subset implementation for low power operation - Google Patents

Instruction subset implementation for low power operation Download PDF

Info

Publication number
US10698472B2
US10698472B2 US15/795,719 US201715795719A US10698472B2 US 10698472 B2 US10698472 B2 US 10698472B2 US 201715795719 A US201715795719 A US 201715795719A US 10698472 B2 US10698472 B2 US 10698472B2
Authority
US
United States
Prior art keywords
processor
feature
instruction
low
instruction thread
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US15/795,719
Other versions
US20190129489A1 (en
Inventor
Elliot H. Mednick
Edward McLellan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Advanced Micro Devices Inc
Original Assignee
Advanced Micro Devices Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Advanced Micro Devices Inc filed Critical Advanced Micro Devices Inc
Priority to US15/795,719 priority Critical patent/US10698472B2/en
Assigned to ADVANCED MICRO DEVICES, INC. reassignment ADVANCED MICRO DEVICES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MCLELLEN, EDWARD, MEDNICK, ELLIOT H.
Publication of US20190129489A1 publication Critical patent/US20190129489A1/en
Priority to US16/912,035 priority patent/US11880260B2/en
Application granted granted Critical
Publication of US10698472B2 publication Critical patent/US10698472B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/26Power supply means, e.g. regulation thereof
    • G06F1/32Means for saving power
    • G06F1/3203Power management, i.e. event-based initiation of a power-saving mode
    • G06F1/3234Power saving characterised by the action undertaken
    • G06F1/3243Power saving in microcontroller unit
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/26Power supply means, e.g. regulation thereof
    • G06F1/32Means for saving power
    • G06F1/3203Power management, i.e. event-based initiation of a power-saving mode
    • G06F1/3234Power saving characterised by the action undertaken
    • G06F1/324Power saving characterised by the action undertaken by lowering clock frequency
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/26Power supply means, e.g. regulation thereof
    • G06F1/32Means for saving power
    • G06F1/3203Power management, i.e. event-based initiation of a power-saving mode
    • G06F1/3234Power saving characterised by the action undertaken
    • G06F1/3287Power saving characterised by the action undertaken by switching off individual functional units in the computer system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/26Power supply means, e.g. regulation thereof
    • G06F1/32Means for saving power
    • G06F1/3203Power management, i.e. event-based initiation of a power-saving mode
    • G06F1/3234Power saving characterised by the action undertaken
    • G06F1/3293Power saving characterised by the action undertaken by switching to a less power-consuming processor, e.g. sub-CPU
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/1666Error detection or correction of the data by redundancy in hardware where the redundant component is memory or memory area
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0806Multiuser, multiprocessor or multiprocessing cache systems
    • G06F12/0811Multiuser, multiprocessor or multiprocessing cache systems with multilevel cache hierarchies
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0875Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches with dedicated cache, e.g. instruction or stack
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • G06F3/0619Improving the reliability of storage systems in relation to data integrity, e.g. data losses, bit errors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0625Power saving in storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0646Horizontal data movement in storage systems, i.e. moving data in between storage devices or systems
    • G06F3/065Replication mechanisms
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device
    • G06F3/068Hybrid storage device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3836Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution
    • G06F9/3851Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution from multiple instruction streams, e.g. multistreaming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0806Multiuser, multiprocessor or multiprocessing cache systems
    • G06F12/084Multiuser, multiprocessor or multiprocessing cache systems with a shared cache
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1016Performance improvement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/28Using a specific disk cache architecture
    • G06F2212/285Redundant cache memory
    • G06F2212/286Mirrored cache memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/45Caching of specific data in cache memory
    • G06F2212/452Instruction code
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/30076Arrangements for executing specific machine instructions to perform miscellaneous control operations, e.g. NOP
    • G06F9/3009Thread control instructions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3877Concurrent instruction execution, e.g. pipeline, look ahead using a slave processor, e.g. coprocessor
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Definitions

  • Battery-powered computing devices such as mobile devices
  • a typical mobile device operates over a wide performance range according to workload requirements. Different performance ranges are conventionally mapped to different operating modes, with power consumption proportionally related to performance. For example, in a low-power mode, the mobile device may provide a small amount of computational capacity, such as to maintain radio contact with a cellular tower and consume a relatively low amount of power. In a high-performance mode, the mobile device may be subject to high processing intensity tasks such as gaming or web browsing, and therefore consume a relatively high amount of power.
  • SoC System on Chip
  • FIG. 1 is a block diagram illustrating a heterogeneous processor system in accordance with at least some embodiments.
  • FIG. 2 is a block diagram illustrating an example operation of a heterogeneous processor system in accordance with at least some embodiments.
  • FIG. 3 is a block diagram illustrating another example operation of a heterogeneous processor system in accordance with at least some embodiments.
  • FIG. 4 is a flow diagram illustrating an operation of a heterogeneous processor system in accordance with at least some embodiments.
  • FIG. 5 is a flow diagram illustrating another operation of a heterogeneous processor system in accordance with at least some embodiments.
  • FIGS. 1-5 describe systems and methods for utilizing heterogeneous processors in which a low-feature processor implements only a subset of a full instruction set. By utilizing a low-feature processor intended for power execution of lower priority tasks, overall power consumption by the processor is decreased.
  • a heterogeneous processor system includes a first processor and a second processor implementing an instruction set architecture (ISA) including a set of ISA features.
  • the first processor is a high-feature processor configured to support a first subset (or entirety) of the set of ISA features.
  • the second processor is a low-feature processor configured to support a subset of the set of ISA features.
  • the heterogeneous processor system may turn off high-feature processors when not in use to conserve energy. Because the low-feature processors use less power, energy efficiency may increase (e.g., such as while operating mobile devices on batteries) the more the low-feature processor is used (instead of using the high-feature processors).
  • the low-feature processors may support instruction execution of low priority processes such as operating system (OS) maintenance, timer support and various monitor functions that are used to allow a device to appear powered on and available through periodic wake ups while most of the time it is in fact powered off. By minimizing the power needed to support these operations, battery life can be greatly extended, thereby improving the efficiency of lower power operations and improving battery life. Accordingly, instructions executed on the low-feature processors offer improvements in power savings relative to other implementations that employ other low power techniques, but continue to execute those instructions on full-feature processors.
  • OS operating system
  • FIG. 1 is a block diagram illustrating a heterogeneous processor system 100 in accordance with at least some embodiments.
  • the heterogeneous processor system 100 includes a plurality of processor clusters, such as processor cluster 102 and processor cluster 104 .
  • Each processor cluster 102 , 104 may include one or more processors or processor cores (e.g., central processing unit (CPU), graphics processing unit (GPU), digital signal processor (DSP), and the like).
  • the processor cluster 102 includes a plurality of central processing unit (CPU) cores 106 and 108 , which are relatively faster and consume more power than the CPU cores 110 and 112 in the processor cluster 104 .
  • CPU central processing unit
  • GPU graphics processing unit
  • DSP digital signal processor
  • the processor clusters 102 and 104 include cache memory used by the corresponding processors or processor cores in the clusters to reduce the average time to access data from a main memory 114 .
  • the cache memory and the main memory 114 are organized as a hierarchy of cache levels (e.g., level one (L1), level two (L2), level three (L3)).
  • L1 level one
  • L2 level two
  • L3 level three
  • the processor cluster 102 includes a cache 116 shared by the CPU cores 106 and 108 .
  • the processor cluster 104 includes a cache 118 shared by the CPU cores 110 and 112 . It should be appreciated that the caches 116 and 118 may be implemented with or without a hierarchy of cache levels.
  • the heterogeneous processor system 100 includes a shared cache 120 shared by the processor clusters 102 and 104 .
  • the heterogeneous processor system 100 also includes a cache controller 122 for monitoring the utilization and performance of the processor clusters 102 and 104 to manage task migration between processor clusters.
  • the cache controller 122 is configured to monitor the workload of the processor(s) in the processor clusters 102 and 104 and the cache demand of each cache. For example, regarding processor workload, the cache controller 122 may determine processor or CPU utilization, loading, processing latency, instruction execution times, and the like. Additionally, regarding cache demand, the cache controller 122 may monitor memory access (e.g., read and/or write access) to caches 116 - 120 and determine delays due to, for example, cache misses that may impact the performance of current tasks running on the processor clusters.
  • memory access e.g., read and/or write access
  • the processor clusters 102 and 104 of heterogeneous processor system 100 are asymmetric relative to each other.
  • the CPU cores 106 and 108 of the processor cluster 102 may be distinguished from the CPU cores 110 and 112 of the processor cluster 104 by, for example, having micro-architectural features that enable relatively more performance but at the cost of additional power when executing a thread.
  • the processor clusters 102 and 104 utilize, relative to each other, different cache memory sizes, different cache hierarchies, in-order vs. out-of-order execution units, speculative vs. non-speculative execution, different branch prediction mechanisms, and/or different compositions of execution units.
  • the heterogeneous processor system 100 includes a feature set of an instruction set architecture (ISA) including: (1) an instruction set, (2) a set of resources (e.g., registers and modes for addressing memory) accessible by the instructions of the instruction set, and (3) a set of operating modes in which processors operate to process the instructions of the instruction set.
  • ISA feature set for a processor is defined as the set of features specified by the ISA that the processor supports.
  • the features may include the instruction set of the ISA supported by the processor, the set of operating modes of the ISA supported by the processor and/or the set of resources of the ISA included by the processor.
  • the processor cluster 102 is asymmetric relative to the processor cluster 104 in that they support different sets or subsets of features of any given ISA feature set.
  • the processor cluster 102 supports a set of features including all of the features of the ISA and the processor cluster 104 supports a subset of features including less than all of the features of the ISA.
  • both the processor cluster 102 and the processor cluster 104 support subsets of features including less than all of the features of the ISA.
  • the processor(s) and core(s) of processor cluster 104 generally supports a subset of features that is smaller than the subset of features supported by the processor(s) and core(s) of processor cluster 102 . Accordingly, the processor clusters 102 and 104 are hereinafter referred to as the “high-feature processor” and “low-feature processor,” respectively, for ease of reference.
  • the high-feature processor 102 supports a set of features including all of the features of the ISA
  • subsets of instructions executable at the low-feature processor 104 are also executable at the high-feature processor 102 .
  • at least a portion of the instructions executable at the high-feature processor 102 may not necessarily be executable at the low-feature processor 104 .
  • the combination of the set of features supported by the high-feature processor 102 and the subset of features supported by the low-feature processor 104 represents all of the features of the ISA supported by the heterogeneous processor system 100 .
  • the high-feature processor 102 supports features of the ISA generally associated with providing high performance and the low-feature processor 104 supports features generally associated with higher power efficiency. Switching between the high-feature processor 102 and the low-feature processor 104 for execution of instruction threads allows for improved power efficiency. Additionally, by supporting less than the entirety of an ISA feature, the low-feature processor 104 enables a lower power implementation than systems utilizing, for example, heterogeneous processors that both implement identical ISAs but operate with different instructions per cycle (IPC) or having different operating frequencies.
  • IPC instructions per cycle
  • FIG. 2 is a diagram illustrating an example operation of a heterogeneous processor system in accordance with at least some embodiments.
  • the heterogeneous processor system 200 of FIG. 2 includes a high-feature processor 202 and a low-feature processor 204 .
  • the heterogeneous processor system 200 includes an L1 cache 206 for the high-feature processor 202 and a L1 cache 208 for the low-feature processor 204 .
  • the heterogeneous processor system 200 further includes a L2 cache 210 shared by the high-feature processor 202 and the low-feature processor 204 .
  • the heterogeneous processor system 200 transitions to a lower-power mode by switching execution of a thread from the high-feature processor 202 to the low-feature processor 204 .
  • This switch of execution includes migrating data, such as thread state, used by the high-feature processor 202 from its L1 cache 206 to the L2 cache 210 shared and accessible by both the high-feature processor 202 to the low-feature processor 204 .
  • the thread state of the high-feature processor 202 may be saved to other shared storage, such as lower-level cache (e.g., L3 cache) or to system memory.
  • the high-feature processor 202 While the state of the high-feature processor 202 is flushed to shared L2 cache 210 , the high-feature processor 202 also saves its registers to a shared register 212 . Subsequently, the low-feature processor 204 to which execution is being switched restores to itself the thread state from the shared L2 cache 210 and also receives the register data from the shared register 212 .
  • the heterogeneous processor system 200 transitions to a higher-power mode by switching execution of a thread from the low-feature processor 204 back to the high-feature processor 202 . Due to the low-feature processor 204 supporting only a subset of the features defined by an ISA, the heterogeneous processor system 200 also transitions execution from the low-feature processor 204 to the high-feature processor 202 when encountering an instruction that the low-feature processor 204 does not support.
  • This switch of execution includes migrating data, such as thread state, used by the low-feature processor 204 from its L1 cache 208 to the L2 cache 210 shared and accessible by both the high-feature processor 202 to the low-feature processor 204 .
  • the thread state of the low-feature processor 204 may be saved to other shared storage, such as lower-level cache (e.g., L3 cache) or to system memory. While the state of the low-feature processor 204 is flushed to shared L2 cache 210 , the low-feature processor 204 also saves its registers to a shared register 212 . Subsequently, the high-feature processor 202 to which execution is being switched restores to itself the thread state from the shared L2 cache 210 and also receives the register data from the shared register 212 .
  • lower-level cache e.g., L3 cache
  • FIG. 3 is a diagram illustrating another example operation of a heterogeneous processor system in accordance with at least some embodiments.
  • the heterogeneous processor system 300 of FIG. 3 includes a high-feature processor 302 and a low-feature processor 304 .
  • the heterogeneous processor system 300 includes a L1 cache 306 for the high-feature processor 302 and a L1 cache 308 for the low-feature processor 304 .
  • the heterogeneous processor system 300 further includes a L2 cache 310 shared by the high-feature processor 302 and the low-feature processor 304 .
  • the heterogeneous processor system 300 transitions to a lower-power mode by switching execution of a thread from the high-feature processor 302 to the low-feature processor 304 .
  • states are maintained in consistency by writing the states of both the high-feature processor 302 and the low-feature processor 304 to the L1 caches 306 , 308 (i.e., shadowing).
  • the heterogeneous processor system 300 transitions to a higher-power mode by switching execution of a thread from the low-feature processor 304 back to the high-feature processor 302 . Due to the low-feature processor 304 supporting only a subset of the features defined by an ISA, the heterogeneous processor system 300 also transitions execution from the low-feature processor 304 to the high-feature processor 302 when encountering an instruction that the low-feature processor 304 does not support.
  • the high-feature processor 302 to which execution is being switched restores to itself the thread state from its L1 cache 306 , which contains a copy of thread state data of the low-feature processor 304 .
  • the low-feature processor 304 being a simpler processor (e.g., due to its micro-architecture only having to support a subset of the full ISA supported by heterogeneous processor system 300 )
  • shadow writes of states from the low-feature processor 304 to the L1 cache 306 are faster than the opposite direction of data migration.
  • a first of the heterogeneous processors (e.g., low-feature processor 304 of FIG. 3 ) is executing an application thread in a low-power mode.
  • low-power modes include: a mode entered by the processor in response to an instruction that instructs the processor to halt execution of a thread, a mode in which an external bus clock to the processor is disabled, a mode in which the processor disables the clock signals to a portion of its circuitry, a mode in which the processor disables power to a portion of its circuitry, and the like.
  • the low power modes may include various Advanced Configuration and Power Interface (ACPI) Processor states, more commonly known as C-states.
  • ACPI Advanced Configuration and Power Interface
  • the first processor which is currently executing the thread detects that the thread is attempting to employ a feature of an ISA feature set that is unsupported by the first processor.
  • the first processor may detect the unsupported feature in different ways. For example, an instruction decoder may decode an instruction that is unsupported by the first processor or an execution unit may detect that an instruction is attempting to access a control register that is unsupported by the first processor and/or access a particular control register bit or field place the processor into a particular operating mode defined by the ISA (but unsupported by the first processor).
  • the first processor stops executing the thread in response to detecting the attempt by the thread to employ the unsupported feature at block 404 . For example, if an instruction decoder decodes an unsupported instruction, it may trap to a microcode routine that handles illegal instruction exceptions, and the microcode routine may stop the execution of subsequent instructions of the thread.
  • the first processor indicates a switch to a second processor (e.g., the high-feature processor 302 of FIG. 3 ) to execute the thread and saves the thread state to a shared memory location. For example, as described relative to FIGS. 2-3 , the thread state may be flushed to a shared L2 cache.
  • the first processor may also transfer other data that is not necessary for the second processor to execute the thread, but which may nevertheless enable the second processor to execute the thread faster, such as some or all of the contents of one or more cache memories of the first processor.
  • the second processor e.g., the high-feature processor 302 of FIG. 3 ) restores the thread state from the shared memory location and begins executing the thread.
  • FIG. 5 is a flow diagram illustrating another example method 500 of operation of a heterogeneous processor system in accordance with at least some embodiments.
  • a first of the heterogeneous processors e.g., low-feature processor 304 of FIG. 3
  • a low-power mode include: a mode entered by the processor in response to an instruction that instructs the processor to halt execution of a thread, a mode in which an external bus clock to the processor is disabled, a mode in which the processor disables the clock signals to a portion of its circuitry, a mode in which the processor disables power to a portion of its circuitry, and the like.
  • the low power modes may include various Advanced Configuration and Power Interface (ACPI) Processor states, more commonly known as C-states.
  • ACPI Advanced Configuration and Power Interface
  • a thread state of the thread executing at the first processor is written to a private cache of the first processor (e.g., L1 cache 308 of FIG. 3 ). Additionally, the thread state of the thread executing at the first processor is shadow written to a private cache of a second processor (e.g., L1 cache 306 of high-feature processor 302 of FIG. 3 ). In addition to the thread state, the first processor may also transfer other data, such as register data.
  • the first processor which is currently executing the thread detects that the thread is attempting to employ a feature of an ISA feature set that is unsupported by the first processor.
  • the first processor may detect the unsupported feature in different ways. For example, an instruction decoder may decode an instruction that is unsupported by the first processor or an execution unit may detect that an instruction is attempting to access a control register that is unsupported by the first processor and/or access a particular control register bit or field place the processor into a particular operating mode defined by the ISA (but unsupported by the first processor).
  • the first processor stops executing the thread in response to detecting the attempt by the thread to employ the unsupported feature at block 506 . For example, if an instruction decoder decodes an unsupported instruction, it may trap to a microcode routine that handles illegal instruction exceptions, and the microcode routine may stop the execution of subsequent instructions of the thread.
  • the first processor also indicates a switch to a second processor (e.g., the high-feature processor 302 of FIG. 3 ) to execute the thread.
  • the second processor e.g., the high-feature processor 302 of FIG. 3
  • restores the thread state from its private cache e.g., the L1 cache 306 of FIG. 3
  • Switching operation from the high-feature processor to the low-feature processor costs performance but saves power.
  • the low-feature processor not only consumes less power than the high-feature processor but is also generally more power efficient as well. It should be noted that the power-to-performance ratio is not linear. Power consumed increases at approximately the square of operating frequency, and also increases geometrically relative to the amount of hardware needed to improve inter-process communications (IPC). Accordingly, for example, low-feature processors are not only lower power but their performance per watt is also better than higher-feature, faster processors.
  • the heterogeneous processor systems described herein further include a hardware abstraction layer implemented at the low-feature processor to reduce incidences where a switch back to the high-feature processor is needed.
  • the hardware abstraction layer emulates opcodes that were not implemented in the micro-architecture hardware of the low-feature processor.
  • a subroutine may be provided in a library that performs divides.
  • the low-feature processor may enter an interrupt routine and execute a sequence of instructions written in the library to perform the divide rather than switching execution back to the high-feature processor. This process is computationally-inefficient, but operates to allow for handling of infrequently encountered instructions. Accordingly, if there is an instruction that the low-feature processor encounters in a lower power mode and high performance is not required, that instruction can be emulated in the hardware abstraction layer instead of switching to a higher power mode.
  • the high-feature processor is instrumented with monitors to assist in the determination of when to switch to the low-feature processor. Accordingly, while operating on the low-feature processor, if an operation that is not implemented is encountered, or if the use of emulation software exceeds a threshold, an exception mechanism forces a switch back to the high-feature processor. In this way, the low-feature processor can be used to balance power and performance, and extend the useful low power range of the combined heterogeneous processor system.
  • a computer readable storage medium may include any storage medium, or combination of storage media, accessible by a computer system during use to provide instructions and/or data to the computer system.
  • Such storage media can include, but is not limited to, optical media (e.g., compact disc (CD), digital versatile disc (DVD), Blu-Ray disc), magnetic media (e.g., floppy disc, magnetic tape, or magnetic hard drive), volatile memory (e.g., random access memory (RAM) or cache), non-volatile memory (e.g., read-only memory (ROM) or Flash memory), or microelectromechanical systems (MEMS)-based storage media.
  • optical media e.g., compact disc (CD), digital versatile disc (DVD), Blu-Ray disc
  • magnetic media e.g., floppy disc, magnetic tape, or magnetic hard drive
  • volatile memory e.g., random access memory (RAM) or cache
  • non-volatile memory e.g., read-only memory (ROM) or Flash memory
  • MEMS microelectro
  • the computer readable storage medium may be embedded in the computing system (e.g., system RAM or ROM), fixedly attached to the computing system (e.g., a magnetic hard drive), removably attached to the computing system (e.g., an optical disc or Universal Serial Bus (USB)-based Flash memory), or coupled to the computer system via a wired or wireless network (e.g., network accessible storage (NAS)).
  • system RAM or ROM system RAM or ROM
  • USB Universal Serial Bus
  • NAS network accessible storage
  • certain aspects of the techniques described above may implemented by one or more processors of a processing system executing software.
  • the software includes one or more sets of executable instructions stored or otherwise tangibly embodied on a non-transitory computer readable storage medium.
  • the software can include the instructions and certain data that, when executed by the one or more processors, manipulate the one or more processors to perform one or more aspects of the techniques described above.
  • the non-transitory computer readable storage medium can include, for example, a magnetic or optical disk storage device, solid state storage devices such as Flash memory, a cache, random access memory (RAM) or other non-volatile memory device or devices, and the like.
  • the executable instructions stored on the non-transitory computer readable storage medium may be in source code, assembly language code, object code, or other instruction format that is interpreted or otherwise executable by one or more processors.

Abstract

A heterogeneous processor system includes a first processor implementing an instruction set architecture (ISA) including a set of ISA features and configured to support a first subset of the set of ISA features. The heterogeneous processor system also includes a second processor implementing the ISA including the set of ISA features and configured to support a second subset of the set of ISA features, wherein the first subset and the second subset of the set of ISA features are different from each other. When the first subset includes an entirety of the set of ISA features, the lower-feature second processor is configured to execute an instruction thread by consuming less power and with lower performance than the first processor.

Description

BACKGROUND
Battery-powered computing devices, such as mobile devices, have become commonplace. A typical mobile device operates over a wide performance range according to workload requirements. Different performance ranges are conventionally mapped to different operating modes, with power consumption proportionally related to performance. For example, in a low-power mode, the mobile device may provide a small amount of computational capacity, such as to maintain radio contact with a cellular tower and consume a relatively low amount of power. In a high-performance mode, the mobile device may be subject to high processing intensity tasks such as gaming or web browsing, and therefore consume a relatively high amount of power.
Processor cores within System on Chip (SoC) designs generally utilize more power as performance increases due to an increase in clock frequency and/or usage of additional logic elements. Increasing battery life is important to users of mobile devices that operate on battery power, sometimes for extended periods of time. Accordingly, there is a need for efficient, low power systems that deliver scalable computational capacity with diminishing energy cost.
BRIEF DESCRIPTION OF THE DRAWINGS
The present disclosure may be better understood, and its numerous features and advantages made apparent to those skilled in the art by referencing the accompanying drawings. The use of the same reference symbols in different drawings indicates similar or identical items.
FIG. 1 is a block diagram illustrating a heterogeneous processor system in accordance with at least some embodiments.
FIG. 2 is a block diagram illustrating an example operation of a heterogeneous processor system in accordance with at least some embodiments.
FIG. 3 is a block diagram illustrating another example operation of a heterogeneous processor system in accordance with at least some embodiments.
FIG. 4 is a flow diagram illustrating an operation of a heterogeneous processor system in accordance with at least some embodiments.
FIG. 5 is a flow diagram illustrating another operation of a heterogeneous processor system in accordance with at least some embodiments.
DETAILED DESCRIPTION
Reducing power in high-performance processors can be difficult due to the techniques required to achieve high performance, such as implementation of complex instructions, deep pipelining, superscalar capabilities and aggressive speculative operations. Conventional processors generally employ multiple processor cores, wherein each of the processor cores has the same performance, frequency, and power relative to each other. An operating system schedules tasks to each core equally and performs similar power management on each core. However, by utilizing processors having different levels of performance on a common die, the processors may be employed differently. As described herein, rather than lowering power required for operation by scaling the frequency and/or voltage of processors, FIGS. 1-5 describe systems and methods for utilizing heterogeneous processors in which a low-feature processor implements only a subset of a full instruction set. By utilizing a low-feature processor intended for power execution of lower priority tasks, overall power consumption by the processor is decreased.
In some embodiments, a heterogeneous processor system includes a first processor and a second processor implementing an instruction set architecture (ISA) including a set of ISA features. The first processor is a high-feature processor configured to support a first subset (or entirety) of the set of ISA features. The second processor is a low-feature processor configured to support a subset of the set of ISA features. By implementing only a subset of the complete set of ISA features at the low-feature processor, greater power reduction is achieved when switching tasks between the high-feature processor and the low-feature processor.
The heterogeneous processor system may turn off high-feature processors when not in use to conserve energy. Because the low-feature processors use less power, energy efficiency may increase (e.g., such as while operating mobile devices on batteries) the more the low-feature processor is used (instead of using the high-feature processors). For example, the low-feature processors may support instruction execution of low priority processes such as operating system (OS) maintenance, timer support and various monitor functions that are used to allow a device to appear powered on and available through periodic wake ups while most of the time it is in fact powered off. By minimizing the power needed to support these operations, battery life can be greatly extended, thereby improving the efficiency of lower power operations and improving battery life. Accordingly, instructions executed on the low-feature processors offer improvements in power savings relative to other implementations that employ other low power techniques, but continue to execute those instructions on full-feature processors.
FIG. 1 is a block diagram illustrating a heterogeneous processor system 100 in accordance with at least some embodiments. In the depicted example, the heterogeneous processor system 100 includes a plurality of processor clusters, such as processor cluster 102 and processor cluster 104. Each processor cluster 102, 104 may include one or more processors or processor cores (e.g., central processing unit (CPU), graphics processing unit (GPU), digital signal processor (DSP), and the like). In the embodiment of FIG. 1, the processor cluster 102 includes a plurality of central processing unit (CPU) cores 106 and 108, which are relatively faster and consume more power than the CPU cores 110 and 112 in the processor cluster 104.
The processor clusters 102 and 104 include cache memory used by the corresponding processors or processor cores in the clusters to reduce the average time to access data from a main memory 114. In some embodiments, the cache memory and the main memory 114 are organized as a hierarchy of cache levels (e.g., level one (L1), level two (L2), level three (L3)). For example, in the embodiment illustrated in FIG. 1, the processor cluster 102 includes a cache 116 shared by the CPU cores 106 and 108. The processor cluster 104 includes a cache 118 shared by the CPU cores 110 and 112. It should be appreciated that the caches 116 and 118 may be implemented with or without a hierarchy of cache levels. Further, the heterogeneous processor system 100 includes a shared cache 120 shared by the processor clusters 102 and 104.
The heterogeneous processor system 100 also includes a cache controller 122 for monitoring the utilization and performance of the processor clusters 102 and 104 to manage task migration between processor clusters. In some embodiments, the cache controller 122 is configured to monitor the workload of the processor(s) in the processor clusters 102 and 104 and the cache demand of each cache. For example, regarding processor workload, the cache controller 122 may determine processor or CPU utilization, loading, processing latency, instruction execution times, and the like. Additionally, regarding cache demand, the cache controller 122 may monitor memory access (e.g., read and/or write access) to caches 116-120 and determine delays due to, for example, cache misses that may impact the performance of current tasks running on the processor clusters.
As illustrated in FIG. 1, the processor clusters 102 and 104 of heterogeneous processor system 100 are asymmetric relative to each other. The CPU cores 106 and 108 of the processor cluster 102 may be distinguished from the CPU cores 110 and 112 of the processor cluster 104 by, for example, having micro-architectural features that enable relatively more performance but at the cost of additional power when executing a thread. For example, in some embodiments, the processor clusters 102 and 104 utilize, relative to each other, different cache memory sizes, different cache hierarchies, in-order vs. out-of-order execution units, speculative vs. non-speculative execution, different branch prediction mechanisms, and/or different compositions of execution units.
The heterogeneous processor system 100 includes a feature set of an instruction set architecture (ISA) including: (1) an instruction set, (2) a set of resources (e.g., registers and modes for addressing memory) accessible by the instructions of the instruction set, and (3) a set of operating modes in which processors operate to process the instructions of the instruction set. Accordingly, an ISA feature set for a processor is defined as the set of features specified by the ISA that the processor supports. The features may include the instruction set of the ISA supported by the processor, the set of operating modes of the ISA supported by the processor and/or the set of resources of the ISA included by the processor.
In the context of the heterogeneous processors described herein, the processor cluster 102 is asymmetric relative to the processor cluster 104 in that they support different sets or subsets of features of any given ISA feature set. In some embodiments, the processor cluster 102 supports a set of features including all of the features of the ISA and the processor cluster 104 supports a subset of features including less than all of the features of the ISA. In other embodiments, both the processor cluster 102 and the processor cluster 104 support subsets of features including less than all of the features of the ISA. However, the processor(s) and core(s) of processor cluster 104 generally supports a subset of features that is smaller than the subset of features supported by the processor(s) and core(s) of processor cluster 102. Accordingly, the processor clusters 102 and 104 are hereinafter referred to as the “high-feature processor” and “low-feature processor,” respectively, for ease of reference.
For embodiments in which the high-feature processor 102 supports a set of features including all of the features of the ISA, subsets of instructions executable at the low-feature processor 104 are also executable at the high-feature processor 102. However, at least a portion of the instructions executable at the high-feature processor 102 may not necessarily be executable at the low-feature processor 104. The combination of the set of features supported by the high-feature processor 102 and the subset of features supported by the low-feature processor 104 represents all of the features of the ISA supported by the heterogeneous processor system 100. Although the example of FIG. 1 illustrates two different processors, those skilled in the art will recognize that other embodiments may utilize three or more heterogeneous processors without departing from the scope of the disclosure described herein. The high-feature processor 102 supports features of the ISA generally associated with providing high performance and the low-feature processor 104 supports features generally associated with higher power efficiency. Switching between the high-feature processor 102 and the low-feature processor 104 for execution of instruction threads allows for improved power efficiency. Additionally, by supporting less than the entirety of an ISA feature, the low-feature processor 104 enables a lower power implementation than systems utilizing, for example, heterogeneous processors that both implement identical ISAs but operate with different instructions per cycle (IPC) or having different operating frequencies.
FIG. 2 is a diagram illustrating an example operation of a heterogeneous processor system in accordance with at least some embodiments. The heterogeneous processor system 200 of FIG. 2 includes a high-feature processor 202 and a low-feature processor 204. The heterogeneous processor system 200 includes an L1 cache 206 for the high-feature processor 202 and a L1 cache 208 for the low-feature processor 204. The heterogeneous processor system 200 further includes a L2 cache 210 shared by the high-feature processor 202 and the low-feature processor 204.
When the high-feature processor 202 is being underutilized (e.g., such as determined by cache controller 122 of FIG. 1 or another system performance monitor), the heterogeneous processor system 200 transitions to a lower-power mode by switching execution of a thread from the high-feature processor 202 to the low-feature processor 204. This switch of execution includes migrating data, such as thread state, used by the high-feature processor 202 from its L1 cache 206 to the L2 cache 210 shared and accessible by both the high-feature processor 202 to the low-feature processor 204. In other embodiments, the thread state of the high-feature processor 202 may be saved to other shared storage, such as lower-level cache (e.g., L3 cache) or to system memory. While the state of the high-feature processor 202 is flushed to shared L2 cache 210, the high-feature processor 202 also saves its registers to a shared register 212. Subsequently, the low-feature processor 204 to which execution is being switched restores to itself the thread state from the shared L2 cache 210 and also receives the register data from the shared register 212.
When the low-feature processor 204 is being overutilized (e.g., such as determined by cache controller 122 of FIG. 1 or another system performance monitor), the heterogeneous processor system 200 transitions to a higher-power mode by switching execution of a thread from the low-feature processor 204 back to the high-feature processor 202. Due to the low-feature processor 204 supporting only a subset of the features defined by an ISA, the heterogeneous processor system 200 also transitions execution from the low-feature processor 204 to the high-feature processor 202 when encountering an instruction that the low-feature processor 204 does not support.
This switch of execution includes migrating data, such as thread state, used by the low-feature processor 204 from its L1 cache 208 to the L2 cache 210 shared and accessible by both the high-feature processor 202 to the low-feature processor 204. In other embodiments, the thread state of the low-feature processor 204 may be saved to other shared storage, such as lower-level cache (e.g., L3 cache) or to system memory. While the state of the low-feature processor 204 is flushed to shared L2 cache 210, the low-feature processor 204 also saves its registers to a shared register 212. Subsequently, the high-feature processor 202 to which execution is being switched restores to itself the thread state from the shared L2 cache 210 and also receives the register data from the shared register 212.
FIG. 3 is a diagram illustrating another example operation of a heterogeneous processor system in accordance with at least some embodiments. The heterogeneous processor system 300 of FIG. 3 includes a high-feature processor 302 and a low-feature processor 304. The heterogeneous processor system 300 includes a L1 cache 306 for the high-feature processor 302 and a L1 cache 308 for the low-feature processor 304. The heterogeneous processor system 300 further includes a L2 cache 310 shared by the high-feature processor 302 and the low-feature processor 304.
When the high-feature processor 302 is being underutilized (e.g., such as determined by cache controller 122 of FIG. 1 or another system performance monitor), the heterogeneous processor system 300 transitions to a lower-power mode by switching execution of a thread from the high-feature processor 302 to the low-feature processor 304. However, rather than flushing thread states to the L2 cache 310 (such as previously described relative to FIG. 2), states are maintained in consistency by writing the states of both the high-feature processor 302 and the low-feature processor 304 to the L1 caches 306, 308 (i.e., shadowing). For example, when a thread state of a thread executing at the high-feature processor 302 is written to L1 cache 306, that same thread state is also written to L1 cache 308 of the low feature processor 304. Similarly, register writes (not shown) are also shadowed. Subsequently, the low-feature processor 304 to which execution is being switched restores to itself the thread state from its L1 cache 308, which contains a copy of thread state data of the high-feature processor 302.
When the low-feature processor 304 is being overutilized (e.g., such as determined by cache controller 122 of FIG. 1 or another system performance monitor), the heterogeneous processor system 300 transitions to a higher-power mode by switching execution of a thread from the low-feature processor 304 back to the high-feature processor 302. Due to the low-feature processor 304 supporting only a subset of the features defined by an ISA, the heterogeneous processor system 300 also transitions execution from the low-feature processor 304 to the high-feature processor 302 when encountering an instruction that the low-feature processor 304 does not support. Similar to the operations discussed above, the high-feature processor 302 to which execution is being switched restores to itself the thread state from its L1 cache 306, which contains a copy of thread state data of the low-feature processor 304. Additionally, with the low-feature processor 304 being a simpler processor (e.g., due to its micro-architecture only having to support a subset of the full ISA supported by heterogeneous processor system 300), shadow writes of states from the low-feature processor 304 to the L1 cache 306 are faster than the opposite direction of data migration.
Referring now to FIG. 4, illustrated is a method 400 of operation of a heterogeneous processor system in accordance with at least some embodiments. At block 402, a first of the heterogeneous processors (e.g., low-feature processor 304 of FIG. 3) is executing an application thread in a low-power mode. Examples of low-power modes include: a mode entered by the processor in response to an instruction that instructs the processor to halt execution of a thread, a mode in which an external bus clock to the processor is disabled, a mode in which the processor disables the clock signals to a portion of its circuitry, a mode in which the processor disables power to a portion of its circuitry, and the like. The low power modes may include various Advanced Configuration and Power Interface (ACPI) Processor states, more commonly known as C-states.
At block 404, the first processor which is currently executing the thread (e.g., low-feature processor 304 of FIG. 3), detects that the thread is attempting to employ a feature of an ISA feature set that is unsupported by the first processor. Depending upon the nature of the particular feature, the first processor may detect the unsupported feature in different ways. For example, an instruction decoder may decode an instruction that is unsupported by the first processor or an execution unit may detect that an instruction is attempting to access a control register that is unsupported by the first processor and/or access a particular control register bit or field place the processor into a particular operating mode defined by the ISA (but unsupported by the first processor).
At block 406, the first processor stops executing the thread in response to detecting the attempt by the thread to employ the unsupported feature at block 404. For example, if an instruction decoder decodes an unsupported instruction, it may trap to a microcode routine that handles illegal instruction exceptions, and the microcode routine may stop the execution of subsequent instructions of the thread. At block 408, the first processor indicates a switch to a second processor (e.g., the high-feature processor 302 of FIG. 3) to execute the thread and saves the thread state to a shared memory location. For example, as described relative to FIGS. 2-3, the thread state may be flushed to a shared L2 cache. In addition to the thread state, the first processor may also transfer other data that is not necessary for the second processor to execute the thread, but which may nevertheless enable the second processor to execute the thread faster, such as some or all of the contents of one or more cache memories of the first processor. At block 410, the second processor (e.g., the high-feature processor 302 of FIG. 3) restores the thread state from the shared memory location and begins executing the thread.
FIG. 5 is a flow diagram illustrating another example method 500 of operation of a heterogeneous processor system in accordance with at least some embodiments. At block 502, a first of the heterogeneous processors (e.g., low-feature processor 304 of FIG. 3) is executing an application thread in a low-power mode. Examples of low-power modes include: a mode entered by the processor in response to an instruction that instructs the processor to halt execution of a thread, a mode in which an external bus clock to the processor is disabled, a mode in which the processor disables the clock signals to a portion of its circuitry, a mode in which the processor disables power to a portion of its circuitry, and the like. The low power modes may include various Advanced Configuration and Power Interface (ACPI) Processor states, more commonly known as C-states.
At block 504, a thread state of the thread executing at the first processor is written to a private cache of the first processor (e.g., L1 cache 308 of FIG. 3). Additionally, the thread state of the thread executing at the first processor is shadow written to a private cache of a second processor (e.g., L1 cache 306 of high-feature processor 302 of FIG. 3). In addition to the thread state, the first processor may also transfer other data, such as register data.
At block 506, the first processor which is currently executing the thread (e.g., low-feature processor 304 of FIG. 3), detects that the thread is attempting to employ a feature of an ISA feature set that is unsupported by the first processor. Depending upon the nature of the particular feature, the first processor may detect the unsupported feature in different ways. For example, an instruction decoder may decode an instruction that is unsupported by the first processor or an execution unit may detect that an instruction is attempting to access a control register that is unsupported by the first processor and/or access a particular control register bit or field place the processor into a particular operating mode defined by the ISA (but unsupported by the first processor).
At block 508, the first processor stops executing the thread in response to detecting the attempt by the thread to employ the unsupported feature at block 506. For example, if an instruction decoder decodes an unsupported instruction, it may trap to a microcode routine that handles illegal instruction exceptions, and the microcode routine may stop the execution of subsequent instructions of the thread. The first processor also indicates a switch to a second processor (e.g., the high-feature processor 302 of FIG. 3) to execute the thread. At block 510, the second processor (e.g., the high-feature processor 302 of FIG. 3) restores the thread state from its private cache (e.g., the L1 cache 306 of FIG. 3) and begins executing the thread.
It should be understood that although the blocks of the embodiments of FIGS. 4 and 5 are described in a particular order, some blocks may be performed in a different order and/or in parallel with one another. For example, the operations may occur in the opposite order to migrate data from the high-feature processor to the low-feature processor for reduction in power consumption, as previously discussed in more detail with respect to FIGS. 2-3.
Switching operation from the high-feature processor to the low-feature processor costs performance but saves power. Further, the low-feature processor not only consumes less power than the high-feature processor but is also generally more power efficient as well. It should be noted that the power-to-performance ratio is not linear. Power consumed increases at approximately the square of operating frequency, and also increases geometrically relative to the amount of hardware needed to improve inter-process communications (IPC). Accordingly, for example, low-feature processors are not only lower power but their performance per watt is also better than higher-feature, faster processors.
When operating in low power modes, switching to the low-feature processor allows for a disproportionate savings of power relative to the amount of performance lost. By implementing only a portion of an ISA instruction set at the low-feature processor, the amount of hardware necessary for the low-feature processor may be decreased and the power differential of operations between the high-feature processor and the low-feature processor increased. In one example embodiment, computationally-intensive instructions such as floating point and operations such as translation lookaside buffer (TLB) table walks are not implemented in the low-feature processor, which decreases the power required for operation of the low-feature processor and also reduces the overhead of having the low-feature processor in addition to the operations of the high-feature processor.
In some embodiments, the heterogeneous processor systems described herein further include a hardware abstraction layer implemented at the low-feature processor to reduce incidences where a switch back to the high-feature processor is needed. In particular, the hardware abstraction layer emulates opcodes that were not implemented in the micro-architecture hardware of the low-feature processor. For example, in a low-feature processor that does not include any divide units, a subroutine may be provided in a library that performs divides. When a divide instruction is encountered at the low-feature processor, the low-feature processor may enter an interrupt routine and execute a sequence of instructions written in the library to perform the divide rather than switching execution back to the high-feature processor. This process is computationally-inefficient, but operates to allow for handling of infrequently encountered instructions. Accordingly, if there is an instruction that the low-feature processor encounters in a lower power mode and high performance is not required, that instruction can be emulated in the hardware abstraction layer instead of switching to a higher power mode.
Extensive use of this emulation support would lead to poor power efficiency, so in addition, the high-feature processor is instrumented with monitors to assist in the determination of when to switch to the low-feature processor. Accordingly, while operating on the low-feature processor, if an operation that is not implemented is encountered, or if the use of emulation software exceeds a threshold, an exception mechanism forces a switch back to the high-feature processor. In this way, the low-feature processor can be used to balance power and performance, and extend the useful low power range of the combined heterogeneous processor system.
A computer readable storage medium may include any storage medium, or combination of storage media, accessible by a computer system during use to provide instructions and/or data to the computer system. Such storage media can include, but is not limited to, optical media (e.g., compact disc (CD), digital versatile disc (DVD), Blu-Ray disc), magnetic media (e.g., floppy disc, magnetic tape, or magnetic hard drive), volatile memory (e.g., random access memory (RAM) or cache), non-volatile memory (e.g., read-only memory (ROM) or Flash memory), or microelectromechanical systems (MEMS)-based storage media. The computer readable storage medium may be embedded in the computing system (e.g., system RAM or ROM), fixedly attached to the computing system (e.g., a magnetic hard drive), removably attached to the computing system (e.g., an optical disc or Universal Serial Bus (USB)-based Flash memory), or coupled to the computer system via a wired or wireless network (e.g., network accessible storage (NAS)).
In some embodiments, certain aspects of the techniques described above may implemented by one or more processors of a processing system executing software. The software includes one or more sets of executable instructions stored or otherwise tangibly embodied on a non-transitory computer readable storage medium. The software can include the instructions and certain data that, when executed by the one or more processors, manipulate the one or more processors to perform one or more aspects of the techniques described above. The non-transitory computer readable storage medium can include, for example, a magnetic or optical disk storage device, solid state storage devices such as Flash memory, a cache, random access memory (RAM) or other non-volatile memory device or devices, and the like. The executable instructions stored on the non-transitory computer readable storage medium may be in source code, assembly language code, object code, or other instruction format that is interpreted or otherwise executable by one or more processors.
Note that not all of the activities or elements described above in the general description are required, that a portion of a specific activity or device may not be required, and that one or more further activities may be performed, or elements included, in addition to those described. Still further, the order in which activities are listed are not necessarily the order in which they are performed. Also, the concepts have been described with reference to specific embodiments. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the present disclosure as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of the present disclosure.
Benefits, other advantages, and solutions to problems have been described above with regard to specific embodiments. However, the benefits, advantages, solutions to problems, and any feature(s) that may cause any benefit, advantage, or solution to occur or become more pronounced are not to be construed as a critical, required, or essential feature of any or all the claims. Moreover, the particular embodiments disclosed above are illustrative only, as the disclosed subject matter may be modified and practiced in different but equivalent manners apparent to those skilled in the art having the benefit of the teachings herein. No limitations are intended to the details of construction or design herein shown, other than as described in the claims below. It is therefore evident that the particular embodiments disclosed above may be altered or modified and all such variations are considered within the scope of the disclosed subject matter. Accordingly, the protection sought herein is as set forth in the claims below.

Claims (17)

What is claimed is:
1. A heterogeneous processor system, comprising:
a first processor implementing an instruction set architecture (ISA) including a set of ISA features and configured to support a first subset of the set of ISA features;
a second processor implementing the ISA including the set of ISA features and configured to support a second subset of the set of ISA features, wherein the first subset and the second subset of the set of ISA features are different from each other; and
wherein the second processor is further configured to write a state of an instruction thread being executed at the second processor to a first private cache memory associated with the second processor and shadow write a copy of the state of the instruction thread to a second private cache memory associated with the first processor.
2. The heterogeneous processor system of claim 1, wherein the first subset comprises an entirety of the set of ISA features.
3. The heterogeneous processor system of claim 1, wherein the second processor is further configured to:
execute an instruction thread by consuming less power and with lower performance than the first processor.
4. The heterogeneous processor system of claim 1, wherein the second processor is further configured to:
detect that an instruction thread, while being executed at the second processor, encounters an instruction not supported by the second subset of the set of ISA features; and
save, in response to encountering the instruction not supported by the second subset, a copy of a state of the instruction thread to a shared memory.
5. The heterogeneous processor system of claim 4, wherein the first processor is further configured to:
restore the state of the instruction thread from the shared memory and to continue execution of the instruction thread.
6. The heterogeneous processor system of claim 4, wherein the first processor is configured to:
restore, in response to encountering an instruction not supported by the second subset of the set of ISA features, the state of the instruction thread from the second private cache memory of the first processor and to continue execution of the instruction thread.
7. The heterogeneous processor system of claim 1, wherein the first processor is further configured to:
detect, while executing an instruction thread at the first processor, an inefficient utilization of the first processor; and
switch execution of the instruction thread from the first processor to the second processor in response to the inefficient utilization.
8. A method, comprising: detecting that an instruction thread, while being executed by a low-feature processor that supports a subset of a set of instruction set architecture (ISA) features, encounters an instruction not supported by the subset of the set of ISA features or by a hardware abstraction layer emulating at least one opcode not supported by a micro-architecture of the low-feature processor; switching execution of the instruction thread to a high-feature processor that supports an entirety of the set of ISA features; and wherein the low-feature processor is further configured to write a state of an instruction thread being executed at the low-feature processor to a first private cache memory associated with the low-feature processor and shadow write a copy of the state of the instruction thread to a second private cache memory associated with the high-feature processor.
9. The method of claim 8, further comprising:
saving, in response to encountering the instruction not supported by the subset of the set of ISA features, a copy of a state of the instruction thread to a shared memory.
10. The method of claim 9, further comprising:
stopping execution of the instruction thread at the low-feature processor;
restoring the state of the instruction thread from the shared memory to the high-feature processor; and
continuing execution of the instruction thread at the high-feature processor.
11. The method of claim 10, further comprising:
saving, in response to encountering the instruction not supported by the subset of the set of ISA features, a copy of register data to a shared register; and
migrating the register data from the shared register to the high-feature processor.
12. The method of claim 8, further comprising:
stopping execution of the instruction thread at the low-feature processor;
restoring, in response to encountering the instruction not supported by the subset of the set of ISA features, the state of the instruction thread from the cache memory of the high-feature processor; and
continuing execution of the instruction thread at the high-feature processor.
13. A method, comprising: detecting, while executing an instruction thread, an inefficient utilization of a high- feature processor that supports a set of instruction set architecture (ISA) features; switching execution of the instruction thread from the high-feature processor to a low- feature processor that supports a subset of the set of ISA features and is configured to emulate, at a hardware abstraction layer, at least one opcode not supported by a micro-architecture of the low-feature processor; and wherein the low-feature processor is further configured to write a state of an instruction thread being executed at the low-feature processor to a first private cache memory associated with the low-feature processor and shadow write a copy of the state of the instruction thread to a second private cache memory associated with the high-feature processor.
14. The method of claim 13, further comprising:
saving, in response to detecting the inefficient utilization of the high-feature processor, a copy of a state of the instruction thread to a shared memory.
15. The method of claim 14, further comprising:
stopping execution of the instruction thread at the high-feature processor;
restoring the state of the instruction thread from the shared memory to the low-feature processor; and
continuing execution of the instruction thread at the low-feature processor.
16. The method of claim 14, further comprising:
saving, in response detecting the inefficient utilization of the high-feature processor, a copy of register data to a shared register; and
migrating the register data from the shared register to the low-feature processor.
17. The method of claim 13, further comprising:
stopping, in response to detecting the inefficient utilization of the high-feature processor, execution of the instruction thread at the high-feature processor;
restoring the state of the instruction thread from the cache memory of the low-feature processor; and
continuing execution of the instruction thread at the low-feature processor.
US15/795,719 2017-10-27 2017-10-27 Instruction subset implementation for low power operation Active 2038-01-27 US10698472B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US15/795,719 US10698472B2 (en) 2017-10-27 2017-10-27 Instruction subset implementation for low power operation
US16/912,035 US11880260B2 (en) 2017-10-27 2020-06-25 Instruction subset implementation for low power operation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/795,719 US10698472B2 (en) 2017-10-27 2017-10-27 Instruction subset implementation for low power operation

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/912,035 Continuation US11880260B2 (en) 2017-10-27 2020-06-25 Instruction subset implementation for low power operation

Publications (2)

Publication Number Publication Date
US20190129489A1 US20190129489A1 (en) 2019-05-02
US10698472B2 true US10698472B2 (en) 2020-06-30

Family

ID=66244814

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/795,719 Active 2038-01-27 US10698472B2 (en) 2017-10-27 2017-10-27 Instruction subset implementation for low power operation
US16/912,035 Active 2038-03-08 US11880260B2 (en) 2017-10-27 2020-06-25 Instruction subset implementation for low power operation

Family Applications After (1)

Application Number Title Priority Date Filing Date
US16/912,035 Active 2038-03-08 US11880260B2 (en) 2017-10-27 2020-06-25 Instruction subset implementation for low power operation

Country Status (1)

Country Link
US (2) US10698472B2 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11900004B2 (en) * 2018-12-07 2024-02-13 Lg Electronics Inc. Vehicle apparatus and control method
KR20210017249A (en) * 2019-08-07 2021-02-17 삼성전자주식회사 An electronic device for executing instructions using processor cores and various versions of ISAs(instruction set architectures)
US11586472B2 (en) * 2019-12-10 2023-02-21 Advanced Micro Devices, Inc. Method of task transition between heterogenous processors

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5481693A (en) * 1994-07-20 1996-01-02 Exponential Technology, Inc. Shared register architecture for a dual-instruction-set CPU
US5685009A (en) * 1994-07-20 1997-11-04 Exponential Technology, Inc. Shared floating-point registers and register port-pairing in a dual-architecture CPU
US5781750A (en) * 1994-01-11 1998-07-14 Exponential Technology, Inc. Dual-instruction-set architecture CPU with hidden software emulation mode
US6076155A (en) * 1995-10-24 2000-06-13 S3 Incorporated Shared register architecture for a dual-instruction-set CPU to facilitate data exchange between the instruction sets
US20040123087A1 (en) * 2002-12-19 2004-06-24 Dale Morris Instruction set reconciliation for heterogeneous symmetric-multiprocessor systems
US20040199727A1 (en) * 2003-04-02 2004-10-07 Narad Charles E. Cache allocation
US20080263324A1 (en) * 2006-08-10 2008-10-23 Sehat Sutardja Dynamic core switching
US20090216950A1 (en) * 2008-02-26 2009-08-27 Mccalpin John D Push for Sharing Instruction
US20140298060A1 (en) * 2013-03-26 2014-10-02 Via Technologies, Inc. Asymmetric multi-core processor with native switching mechanism
US20170220474A1 (en) * 2016-02-01 2017-08-03 Oracle International Corporation System and Method for Promoting Reader Groups for Lock Cohorting
US20180239702A1 (en) * 2017-02-23 2018-08-23 Advanced Micro Devices, Inc. Locality-aware and sharing-aware cache coherence for collections of processors

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6349361B1 (en) * 2000-03-31 2002-02-19 International Business Machines Corporation Methods and apparatus for reordering and renaming memory references in a multiprocessor computer system
US20060112226A1 (en) * 2004-11-19 2006-05-25 Hady Frank T Heterogeneous processors sharing a common cache
WO2013101139A1 (en) * 2011-12-30 2013-07-04 Intel Corporation Providing an asymmetric multicore processor system transparently to an operating system
US20160162293A1 (en) * 2013-03-26 2016-06-09 Via Technologies, Inc. Asymmetric processor with cores that support different isa instruction subsets
WO2014196976A1 (en) * 2013-06-06 2014-12-11 Empire Technology Development, Llc Shared cache data movement in thread migration
US9298623B2 (en) * 2013-09-26 2016-03-29 Globalfoundries Inc. Identifying high-conflict cache lines in transactional memory computing environments
US9298626B2 (en) * 2013-09-26 2016-03-29 Globalfoundries Inc. Managing high-conflict cache lines in transactional memory computing environments
US10339023B2 (en) * 2014-09-25 2019-07-02 Intel Corporation Cache-aware adaptive thread scheduling and migration
US11513805B2 (en) * 2016-08-19 2022-11-29 Wisconsin Alumni Research Foundation Computer architecture with synergistic heterogeneous processors
KR102520017B1 (en) * 2016-12-31 2023-04-11 인텔 코포레이션 Systems, methods, and apparatuses for heterogeneous computing

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5781750A (en) * 1994-01-11 1998-07-14 Exponential Technology, Inc. Dual-instruction-set architecture CPU with hidden software emulation mode
US5481693A (en) * 1994-07-20 1996-01-02 Exponential Technology, Inc. Shared register architecture for a dual-instruction-set CPU
US5685009A (en) * 1994-07-20 1997-11-04 Exponential Technology, Inc. Shared floating-point registers and register port-pairing in a dual-architecture CPU
US6076155A (en) * 1995-10-24 2000-06-13 S3 Incorporated Shared register architecture for a dual-instruction-set CPU to facilitate data exchange between the instruction sets
US20040123087A1 (en) * 2002-12-19 2004-06-24 Dale Morris Instruction set reconciliation for heterogeneous symmetric-multiprocessor systems
US20040199727A1 (en) * 2003-04-02 2004-10-07 Narad Charles E. Cache allocation
US20080263324A1 (en) * 2006-08-10 2008-10-23 Sehat Sutardja Dynamic core switching
US20090216950A1 (en) * 2008-02-26 2009-08-27 Mccalpin John D Push for Sharing Instruction
US20140298060A1 (en) * 2013-03-26 2014-10-02 Via Technologies, Inc. Asymmetric multi-core processor with native switching mechanism
US20170220474A1 (en) * 2016-02-01 2017-08-03 Oracle International Corporation System and Method for Promoting Reader Groups for Lock Cohorting
US20180239702A1 (en) * 2017-02-23 2018-08-23 Advanced Micro Devices, Inc. Locality-aware and sharing-aware cache coherence for collections of processors

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Carvalho, Adriano, et al. "Full virtualization on low-end hardware: A case study." IECON 2016-42nd Annual Conference of the IEEE Industrial Electronics Society. IEEE, 2016. (Year: 2016). *
Carvalho, Adriano, et al. "Full virtualization on low-end hardware: A case study." IECON 2016—42nd Annual Conference of the IEEE Industrial Electronics Society. IEEE, 2016. (Year: 2016). *
Dinaburg, Artem, et al. "Ether: malware analysis via hardware virtualization extensions." Proceedings of the 15th ACM conference on Computer and communications security. ACM, 2008. (Year: 2008). *

Also Published As

Publication number Publication date
US11880260B2 (en) 2024-01-23
US20200393887A1 (en) 2020-12-17
US20190129489A1 (en) 2019-05-02

Similar Documents

Publication Publication Date Title
US11755099B2 (en) Dynamic core selection for heterogeneous multi-core systems
US11880260B2 (en) Instruction subset implementation for low power operation
US10664039B2 (en) Power efficient processor architecture
US9600283B2 (en) Single instruction for specifying a subset of registers to save prior to entering low-power mode, and for specifying a pointer to a function executed after exiting low-power mode
US9720730B2 (en) Providing an asymmetric multicore processor system transparently to an operating system
US9600058B2 (en) Enhancing power-performance efficiency in a computer system when bursts of activity occurs when operating in low power
WO2013095869A1 (en) A method, apparatus, and system for energy efficiency and energy conservation including autonomous hardware-based deep power down in devices
Mittal et al. EnCache: A dynamic profiling-based reconfiguration technique for improving cache energy efficiency
JP6409218B2 (en) Power efficient processor architecture
JP2016212907A (en) Excellent power efficient processor architecture
JP2017021811A (en) Power efficient processor architecture

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: ADVANCED MICRO DEVICES, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MEDNICK, ELLIOT H.;MCLELLEN, EDWARD;SIGNING DATES FROM 20171024 TO 20171103;REEL/FRAME:044149/0274

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

CC Certificate of correction
MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4