US9870228B2 - Prioritising of instruction fetching in microprocessor systems - Google Patents
Prioritising of instruction fetching in microprocessor systems Download PDFInfo
- Publication number
- US9870228B2 US9870228B2 US15/134,510 US201615134510A US9870228B2 US 9870228 B2 US9870228 B2 US 9870228B2 US 201615134510 A US201615134510 A US 201615134510A US 9870228 B2 US9870228 B2 US 9870228B2
- Authority
- US
- United States
- Prior art keywords
- instructions
- thread
- instruction
- threads
- execution
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 239000000872 buffer Substances 0.000 claims abstract description 34
- 238000000034 method Methods 0.000 claims abstract description 15
- 230000015654 memory Effects 0.000 description 29
- 241000953555 Theama Species 0.000 description 5
- 238000007726 management method Methods 0.000 description 3
- 238000012913 prioritisation Methods 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 241001522296 Erithacus rubecula Species 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3802—Instruction prefetching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/04—Generating or distributing clock signals or signals derived directly therefrom
- G06F1/10—Distribution of clock signals, e.g. skew
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/30145—Instruction analysis, e.g. decoding, instruction word fields
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/30—Arrangements for executing machine instructions, e.g. instruction decode
- G06F9/38—Concurrent instruction execution, e.g. pipeline or look ahead
- G06F9/3836—Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution
- G06F9/3851—Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution from multiple instruction streams, e.g. multistreaming
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Program initiating; Program switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/4881—Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
Definitions
- This invention relates to the prioritising of instruction fetching in microprocessor systems and in particular to the implementation of this in multithreaded microprocessor systems.
- EP0891588 there is described a multithreaded microprocessor and data processing management system in which a plurality of executing threads are routed between a plurality of data inputs and a plurality of data outputs via a data processing means.
- the data processing means has access to a data storage means and the system repeatedly determines which routing operations and which data processing operations are capable of being performed, and commences execution of at least one of the routing and data processing operations on each clock cycle.
- a memory management unit is used to fetch data either from an internal memory cache or from an external memory.
- external memory has only a single data path and therefore a memory prearbiter is used to arbitrate between requests from different threads for data from memory.
- EP1738259 proposes a scheme for improved memory arbitration using various metrics attached to the various threads which are executing.
- Improvements to the type of multithreaded processor described above have introduced the concept of what we refer to as “superthreading.” This involves issuing instructions for more than one executing thread on a given clock cycle. At maximum, instructions for all the executing threads can be issued on a single clock cycle which, in a four thread implementation, would involve issuing four instructions per clock cycle. Such an implementation is, however, only fully utilised when all of the possible threads have instructions available and ready to run.
- threads are provided with an instruction buffer which may, for example, hold up to eight instructions for a thread.
- the instruction buffer is filled by using an instruction fetch routine which is able to fetch instructions in advance of the current instruction to be used and preferably is also able to determine actions such as branch or hardware loop prediction which may result from instructions.
- Preferred embodiments of the present invention address the problem of efficient arbitration between instruction fetching for different executing threads whilst also attempting to ensure that as many threads as possible have at least one instruction in their buffers for subsequent execution.
- a method for prioritising the fetching of instructions for each of a plurality of executing instruction threads in a multithreaded processor system from at least one source of instructions comprising the steps of: determining a first metric for each thread based on the number of instructions currently buffered for execution on that thread; determining an execution based second metric for each thread; determining a priority order for the threads from the first and second metric; and fetching an instruction from the source for the thread with the highest determined priority that is requesting an instruction.
- FIG. 1 shows a multithreaded processor embodying the invention
- FIG. 2 shows an instruction fetch subsystem for use in the system of FIG. 1 ;
- FIG. 3 shows an instruction fetch unit for use in the instruction fetch subsystem of FIG. 2 ;
- FIG. 4 shows a priority arbiter for use with the instruction fetch subsystem of FIG. 2 .
- the system shown in FIG. 1 comprises a multithreaded processor core. This is capable of running a number of threads each of which provide instructions via an instruction fetch unit 2 . Instructions to be executed by each of the plurality of executing threads are fetched by thread instruction fetch units 2 , one for each thread. Systems may have any number of threads but typically two or four are used at present. In general, these are arranged to operate in parallel subject to their ability to issue instructions via a thread instruction scheduler 4 , but could be used in a system where only one thread can execute on each clock cycle. There are two sources of instructions in this example, an instruction cache 6 which retrieves instructions from external memories, and a core code memory 8 which stores operating system instructions and other instructions of this nature. An arbitrator is positioned between the instruction cache and the thread instruction fetch unit 2 and further arbitrator between the core code memory 8 and the thread instruction fetch units 2 . This is used to determine which thread should be next to retrieve an instruction from each source of instructions.
- the thread instruction scheduler 4 issues instructions to, in this example, each of four executing pipelines which here are represented by arithmetic logic units (ALU) 12 . These execute the plurality of executing threads.
- ALU arithmetic logic units
- MMU Memory Management Unit
- a complete implementation may also incorporate small quantities of on chip RAM that can be accessed instead of the cache sub-system 3 .
- Some systems may incorporate instruction RAM and also data RAM.
- Multithreaded processors have been described in a number of other documents and will be well known to those skilled in the art.
- the embodiment of the invention described herein focuses on the prioritization of instruction fetch to feed an instruction scheduler 4 which is preferably capable of issuing instructions on more than one thread on any given clock cycle.
- a block diagram of a suitable system embodying the invention is shown in FIG. 2 . This shows all the main elements in an instruction fetch sub-system.
- the instruction cache 20 and on chip RAM 22 are loaded with embedded instructions for the microprocessor which may be required to run on various ones of the executing Instruction threads.
- a thread When a thread requires an instruction, it must be fetched from either the instruction cache 20 or from the on chip RAM 22 .
- a thread requesting an instruction sends its request to one of the available sources of instructions, i.e. the one containing the instruction being requested. Each thread can only issue one instruction fetch per clock cycle.
- each source of instructions may be accessed at the same time as the other by different threads.
- each source of instructions has a priority arbiter 24 coupled between it and the instruction fetch units for each of the threads.
- Each priority arbiter is coupled to an instruction fetch unit 26 for each of the threads (thread 0 to thread n). Via this connection, the priority arbiter can receive requests for instructions with associated prioritisation data and can provide fetched instructions to an instruction fetch unit 26 .
- Each instruction fetch unit 26 includes an instruction buffer which can hold up to eight instructions in readiness to feed to an instruction decoder and resource interlocking unit 28 before issuing the instruction to an instruction scheduler 30 which will pass each instruction in turn to a processor for execution.
- the priority arbiters 24 operate by receiving from each thread's instruction thread fetch unit 26 a measure of how much instruction data that instruction fetch unit 26 has left in its instruction buffer or in the process of being loaded into its instruction buffer, i.e. data relating to the number of instructions currently buffered. Therefore, if the instruction buffer can contain eight instructions, and it currently contains five instructions, data indicating that the thread currently has five instructions buffered (or space for three further instructions in its buffer) will be sent to the priority arbiter. The priority arbiter will then be able to compare this data with data received from other threads to determine which a priority order for the threads. It will then give priority to the thread with the highest priority that is also requesting an instruction.
- the priority arbiter 24 can combine this first measure of prioritisation for issuance of instructions with a further measure.
- This second measure is a more sophisticated measurement of the overall priority given to a thread. This is typically a combination of a number of different properties of a thread. These can include a deadline count, which is a time by which a thread must be executed; a delay count, which is the delay since execution of the thread was first requested; a number of instructions yet to execute, and priority, this being the nominal priority given to the thread. The deadline count is the most important of these as it determines when a thread's execution must be complete. Other factors relating to thread priority may be used in measures of this type.
- each priority arbiter 24 will combine the priority measures relating to the need for an Instruction, i.e. the measure based on the fullness of the thread's instruction buffer, and the overall thread priority, which can be based on a number of different factors. Using these, it is possible for each priority arbiter to determine an ordered list of thread priorities when requesting instructions from either the instruction cache or the on Chip RAM on each cycle such that on any given cycle the thread with the highest thus determined priority is the one that is given access to the instruction cache or the on chip RAM as appropriate. A thread will only access one of these sources of instructions at a time. It is not possible for a thread to fetch instructions from both at the same time.
- a multithreaded processor is a microprocessor with execution units, caches and memories which additionally incorporates the concept of multiple threads of executing instructions.
- a thread consists of a stream of instructions acting to utilise the resources of the microprocessor, some of which may be private to that specific thread and some of which may be shared between threads. Resources include registers, bandwidth, and space in the cache and associated memories.
- Each thread in a multithreaded processor has its own program counter, which in turn is used to control instruction fetch for that thread. Therefore, every thread has a piece of hardware to fetch instructions from instruction memory or instruction cache for execution on that thread. It would be possible for each thread to have its own instruction memory or instruction cache but cost and size are minimised by sharing the instruction memories and instruction cache between all the threads and arbitrating between them for the right to access the memory or cache.
- Simple arbitration methods such as round robin can work satisfactory up to a point. Difficulties are encountered, where different threads have different workloads and the balance between the priorities of the threads needs to be implemented such that threads with the greatest demand get priority over other threads when accessing instructions.
- the process for determining which threads need which resources is complex, particularly In super threading processors where more than one thread may be executing instructions at the same time.
- a determination is commenced, however, by passing each instructions through a pre-decode stage prior to making it available to the instruction scheduler 4 as candidate for scheduling for execution. This pre-decode stage is used to pick out the resources that the instruction will require.
- a thread may wish to use an execution unit to send a request for a data memory fetch to load or store some data. If another thread wishes to use either the same execution unit or wishes to use the data memory port for the data memory fetch, then it will be incompatible with execution of the first thread such that the two cannot be issued at the same time. If, however, another thread wishes to perform a set of different instructions such as a program branch which does not require those resources, it will be deemed to be compatible and may be issued for execution at the same time as a thread is used to request a data memory fetch.
- a result of this compatibility testing a flag is generated for each pairing of threads and the status of this flag is determines whether or not threads are compatible. Therefore, the result of testing each thread against all other threads is that for each pair of threads a compatibility flag is generated the status of which determines whether or not threads are compatible for execution, i.e. whether or not only two threads overlapping resource requirements at the same time.
- the threads have a ranked ordering.
- the compatibility status information is used such that each rank in this order, a final compatibility set is prepared by considering each of the lower ranks in turn. If a rank is compatible where the highest priority thread it is added to the set for execution, otherwise it is skipped and the next thread considered.
- a final step to actual scheduling of instructions is that the determination is then made of the highest ranking thread which is able to be issued, i.e. it is not waiting on a resource or register to become available. This thread along with its super threading set (other threads which may execute with it) is then issued together.
- the priority arbiter is configured to give preference, when appropriate, to fetching instructions for threads that have nothing in their thread instruction buffers or are running out of thread instructions in their thread instruction buffers. These threads then take precedence over other threads when fetching instructions, including threads which have a higher overall priority ranking.
- FIG. 3 A preferred embodiment of the thread instruction fetch unit is shown in FIG. 3 .
- Each thread is provided with such an instruction fetch unit. This particular example is shown with reference to obtaining instructions from the instruction cache, it could equally be used with the on chip RAM such as that shown in FIG. 2 .
- the instruction fetch unit comprises an instruction address feed unit 32 , which sends requests for instructions to the instruction cache.
- the instruction fetch engine may be configured to request instructions with sufficient frequency to reduce or eliminate the latency altogether from the stream of issuing instructions.
- each thread instruction fetch unit incorporates an instruction buffer able to hold the returning data for a fixed number of instruction requests. As many instruction requests may be in transit as there are slots in the instruction buffer may be stored.
- the instruction buffer 34 is shown in FIG. 3 . After the instruction buffer instructions pass via an instruction data feed unit 36 to the instruction scheduler.
- data is extracted from the thread instruction fetch unit relating to the fullness of the instruction buffer 38 .
- This may be a data value indicating the number of available slots in the instruction buffer or a data value indicating how many slots have been already occupied.
- This data is provided to a priority arbiter 24 of the type shown in FIG. 2 and is used by the priority arbiter to prioritise instruction fetches for threads which are running short of instructions in comparison with other threads. This then ensures that the instruction buffer for that thread has further instructions provided to it as early as possible so that a larger number of instruction are available to the instruction scheduler.
- the instruction fetch unit of FIG. 3 can be used for each of a number of threads and will provide data that indicates which thread has the instruction buffer with the most instructions in it and which has the least. These are compared by the priority arbiter produce a prioritised list. The thread which has fewest instructions in its instruction buffer can then be prioritised for the next instruction fetch on the next clock cycle (providing it is requesting an instruction). At the same time, at least one of the threads will issue an instruction to the instruction scheduler and this may in turn change the priority on the next dock cycle. Thus the thread with the highest priority which is also requesting an instruction from a source of instructions will fetch an instruction.
- this system prioritises threads by sorting them into an order based upon the fullness of the instruction buffer 34 in each instruction fetch unit.
- Preferred embodiments of the invention modify this by adding an additional level of arbitration.
- a priority arbiter of the type shown in FIG. 2 is illustrated in more detail in FIG. 4 .
- the priority arbiter 24 receives instruction requests and grants instructions from its associated on chip RAM or instruction cache to each of the instruction fetch engines for the respective threads provided in the multithreaded processor.
- An input to the priority arbiter is the prioritized order determined for the instruction fetch engines for each thread as determined by the fullness of their respective instruction buffers. This may be derived in a separate unit. Alternatively, the relevant data values may be provided to the priority arbiter itself which may then compute the necessary priority order.
- each instruction fetch unit may only fetch enough instructions to fill its instruction buffer, any thread with a full instruction buffer will not request a new fetch on each cycle. Therefore, the priority arbiter will determine the highest priority thread which is requesting a new instruction via its request signal.
- the priority order determined by the priority arbiter 24 or received from a separate unit, is combined with the additional measures for priority discussed above which we refer to at the automatic million instructions per second (MIPS) allocation (AMA) metric.
- MIPS automatic million instructions per second
- AMA automatic million instructions per second
- This may be a combination of e.g. deadline count, delay count and possibly other factors effecting the execution of a thread. Effectively, it is an execution based metric which can be combined with the priority metric determined from the fullness of the instruction buffers.
- the AMA metric is simplified into a simple ordering of the threads, i.e. it is a sorted list of threads. This is in contrast to the fullness of the instruction buffer for each thread which has a list of data regarding buffer capacity.
- the buffer capacity is extremely important and a list is formed and sorted into order of fullest to emptiest of instruction buffers for threads.
- the emptiest is the higher priority. If all of the threads have a different buffer fullness then, in a preferred embodiment of the invention, this is the end of the process.
- the AMA metric is used only to determine the correct order of threads which have the same buffer fullness.
- each thread will have an equal need for an instruction to be fetched. Therefore, the arbiter must then consider the AMA metrics of each thread. These will be ranked 3, 2, 1, and 0. If 3 is the higher priority metric and 0 is the lowest, the thread which has an AMA of 3 will have the highest priority for the next instruction fetch.
- threads 0, 1, 2, and 3 have a need for instructions which are given priorities 1, 2, 2, 0 respectively and the AMA metrics for each thread are 3, 2, 1, and 0 respectively, then the final priority order will be 1, 3, 2, and 0. This is because threads 1 and 2 have the highest need for instructions and of these two threads, thread 1 has an AMA priority of 3, and thread 2 has an AMA priority of 2.
- each of the threads is checked internally each clock cycle to determine whether or not they are requesting an instruction from any particular memory. If the highest priority thread is requesting an instruction and is not blocked in some way then the request is issued. This will alter the need metric which will be employed on future cycles. If the highest priority thread is not making a request then all other priorities will be considered in turn to see if they can be used. If no thread can request an instruction then no request is made on that cycle.
- an instruction Once an instruction has been fetched, it is supplied directly from the on chip RAM or the instruction cache to the thread instruction fetch unit.
- the order in which instructions reach the instruction fetch units may vary in dependence on the source of instructions being fetched and their location within each source.
- the instruction cache and the instruction memory may respectively be accessed by different threads at the same time. This necessitates two priority arbiters.
- the priority order list for the arbiters may be shared between the two devices.
- preferred embodiments provide an arbiter for each source of instructions which is coupled to each of the instruction fetch units for executing threads. These receive instruction fetch requests from each thread which requires an instruction at various times and arbiter between requests to determine which thread should receive an instruction from that source. Each thread can only issue one instruction fetch per clock cycle. Preferably, a thread will not request an instruction from a different source while it still has requests outstanding from a first source.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Advance Control (AREA)
- Memory System Of A Hierarchy Structure (AREA)
Abstract
Description
Claims (19)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/134,510 US9870228B2 (en) | 2008-02-07 | 2016-04-21 | Prioritising of instruction fetching in microprocessor systems |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB0802314.5 | 2008-02-07 | ||
GB0802314A GB2457265B (en) | 2008-02-07 | 2008-02-07 | Prioritising of instruction fetching in microprocessor systems |
US12/322,942 US9348600B2 (en) | 2008-02-07 | 2009-02-09 | Prioritising of instruction fetching in microprocessor systems |
US15/134,510 US9870228B2 (en) | 2008-02-07 | 2016-04-21 | Prioritising of instruction fetching in microprocessor systems |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/322,942 Continuation US9348600B2 (en) | 2008-02-07 | 2009-02-09 | Prioritising of instruction fetching in microprocessor systems |
Publications (2)
Publication Number | Publication Date |
---|---|
US20160232007A1 US20160232007A1 (en) | 2016-08-11 |
US9870228B2 true US9870228B2 (en) | 2018-01-16 |
Family
ID=39204442
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/322,942 Active 2031-08-01 US9348600B2 (en) | 2008-02-07 | 2009-02-09 | Prioritising of instruction fetching in microprocessor systems |
US15/134,510 Active 2029-04-21 US9870228B2 (en) | 2008-02-07 | 2016-04-21 | Prioritising of instruction fetching in microprocessor systems |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/322,942 Active 2031-08-01 US9348600B2 (en) | 2008-02-07 | 2009-02-09 | Prioritising of instruction fetching in microprocessor systems |
Country Status (5)
Country | Link |
---|---|
US (2) | US9348600B2 (en) |
EP (1) | EP2240850B1 (en) |
JP (1) | JP5607545B2 (en) |
GB (1) | GB2457265B (en) |
WO (1) | WO2009098489A1 (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101325063B (en) * | 2007-06-12 | 2011-02-16 | 建兴电子科技股份有限公司 | Method for searching locating point position in holography storage system |
JP5173711B2 (en) * | 2008-09-30 | 2013-04-03 | ルネサスエレクトロニクス株式会社 | Multi-thread processor and hardware thread scheduling method |
TWI474280B (en) * | 2010-04-21 | 2015-02-21 | Via Tech Inc | System and method for improving throughput of a graphics processing unit |
US8904115B2 (en) * | 2010-09-28 | 2014-12-02 | Texas Instruments Incorporated | Cache with multiple access pipelines |
US9189283B2 (en) | 2011-03-03 | 2015-11-17 | Hewlett-Packard Development Company, L.P. | Task launching on hardware resource for client |
US9645823B2 (en) | 2011-03-03 | 2017-05-09 | Hewlett-Packard Development Company, L.P. | Hardware controller to choose selected hardware entity and to execute instructions in relation to selected hardware entity |
US8738830B2 (en) | 2011-03-03 | 2014-05-27 | Hewlett-Packard Development Company, L.P. | Hardware interrupt processing circuit |
JP5803972B2 (en) | 2013-04-18 | 2015-11-04 | 株式会社デンソー | Multi-core processor |
JP6384380B2 (en) | 2015-03-27 | 2018-09-05 | 富士通株式会社 | Arithmetic processing device and control method of arithmetic processing device |
US20170139716A1 (en) | 2015-11-18 | 2017-05-18 | Arm Limited | Handling stalling event for multiple thread pipeline, and triggering action based on information access delay |
US10353711B2 (en) * | 2016-09-06 | 2019-07-16 | Apple Inc. | Clause chaining for clause-based instruction execution |
CN116324716A (en) * | 2020-10-14 | 2023-06-23 | 瑞典爱立信有限公司 | Apparatus and method for simultaneous multithreading instruction scheduling in a microprocessor |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0583089A2 (en) | 1992-08-12 | 1994-02-16 | Advanced Micro Devices, Inc. | Instruction decoder |
EP0891588B1 (en) | 1996-04-04 | 2001-12-12 | Imagination Technologies Limited | A data processing management system |
GB2372349A (en) | 2001-02-19 | 2002-08-21 | Imagination Tech Ltd | Control of priority and instruction rates on a multithreaded processor |
US20040003203A1 (en) | 2002-06-28 | 2004-01-01 | Fujitsu Limited | Instruction fetch control device and instruction fetch control method |
US6938147B1 (en) | 1999-05-11 | 2005-08-30 | Sun Microsystems, Inc. | Processor with multiple-thread, vertically-threaded pipeline |
US20050235126A1 (en) | 2004-04-16 | 2005-10-20 | Ko Yin N | Dynamic load balancing |
US20060101238A1 (en) | 2004-09-17 | 2006-05-11 | Pradip Bose | Adaptive fetch gating in multithreaded processors, fetch control and method of controlling fetches |
US20060179274A1 (en) | 2005-02-04 | 2006-08-10 | Mips Technologies, Inc. | Instruction/skid buffers in a multithreading microprocessor |
US7657883B2 (en) | 2005-02-04 | 2010-02-02 | Mips Technologies, Inc. | Instruction dispatch scheduler employing round-robin apparatus supporting multiple thread priorities for use in multithreading microprocessor |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07281954A (en) * | 1994-04-11 | 1995-10-27 | Hitachi Ltd | Fast pc system |
JP4631442B2 (en) * | 2005-01-13 | 2011-02-16 | セイコーエプソン株式会社 | Processor |
US7490230B2 (en) * | 2005-02-04 | 2009-02-10 | Mips Technologies, Inc. | Fetch director employing barrel-incrementer-based round-robin apparatus for use in multithreading microprocessor |
-
2008
- 2008-02-07 GB GB0802314A patent/GB2457265B/en not_active Expired - Fee Related
-
2009
- 2009-02-09 WO PCT/GB2009/000360 patent/WO2009098489A1/en active Application Filing
- 2009-02-09 JP JP2010545556A patent/JP5607545B2/en active Active
- 2009-02-09 US US12/322,942 patent/US9348600B2/en active Active
- 2009-02-09 EP EP09708438.8A patent/EP2240850B1/en not_active Not-in-force
-
2016
- 2016-04-21 US US15/134,510 patent/US9870228B2/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0583089A2 (en) | 1992-08-12 | 1994-02-16 | Advanced Micro Devices, Inc. | Instruction decoder |
EP0891588B1 (en) | 1996-04-04 | 2001-12-12 | Imagination Technologies Limited | A data processing management system |
US6938147B1 (en) | 1999-05-11 | 2005-08-30 | Sun Microsystems, Inc. | Processor with multiple-thread, vertically-threaded pipeline |
GB2372349A (en) | 2001-02-19 | 2002-08-21 | Imagination Tech Ltd | Control of priority and instruction rates on a multithreaded processor |
US20040003203A1 (en) | 2002-06-28 | 2004-01-01 | Fujitsu Limited | Instruction fetch control device and instruction fetch control method |
US20050235126A1 (en) | 2004-04-16 | 2005-10-20 | Ko Yin N | Dynamic load balancing |
WO2005101203A1 (en) | 2004-04-16 | 2005-10-27 | Imagination Technologies Limited | Dynamic load balancing |
EP1738259A1 (en) | 2004-04-16 | 2007-01-03 | Imagination Technologies Limited | Dynamic load balancing |
US20060101238A1 (en) | 2004-09-17 | 2006-05-11 | Pradip Bose | Adaptive fetch gating in multithreaded processors, fetch control and method of controlling fetches |
US20060179274A1 (en) | 2005-02-04 | 2006-08-10 | Mips Technologies, Inc. | Instruction/skid buffers in a multithreading microprocessor |
US7657883B2 (en) | 2005-02-04 | 2010-02-02 | Mips Technologies, Inc. | Instruction dispatch scheduler employing round-robin apparatus supporting multiple thread priorities for use in multithreading microprocessor |
Non-Patent Citations (2)
Title |
---|
Eggers et al., "Simultaneous Multithreading: A Platform for Next-Generation Processors," IEEE Micro, Sep./Oct. 1997, pp. 12-19. |
Lo et al., "Converting Thread-Level Parallelism to Instruction-Level Parallelism via Simultaneous Multithreading," ACM Transactions on Computer Systems, vol. 15, No. 3, Aug. 1997, pp. 322-354. |
Also Published As
Publication number | Publication date |
---|---|
US20090210660A1 (en) | 2009-08-20 |
JP2011511378A (en) | 2011-04-07 |
GB2457265B (en) | 2010-06-09 |
WO2009098489A1 (en) | 2009-08-13 |
GB0802314D0 (en) | 2008-03-12 |
US20160232007A1 (en) | 2016-08-11 |
US9348600B2 (en) | 2016-05-24 |
GB2457265A (en) | 2009-08-12 |
JP5607545B2 (en) | 2014-10-15 |
EP2240850B1 (en) | 2016-08-10 |
EP2240850A1 (en) | 2010-10-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9870228B2 (en) | Prioritising of instruction fetching in microprocessor systems | |
US8082420B2 (en) | Method and apparatus for executing instructions | |
EP2425329B1 (en) | Method and apparatus for scheduling the issue of instructions in a multithreaded microprocessor | |
US7734897B2 (en) | Allocation of memory access operations to memory access capable pipelines in a superscalar data processing apparatus and method having a plurality of execution threads | |
US7822885B2 (en) | Channel-less multithreaded DMA controller | |
US8095779B2 (en) | System and method for optimization within a group priority issue schema for a cascaded pipeline | |
US11782720B2 (en) | Processor architecture with micro-threading control by hardware-accelerated kernel thread | |
US8108654B2 (en) | System and method for a group priority issue schema for a cascaded pipeline | |
US10545892B2 (en) | Multi-thread processor and its interrupt processing method | |
JP6260303B2 (en) | Arithmetic processing device and control method of arithmetic processing device | |
US7818747B1 (en) | Cache-aware scheduling for a chip multithreading processor | |
US9401869B1 (en) | System and methods for sharing memory subsystem resources among datacenter applications | |
JP2012512472A (en) | Resolving contention between data bursts | |
US8006073B1 (en) | Simultaneous speculative threading light mode | |
US7725659B2 (en) | Alignment of cache fetch return data relative to a thread | |
US20140258680A1 (en) | Parallel dispatch of coprocessor instructions in a multi-thread processor | |
US8918786B2 (en) | Generating simulated stall signals based on access speed model or history of requests independent of actual processing or handling of conflicting requests | |
US6915516B1 (en) | Apparatus and method for process dispatching between individual processors of a multi-processor system | |
JP4631442B2 (en) | Processor | |
US11907126B2 (en) | Processor with multiple op cache pipelines | |
CN118585312A (en) | Instruction scheduling method and device, processor, electronic device and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL) |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: HELLOSOFT LIMITED, UNITED KINGDOM Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:IMAGINATION TECHNOLOGIES LIMITED;REEL/FRAME:045136/0975 Effective date: 20171006 |
|
AS | Assignment |
Owner name: MIPS TECH LIMITED, UNITED KINGDOM Free format text: CHANGE OF NAME;ASSIGNOR:HELLOSOFT LIMITED;REEL/FRAME:045168/0922 Effective date: 20171108 |
|
AS | Assignment |
Owner name: MIPS TECH, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MIPS TECH LIMITED;REEL/FRAME:045555/0317 Effective date: 20180216 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.) |
|
AS | Assignment |
Owner name: WAVE COMPUTING LIQUIDATING TRUST, CALIFORNIA Free format text: SECURITY INTEREST;ASSIGNORS:WAVE COMPUTING, INC.;MIPS TECH, LLC;MIPS TECH, INC.;AND OTHERS;REEL/FRAME:055429/0532 Effective date: 20210226 |
|
AS | Assignment |
Owner name: CAPITAL FINANCE ADMINISTRATION, LLC, ILLINOIS Free format text: SECURITY INTEREST;ASSIGNORS:MIPS TECH, LLC;WAVE COMPUTING, INC.;REEL/FRAME:056558/0903 Effective date: 20210611 Owner name: MIPS TECH, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 Owner name: HELLOSOFT, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 Owner name: WAVE COMPUTING (UK) LIMITED, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 Owner name: IMAGINATION TECHNOLOGIES, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 Owner name: CAUSTIC GRAPHICS, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 Owner name: MIPS TECH, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 Owner name: WAVE COMPUTING, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:WAVE COMPUTING LIQUIDATING TRUST;REEL/FRAME:056589/0606 Effective date: 20210611 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: WAVE COMPUTING INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CAPITAL FINANCE ADMINISTRATION, LLC, AS ADMINISTRATIVE AGENT;REEL/FRAME:062251/0251 Effective date: 20221229 Owner name: MIPS TECH, LLC, CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CAPITAL FINANCE ADMINISTRATION, LLC, AS ADMINISTRATIVE AGENT;REEL/FRAME:062251/0251 Effective date: 20221229 |