WO2005013039A2 - Prefetch control in a data processing system - Google Patents

Prefetch control in a data processing system Download PDF

Info

Publication number
WO2005013039A2
WO2005013039A2 PCT/US2004/022438 US2004022438W WO2005013039A2 WO 2005013039 A2 WO2005013039 A2 WO 2005013039A2 US 2004022438 W US2004022438 W US 2004022438W WO 2005013039 A2 WO2005013039 A2 WO 2005013039A2
Authority
WO
WIPO (PCT)
Prior art keywords
prefetch
master
buffer
limit
prefetches
Prior art date
Application number
PCT/US2004/022438
Other languages
English (en)
French (fr)
Other versions
WO2005013039A3 (en
Inventor
William C Moyer
Lea Hwang Lee
Afzal M. Malik
Original Assignee
Freescale Semiconductor, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Freescale Semiconductor, Inc. filed Critical Freescale Semiconductor, Inc.
Priority to EP04778111A priority Critical patent/EP1652092A2/de
Priority to JP2006521871A priority patent/JP4875981B2/ja
Priority to KR1020067002041A priority patent/KR101093317B1/ko
Publication of WO2005013039A2 publication Critical patent/WO2005013039A2/en
Publication of WO2005013039A3 publication Critical patent/WO2005013039A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F13/14Handling requests for interconnection or transfer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0215Addressing or allocation; Relocation with look ahead addressing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Definitions

  • the present invention relates to a data processing system, and more particularly, to prefetch control within a data processing system.
  • Prefetching is commonly used to access information within a data processing system.
  • prefetching information in advance of a request for that information, the latency caused by accessing the information requested by the bus master may be reduced.
  • a number of prefetches are wasted since the bus master may not request access to the prefetched information.
  • Another disadvantage of general prefetching schemes is that the prefetch limit control provided by such systems is based on a fixed policy, allowing less flexibility and control. Therefore, a need exists for a method of prefetching that reduces the number of prefetches that are wasted, resulting in a reduced amount of power consumption and an optimization of data processor performance.
  • FIG. 1 illustrates, in block diagram form, a data processing system in accordance with one embodiment of the present invention
  • FIG. 2 illustrates, in block diagram form, a control register of the data processing system of FIG. 1, in accordance with one embodiment of the present invention
  • FIG. 3 illustrates, in block diagram form, prefetch counters of the data processing system of FIG. 1, in accordance with one embodiment of the present invention
  • FIG. 4 illustrates, in table form, field descriptions of the control register of FIG. 2 in accordance with one embodiment of the present invention.
  • FIG. 5 illustrates, in flow diagram form, operation of the data processing system of FIG. 1 in accordance with one embodiment of the present invention.
  • bus is used to refer to a plurality of signals or conductors which may be used to transfer one or more various types of information, such as data, addresses, control, or status.
  • the conductors as discussed herein may be illustrated or described in reference to being a single conductor, a plurality of conductors, unidirectional conductors, or bidirectional conductors. However, different embodiments may vary the implementation of the conductors. For example, separate unidirectional conductors may be used rather than bidirectional conductors and vice versa.
  • plurality of conductors may be replaced with a single conductor that transfers multiple signals serially or in a time multiplexe Ashd_manner._
  • One embodiment of the present invention relates to dynamically controlling the number of sequential prefetch operations in order to prevent wasted prefetches. For example, by limiting the number of sequential prefetches performed after a buffer miss, the performance and power consumption of the memory control unit can be optimized.
  • limiting the number of sequential prefetches is accomplished by using a set of prefetch buffers, a control register, and prefetch counters, all of which will be further described below.
  • a software programmable control scheme is provided which allows for dynamic control of the number of sequential prefetch operations between buffer misses and may be preprogrammed by the user based upon the properties of the requesting master. For example, within a central processing unit (CPU) instructions are typically executed sequentially until a change of flow instruction or an exception is reached, which causes a discontinuity in sequential execution. The number of instructions executed sequentially is dependent upon the CPU and is a function of the type of application program being executed.
  • a DMA master is provided in which transfers occur sequentially until an end-of -transfer or a channel switch occurs. For this example, long sequences are expected and the prefetch limit corresponding to the DMA master can be adjusted accordingly.
  • One embodiment of the present invention relates to a data processing system including a master, storage circuitry coupled to the master, a control storage circuit which stores a prefetch limit, a prefetch buffer, and prefetch circuitry coupled to the control storage circuit, prefetch buffer, and storage circuitry.
  • the prefetch circuitry selectively prefetches a predetermined number of lines from the storage circuitry into the prefetch buffer, wherein the prefetch limit controls the number of prefetches that occur between misses in the prefetch buffer.
  • Another embodiment of the present invention relates to a method for performing prefetch in a data processing system.
  • a plurality of access requests from a master to access storage circuitry is received and a prefetch limit is used to limit a number of prefetches performed between misses in a prefetch buffer resulting from at least a portion of the plurality of access requests.
  • a prefetch control circuit is provided to store the prefetch limit.
  • Yet another embodiment of the present invention relates to a method for performing a prefetch in a data processing system in which a read request from a master to access storage circuitry is received and it is determined whether the read request results in a hit or a miss in a prefetch buffer.
  • FIG. 1 illustrates one embodiment of data processing system 10.
  • Data processing system 10 includes a master 12 (also referred to as an interconnect master 12), a master 14 (also referred to as an interconnect master 14), a memory array controller 36, a system interconnect 22, I/O circuitry 16, a peripheral 18, other slaves 20, and a memory array 35.
  • Master 12 is bidirectionally coupled to system interconnect 22 via conductors 48
  • master 14 is bidirectionally coupled to system interconnect 22 via conductors 50
  • I/O circuitry is bidirectionally coupled to system interconnect 22 via conductors 52
  • peripheral 18 is bidirectionally coupled to system interconnect 22 via conductors 54
  • other slaves 20 is bidirectionally coupled to system interconnect 22 via conductors 56
  • memory array controller 36 is bidirectionally coupled to system interconnect 22 via conductors 24.
  • Conductors 24 include conductors for communicating a master identifier 26, address/data 27, a R/W signal 28, a burst signal 30, an instruction/data signal 32, and other signals 34.
  • Memory array controller 36 includes a control register 38, request monitor 43, prefetch circuitry 40, prefetch counters 41, and buffer 42 (also referred to as prefetch buffers), and is bidirectionally coupled to memory array 35 via conductors 33.
  • Control register 38 is coupled to request monitor 43 and prefetch circuitry 40, which is coupled to buffer 42 and prefetch counters 41.
  • Buffer 42 includes a buffer 44 and a buffer 46.
  • any number of masters and slaves may be coupled to system interconnect 22 and are not limited to those shown in FIG. 1.
  • all of data processing system 10 may be located on a single integrated circuit or within a same device.
  • data processing system 10 may include any number of separate integrated circuits or separate devices interconnected with each other.
  • the memory and memory controller (such as, for example, memory array 35 and memory array controller 36) may be located on one or more integrated circuits, separate from the rest of data processing system 10.
  • bus master 12 and bus master 14 may be processors capable of executing instructions, such as microprocessors, digital signal processors, etc., or may be any other type of interconnect master, such as direct memory access (DMA) circuitry or debug circuitry.
  • Peripheral 18 may be any type of peripheral, such as a universal asynchronous receiver transmitter (UART), a real time clock (RTC), a keyboard controller, etc.
  • other slaves 20 may include any type of interconnect slaves, such as, for example, a memory accessible by masters 12 and 14, as well as any type of peripheral which resides on the system bus, including the same types of peripherals as peripheral 18.
  • I/O circuitry 16 may include any type of I/O circuitry which receives or provides information external to data processing system 10.
  • memory array controller 36 and memory array 35 correspond to another slave coupled to system interconnect 22.
  • memory array 35 can include any number of arrays.
  • memory array 35 may be referred to as storage circuitry 35.
  • Memory array 35 may be located on a same integrated circuit as masters 12 and 14 or on a separate integrated circuit.
  • memory array 35 may be any type of memory, such as, for example, a read only memory (ROM), a random access memory (RAM), non-volatile memory (e.g. Flash), etc.
  • ROM read only memory
  • RAM random access memory
  • non-volatile memory e.g. Flash
  • memory array 35 may be a memory or other storage located within another peripheral or slave.
  • System interconnect 22 interconnects master 12, master 14, I/O circuitry 16, peripheral 18, other slaves 20, and memory array controller 36.
  • system interconnect 22 is implemented as a system bus operating according to a system bus protocol.
  • system interconnect 22 can be implemented using interconnect circuitry, such as, for example, switching circuitry, which routes information between the various devices.
  • masters 12 and 14 request access of system interconnect 22 to request access to other slaves 20, to peripherals 18, or to memory array 35 via memory array controller 36.
  • a requesting master can provide an access request, via system interconnect 22, to memory array controller 36.
  • the access request can be, for example, a read request or a write request for either data or instructions.
  • Memory array controller 36 in response to a read access request, provides the requested information (data or instructions) back to the requesting master via system interconnect 22.
  • the read access request from a requesting master may also be referred to as a demand fetch.
  • a master identifier 26 is provided to memory array controller 36 which identifies which master is requesting the current access.
  • R/W signal 28 is also provided to memory array controller 36 to indicate whether the current access request is for a read or a write type of access.
  • Burst signal 30 is provided to memory array controller 36 to indicate whether the current access request is a burst or a non-burst type of access.
  • Instruction/data signal 32 is provided to memory array controller 36 to indicate whether the current access request is for an instruction or data.
  • Memory array controller 36 also receives address information corresponding to the current access request and provides the requested information via address/data 27.
  • prefetch circuitry 40 may prefetch information from memory array 35 into buffer 42, such as buffer 44 and buffer 46. Therefore, in response to a read access request from a requesting master, memory array controller 36 may be able to provide the requested information from buffer 42 (if the information was already prefetched) rather than having to fetch the information from memory array 35, which generally has a longer access time as compared to buffer 42.
  • prefetch circuitry includes tag portions (not shown) corresponding to each of buffers 44 and 46, and comparison circuitry (not shown) in order to determine whether the information being requested in a current access request is already located in one of buffers 44 and 46. For example, prefetch circuitry can compare the incoming address corresponding to the current access request via address/data 27 with the tag portions to determine whether or not the information has already been prefetched. If the information has not been prefetched, memory array controller 36 can provide the requested information from memory array 35. In the illustrated embodiment, two buffers are illustrated (buffer 44 and buffer 46). However, in alternate embodiments, buffer 42 may include any number (one or more) of buffers.
  • prefetch circuitry may prefetch information from memory array 35 into one of buffers 44 and 46 while providing information to a requesting master from another one of buffers 44 and 46. That is, the use of more than one buffer allows for a prefetch to memory array 35 to occur at least partially simultaneously with servicing an • access request from a requesting master. Since providing information from buffer 42 has a reduced access time as compared to accessing memory array 35 for responding to an access request from a requesting master, it is desirable to prefetch information from memory array 35 into buffer 42 that will subsequently be requested. However, note that prefetching is speculative in that it is uncertain whether the prefetched information will actually be requested by a master.
  • one embodiment of the present invention controls prefetching of speculative data into buffer 42 utilizing prefetch counters 41 and control register 38 by limiting the number of prefetches between misses in buffer 42 for each master (such as masters 12 and 14). For example, one embodiment uses a control register to store a prefetch limit for each master such that the number of prefetches for each master is based on properties of the requesting master. In another embodiment, the control register may also be used to determine how many lines are prefetched during each prefetch, as will be discussed in more detail below. FIG.
  • control register 38 which includes a prefetch limit field for each master.
  • control register 38 includes master 12 prefetch limit field 60 and master 14 prefetch limit field 62.
  • control register 38 may include more or less fields, as needed, to store the desired prefetch limits.
  • control register 38 may be programmed via instructions from a master such as master 12 or master 14 which are coupled to system interconnect 22.
  • the prefetch limit field values may be provided, for example, by a user or may be programmed at design time.
  • prefetch circuitry 40 is bidirectionally coupled to request monitor 43, which profiles access requests from one or more masters, such as master 12 or master 14, whose profiles may be used to dynamically update the prefetch limit.
  • request monitor 43 may monitor the number of accesses made to buffer 42 between successive buffer misses and update or set the limits accordingly.
  • request monitor 43 can be excluded from the data processing system 10 and the prefetch limit can be, for example, user programmable, or hardwired.
  • FIG. 3 illustrates one embodiment of prefetch counters 41, which includes a prefetch counter corresponding to each master.
  • prefetch counters 41 include a prefetch counter 64 corresponding to master 12, and a prefetch counter 66 corresponding to master 14.
  • data processing system 10 may include a single prefetch counter per memory array controller (such as memory array controller 36) that is shared by one or more masters, such as master 12 or master 14.
  • Prefetch counters 41 can alternatively be referred to as a counter, counters, or any other device (software or hardware) that operates similar to a counter.
  • FIG. 4 illustrates one embodiment of the field definitions of control register 38 of HG. 2.
  • each of fields 60 and 62 are three-bit fields, where each field is capable of having 8 values (000, 001, 010, 011, 100, 101, 110, and 111).
  • prefetch limit fields 60 and 62 are used to control the prefetch algorithm used by the prefetch circuitry.
  • prefetch limit fields 60 and 62 define a limit on the maximum number of sequential prefetches which will be attempted between buffer misses for each corresponding master.
  • Prefetch limit fields 60 and 62 may also be used to define how many lines are prefetched each time.
  • a value of 000 indicates that no prefetching is performed between access requests from a corresponding master that result in misses in buffer 42. That is, if master 12 prefetch limit field 60 is set to 000, then no prefetching occurs between misses in buffer 42 caused by access requests from master 12. Similarly, if master 14 prefetch limit field 62 is set to 000, then no prefetching occurs between misses in buffer 42 caused by access requests from master 14. A value of 001 for prefetch limit fields 60 and 62 indicates that a single additional line (next sequential line in memory array 35) is prefetched on a buffer miss.
  • a value of 010 for prefetch limit fields 60 and 62 indicates that up to two additional lines may be prefetched following each buffer miss before prefetching is halted, where a single additional line (next sequential line in memory array 35) is prefetched on a buffer miss, and the next additional 'sequential line is prefetched on a buffer hit (if not already present). Still referring to the embodiment of FIG. 4, a value of 011 for prefetch limit fields 60 and 62 indicates that up to three additional lines may be prefetched following each buffer miss before prefetching is halted.
  • a single prefetch may be initiated after the initial miss, and a single additional prefetch may be initiated after each subsequent buffer hit (up to a total of three single prefetches).
  • a value of 100 for prefetch limit fields 60 and 62 indicates up to four additional lines may be prefetched following each buffer miss before prefetching is halted. That is, for example, a single prefetch may be initiated after the initial miss, and a single additional prefetch may be initiated after each subsequent buffer hit (up to a total of four single prefetches).
  • a value of 101 for prefetch limit fields 60 and 62 indicates up to five additional lines may be prefetched following each buffer miss before prefetching is halted.
  • a single prefetch may be initiated after the initial miss, and a single additional prefetch may be initiated after each subsequent buffer hit (up to a total of five single prefetches).
  • a value of 110 for prefetch limit fields 60 and 62 indicates an unlimited number of additional lines may be prefetched following each buffer miss. That is, for example, a single additional prefetch is initiated after each buffer hit or miss. (Note that in this embodiment, the value of 111 is reserved and is not being used to set a prefetch operation).
  • control register 38 may include more or less fields for each master utilizing more or less bits as needed.
  • FIG. 5 illustrates, in flow diagram form, operation of data processing system 10 in accordance with one embodiment of the present invention.
  • Flow 70 begins with start 72 and proceeds to block 74 where an access request from a master, such as master 12 or master 14, is received.
  • a master such as master 12 or master 14
  • This access request can be many different types of access requests, such as a read request, a write request, a burst request, a non-burst request, a request for data, a request for instructions, etc. However, for ease of explanation herein, it will be assumed that the access request is a read request.
  • master identifier as provided by master identifier 26 of FIG. 1.
  • different masters may have different master identifiers, for example, master 12 may have an identifier of 0 and master 14 may have an identifier of 1. Therefore, each master can be assigned a unique identifier.
  • some masters may share a same identifier.
  • which master is requesting the access may be determined in different ways rather than by providing a signal such as master identifier 26.
  • the correct prefetch counter and corresponding prefetch limit can be selected from a plurality of prefetch counters and limits based upon the identity of the master identifier. Referring back to FIG. 5, flow then proceeds to decision diamond 78, where it is determined whether the access request results in a hit or miss. If the access request results in a miss, flow then proceeds to block 86, where the prefetch counter (one of the counters in prefetch counters 41) corresponding to the requesting master is set to the prefetch limit (e.g., indicated by one of the prefetch limit fields in control register 38) corresponding to the requesting master. This prefetch limit is used to control the number of prefetches between buffer misses of the requesting master.
  • the prefetch limit may be used to limit the number of prefetches between buffer misses for a corresponding requesting master. Flow then proceeds to ending oval 90.
  • the prefetch counter is decremented by a fixed value, for example, 1, to indicate a prefetch was performed (e.g., in blocks 82 and 89 of FIG.5). The prefetch counter is considered expired when the value of the prefetch counter reaches its lower limit, in this case 0.
  • the initial value of the prefetch counter may be set to 0 (such as in block 86 of FIG. 5), and the prefetch counter may therefore be incremented by a fixed value, for example, 1, to indicate a prefetch was performed (such as in blocks 82 and 89 of FIG. 5).
  • the prefetch counter is considered expired when the value of the counter reaches the corresponding prefetch limit. Note also that in the embodiment of FIG. 5, only a single additional line is fetched during each prefetch (such as in blocks 82 and 89) as defined by the fields of FIG. 4.
  • a predetermined number of lines may be fetched during each prefetch (such as in blocks 82 and 89 of FIG. 5). That is, a predetermined number of lines may be prefetched in response to a hit, a miss, or both, if the lines are not already present in the prefetch buffer. Also, the corresponding prefetch counter may be decremented (or incremented) accordingly so as to keep track of the number of prefetches.
  • prefetch limit fields such as prefetch limit fields 60 and 62 in control register 38 may be used to define a prefetch of any predetermined number of lines after each miss or hit or both (until the corresponding counter expires) rather that just a single additional line.
  • a prefetch counter may be used for each master, however, in an alternate embodiment, a single counter may be shared by multiple masters.
  • a priority scheme may be used where, for example, the priority scheme may be programmed by the user as to which master takes control of the single counter. For example, master 14 may require use of the counter that is currently being used by master 12 and if master 14 has priority over master 12, then master 12 can relinquish control of the counter to master 14.
  • multiple masters may share a single counter.
  • the counter may be shared such that a prefetch limit may limit the number of prefetches between successive misses, regardless of the identity of the requesting master, rather than on a per master basis.
  • the shared counter would be set (such as in block 86 of FIG. 5) each time a miss occurs, regardless of the identity of the requesting master.
  • decision diamond 80 and blocks 82 and 89 of FIG. 5 would operate on this shared counter, regardless of the identity of the requesting master.
  • multiple masters may share a prefetch limit in control register 38. When dealing with different types of masters in a data processing system it may be necessary to optimize performance of the memory control unit.
  • a prefetching limitation scheme that takes into account different access characteristics of the masters and limits the number of prefetches between misses, allowing for dynamic control of the sequential prefetches. For example, instructions in a CPU are typically performed sequentially until a change of flow has been reached, whereas, with a DMA, most transfers occur sequentially until an end-of -transfer or a channel switch occurs. Therefore the prefetch limit fields of the CPU and DMA may be set differently such that, for example, the DMA allows for a greater number of prefetches between misses as compared to the CPU.
  • the prefetch limit fields can be programmed to take into account various differences between masters and to control the number of prefetches that occur between successive misses, as was described above. That is, after the prefetch limit is reached, no further prefetching occurs until the next buffer miss, since the likelihood that a sequential prefetch will be used decreases as the number of sequential prefetches increases. Also, it can be appreciated how the prefetch limit fields may be used to control the number of lines prefetch in response to each hit or miss or both (prior to the prefetch limit being reached) in order to further reduce wasted prefetches. In the foregoing specification, the invention has been described with reference to specific embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Memory System Of A Hierarchy Structure (AREA)
  • Advance Control (AREA)
PCT/US2004/022438 2003-07-31 2004-07-13 Prefetch control in a data processing system WO2005013039A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP04778111A EP1652092A2 (de) 2003-07-31 2004-07-13 Vorabruf-steuerung in einem datenverarbeitungs-system
JP2006521871A JP4875981B2 (ja) 2003-07-31 2004-07-13 データ処理システムにおけるプリフェッチ制御
KR1020067002041A KR101093317B1 (ko) 2003-07-31 2004-07-13 데이터 처리 시스템 내의 프리패치 제어

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/631,136 2003-07-31
US10/631,136 US7200719B2 (en) 2003-07-31 2003-07-31 Prefetch control in a data processing system

Publications (2)

Publication Number Publication Date
WO2005013039A2 true WO2005013039A2 (en) 2005-02-10
WO2005013039A3 WO2005013039A3 (en) 2005-07-21

Family

ID=34115765

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2004/022438 WO2005013039A2 (en) 2003-07-31 2004-07-13 Prefetch control in a data processing system

Country Status (7)

Country Link
US (1) US7200719B2 (de)
EP (1) EP1652092A2 (de)
JP (1) JP4875981B2 (de)
KR (1) KR101093317B1 (de)
CN (1) CN100407165C (de)
TW (1) TWI352293B (de)
WO (1) WO2005013039A2 (de)

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7526604B1 (en) * 2004-08-09 2009-04-28 Nvidia Corporation Command queueing speculative write prefetch
JP2006251923A (ja) * 2005-03-08 2006-09-21 Oki Electric Ind Co Ltd 先読み制御方法
US7346741B1 (en) * 2005-05-10 2008-03-18 Sun Microsystems, Inc. Memory latency of processors with configurable stride based pre-fetching technique
CN101261610B (zh) * 2007-03-06 2014-04-02 西北农林科技大学 多主设备无冲突访问从设备的方法及装置
JP4829191B2 (ja) * 2007-08-30 2011-12-07 株式会社東芝 キャッシュシステム
US9274965B2 (en) * 2008-12-15 2016-03-01 International Business Machines Corporation Prefetching data
US8473689B2 (en) * 2010-07-27 2013-06-25 Texas Instruments Incorporated Predictive sequential prefetching for data caching
US8583894B2 (en) * 2010-09-09 2013-11-12 Advanced Micro Devices Hybrid prefetch method and apparatus
CN102156633A (zh) * 2011-04-18 2011-08-17 北京北大众志微系统科技有限责任公司 预执行指导的数据预取方法及系统
WO2013030628A1 (en) * 2011-09-01 2013-03-07 Freescale Semiconductor, Inc. Integrated circuit device, memory interface module, data processing system and method for providing data access control
US9645934B2 (en) * 2013-09-13 2017-05-09 Samsung Electronics Co., Ltd. System-on-chip and address translation method thereof using a translation lookaside buffer and a prefetch buffer
US11010092B2 (en) 2018-05-09 2021-05-18 Micron Technology, Inc. Prefetch signaling in memory system or sub-system
US10754578B2 (en) 2018-05-09 2020-08-25 Micron Technology, Inc. Memory buffer management and bypass
US10714159B2 (en) 2018-05-09 2020-07-14 Micron Technology, Inc. Indication in memory system or sub-system of latency associated with performing an access command
US10942854B2 (en) 2018-05-09 2021-03-09 Micron Technology, Inc. Prefetch management for memory
US11294808B2 (en) 2020-05-21 2022-04-05 Micron Technology, Inc. Adaptive cache
US11409657B2 (en) 2020-07-14 2022-08-09 Micron Technology, Inc. Adaptive address tracking
US11422934B2 (en) 2020-07-14 2022-08-23 Micron Technology, Inc. Adaptive address tracking
US20220091847A1 (en) * 2020-09-23 2022-03-24 Advanced Micro Devices, Inc. Prefetching from indirect buffers at a processing unit

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040205299A1 (en) * 2003-04-14 2004-10-14 Bearden Brian S. Method of triggering read cache pre-fetch to increase host read throughput
US20040221111A1 (en) * 2003-04-30 2004-11-04 Sun Microsystems, Inc. Computer system including a memory controller configured to perform pre-fetch operations

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4709324A (en) * 1985-11-27 1987-11-24 Motorola, Inc. Data processor control unit having an interrupt service using instruction prefetch redirection
US5146578A (en) 1989-05-01 1992-09-08 Zenith Data Systems Corporation Method of varying the amount of data prefetched to a cache memory in dependence on the history of data requests
JPH0754482B2 (ja) * 1990-08-07 1995-06-07 株式会社日立製作所 計算機システム
US5410653A (en) * 1992-06-16 1995-04-25 International Business Machines Corporation Asynchronous read-ahead disk caching using multiple disk I/O processes and dynamically variable prefetch length
US5619663A (en) 1994-09-16 1997-04-08 Philips Electronics North America Corp. Computer instruction prefetch system
JP2720838B2 (ja) * 1995-06-19 1998-03-04 日本電気株式会社 データ転送装置
US6085291A (en) 1995-11-06 2000-07-04 International Business Machines Corporation System and method for selectively controlling fetching and prefetching of data to a processor
US5802569A (en) * 1996-04-22 1998-09-01 International Business Machines Corp. Computer system having cache prefetching amount based on CPU request types
US6901500B1 (en) * 2000-07-28 2005-05-31 Silicon Graphics, Inc. Method and apparatus for prefetching information and storing the information in a stream buffer
US6578130B2 (en) 2001-10-18 2003-06-10 International Business Machines Corporation Programmable data prefetch pacing
US6832296B2 (en) * 2002-04-09 2004-12-14 Ip-First, Llc Microprocessor with repeat prefetch instruction

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040205299A1 (en) * 2003-04-14 2004-10-14 Bearden Brian S. Method of triggering read cache pre-fetch to increase host read throughput
US20040221111A1 (en) * 2003-04-30 2004-11-04 Sun Microsystems, Inc. Computer system including a memory controller configured to perform pre-fetch operations

Also Published As

Publication number Publication date
US7200719B2 (en) 2007-04-03
TW200519604A (en) 2005-06-16
WO2005013039A3 (en) 2005-07-21
JP2007500897A (ja) 2007-01-18
US20060053256A1 (en) 2006-03-09
KR20060052936A (ko) 2006-05-19
EP1652092A2 (de) 2006-05-03
JP4875981B2 (ja) 2012-02-15
TWI352293B (en) 2011-11-11
CN1813246A (zh) 2006-08-02
KR101093317B1 (ko) 2011-12-14
CN100407165C (zh) 2008-07-30

Similar Documents

Publication Publication Date Title
US7200719B2 (en) Prefetch control in a data processing system
US6832280B2 (en) Data processing system having an adaptive priority controller
US8725987B2 (en) Cache memory system including selectively accessible pre-fetch memory for pre-fetch of variable size data
JP4064924B2 (ja) メモリアクセスを制御する方法及び装置
US5822758A (en) Method and system for high performance dynamic and user programmable cache arbitration
JP2002140289A (ja) 調整可能ワード・サイズ転送とアドレス配列/増加を備えたマイクロコントローラdmaオペレーション
US7444435B2 (en) Non-fenced list DMA command mechanism
US6871246B2 (en) Prefetch control in a data processing system
US6411218B1 (en) Priority-encoding device selection using variable arbitrary rankings
JP5058116B2 (ja) ストリーミングidメソッドによるdmac発行メカニズム
US10120819B2 (en) System and method for cache memory line fill using interrupt indication
JP4431492B2 (ja) 複数のコヒーレンシー・グラニュールをサポートするデータ転送ユニット
JPH1083343A (ja) メモリをアクセスする方法
JPH0644069A (ja) 周辺装置制御用マイクロプロセッサ

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2004778111

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 20048183628

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 2006521871

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 1020067002041

Country of ref document: KR

WWP Wipo information: published in national office

Ref document number: 2004778111

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1020067002041

Country of ref document: KR