US7552277B2 - Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof - Google Patents

Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof Download PDF

Info

Publication number
US7552277B2
US7552277B2 US10/644,210 US64421003A US7552277B2 US 7552277 B2 US7552277 B2 US 7552277B2 US 64421003 A US64421003 A US 64421003A US 7552277 B2 US7552277 B2 US 7552277B2
Authority
US
United States
Prior art keywords
data
cache
cam
buffer
tag
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US10/644,210
Other versions
US20050044317A1 (en
Inventor
Jude A. Rivers
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US10/644,210 priority Critical patent/US7552277B2/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RIVERS, JUDE A.
Publication of US20050044317A1 publication Critical patent/US20050044317A1/en
Application granted granted Critical
Publication of US7552277B2 publication Critical patent/US7552277B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0893Caches characterised by their organisation or structure
    • G06F12/0897Caches characterised by their organisation or structure with two or more cache hierarchy levels
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Definitions

  • the present invention is related to cache memories and more particularly to storing and accessing data in a cache memory for reduced energy consumption.
  • RAMs Random access memories
  • DRAMs Dynamic RAMs
  • SRAMs High performance Static RAMs
  • SRAMs are somewhat less dense (and generally more expensive per bit) than DRAMs, but expend more power in each access to achieve speed, i.e., provide better access times than DRAMs at the cost of higher power.
  • Content addressable memories CAMs, which also are well known in the art, relate memory locations to detectable values (i.e., location content) and have two modes of operation.
  • the CAM In a storage mode of operation the CAM accepts data for particular locations (e.g., reading/writing to CAM locations), similar to loading a RAM or loading data in a register file.
  • CAM storage locations are identified by and selected by what the locations contain.
  • a particular identifying value, typically called a Comparand is provided, and the array is searched for a match by comparing array contents to the Comparand.
  • Caching is known as an effective technique for increasing microprocessor performance.
  • Typical cache memories are organized with data stored in blocks and, data and tag information in a cache line for each cached data block.
  • Each data block is identified by one of n tags, where each tag may be a virtual index into the cache.
  • the tag normally, includes the upper bits of a virtual address in combination with an address space identifier that is unique to a particular process. Locating a block in cache requires searching cache line data for the virtual address, i.e., the tag, which may be located in one and only one cache location. So, unfortunately, caching is also a major contributor to microprocessor system energy consumption.
  • CAMs work well for cache memory applications, especially for finding a particular tag associated with a selected virtual memory address.
  • an n-way associative cache memory does n tag and data checks in CAM in parallel and, provided the selected block is in cache, quickly locates the tag for the selected block and ignores the rest.
  • tags 52 are stored in CAM 54 and associated data 56 is stored in a bank store (BS) 58 , typically SRAM.
  • BS bank store
  • the CAMRAM 50 is an m (4 in this example) bank 60 cache. Each bank 60 is identified by a bank tag 62 . If the incoming tag 52 matches one of the n entries in the CAM 54 , that match 64 selects a corresponding data block in BS 58 , which is made available for access 66 , e.g., as output or for a cached store. Otherwise, a miss 68 is returned and the incoming request is directed to data located elsewhere, e.g., in main memory.
  • Standard cache memories store data and tag information in the RAM of a cache line.
  • the hardware finds the data based on the virtual address, reads the data and checks the tag against the value stored in the line.
  • the tag for a virtually indexed cache includes the upper bits of the virtual address and an address space identifier, which is unique to a process.
  • An n-way associative cache memory does n tag and data checks in parallel, throwing out the value of all but one of them. While associativity is good and lowers cache miss rates while improving microprocessor performance, the redundant work it requires has a high energy cost. Direct-mapped caches, with associativity of 1, only read one tag and one data word/block and have lower hit energy.
  • CAMRAM caching facilitates higher associativity and can reduce power consumption because of its sequential tag and data access.
  • the search tag of the incoming address is broadcast to the tags depository i.e., the CAM.
  • a matching tag locates the blocks in cache RAM that is requested for access, i.e., requested for a read operation or cached for storage in a store operation.
  • CAM-tag caches are often subbanked with a multi phased access.
  • the CAM-tag compare is the first access phase, where each CAM cell compares its stored value in place with an arriving address. If there is a match in the first phase, the actual data read or write to cache occurs in the next phase.
  • CAM-tag caches still use a significant amount of power finding the associated data in the first phase because the arriving address is broadcast to all of the CAM bank locations. Typically, more than half of CAMRAM cache power is consumed in the CAM-tag checking phase. Consequently, CAMRAM power is directly related to the number of bank entries, i.e., the larger the bank, the more power required. For an energy-efficient cache design, therefore, the designer must find the proper mix of associativity, size, structure configuration, and partitioning to achieve an acceptable energy consumption level. Achieving such a mix without proper regards to the inherent code and data behavior of targeted workloads has been difficult.
  • the present invention relates to a cache memory, content addressable memory, random access memory (CAMRAM) cache and method of managing a cache.
  • a cache buffer receives incoming data and buffers a storage array.
  • the cache buffer holds a number of most recently accessed data blocks. In any access, cache buffer locations are checked before checking the storage array.
  • FIG. 1 shows an example of a state of the art CAMRAM
  • FIGS. 2A-B show data flow examples of cache accesses according to a preferred embodiment of the present invention
  • FIG. 3 shows an example of a preferred embodiment single bank cache memory according to the present invention
  • FIG. 4 shows an example of a multibank CAMRAM cache 130 embodiment
  • FIG. 5 shows a flow diagram for a load access (read) in a preferred multibank CAMRAM cache
  • FIG. 6 shows a flow diagram for a store access (write) to a preferred multibank CAMRAM cache.
  • FIGS. 2A-B show data flow examples 80 A-B of cache accesses according to a preferred embodiment of the present invention with like numbered elements labeled identically.
  • a line in 82 e.g., from main memory or level 2(L2) cache (not shown) is first brought into a staging buffer 84 .
  • a CPU data request is first presented in parallel to the staging buffer 84 and the hollow buffer 88 (which holds a number (i) of the most recently accessed lines from the holding cache); and, when the item is not found the request is then presented to the holding cache 86 .
  • FIG. 2B is substantially similar to that of FIG.
  • the buffers 84 , 88 hold the most active cache lines and buffer the holding cache 86 , which is dormant unless a data request is directed to data that is not found in the buffers 84 , 88 . So, for any cache access, first the cache buffers, staging buffer 84 and hollow buffer 88 , are checked. Only if the target data are not found in this first check, are the remaining holding cache 86 locations checked. Accordingly, since analysis has shown that more than 85% of all data accesses are found in this initial search, limiting the initial search to the cache buffer 84 , 88 substantially reduces cache power consumption.
  • CAM content addressable memory
  • RAM random access memory
  • CAMRAM cache CAM random access memory
  • present invention has application to any suitable memory architecture wherein at any one time, a few locations may be accessed more frequently than the remaining locations.
  • FIG. 3 shows a CAMRAM example of a preferred embodiment cache memory 100 according to the present invention, a single bank cache memory in this example.
  • the cache memory is a CAMRAM cache with a two phase access for improved (lower) power consumption.
  • data storage includes an n block deep holding cache (HC) 106 or bank store and a shallow or hollow output buffer, i.e., an i block deep hollow buffer (HB) 108 .
  • HC n block deep holding cache
  • HB i block deep hollow buffer
  • a block may be a single word or multiple words and, further, includes any portion of a block or subblock.
  • the n tags in the n-CAM 102 are each associated with corresponding blocks in HC 106 and i-CAM tags being associated with corresponding blocks in HB 108 .
  • the HB 108 is a fast and low power memory array, e.g., an i-stage register file, that is accessible externally, i.e., both readable and writeable.
  • Incoming cache lines are stored as lines in a cache staging buffer (CSB) 110 which serves as an input staging buffer for external data being passed, e.g., to/from a second level (L2) cache and data from a cache hit is passed to/from a stored bank buffer (SBB) 112 .
  • CSB cache staging buffer
  • L2 second level
  • SBB stored bank buffer
  • Only the CSB 110 can write to the HC 106 .
  • the CSB 110 can be both read from and written into by the microprocessor, and written into by a higher level memory.
  • CAMRAM cache 100 Power is reduced in CAMRAM cache 100 over prior art CAMRAM caches without significant performance degradation because the buffers contain the most likely target locations and the initial search is directed to those location.
  • the i most recently accessed cache data blocks which have the highest likelihood of being requested in immediately subsequent accesses, are held in the HB 108 . So, in any immediately subsequent access, the incoming tag is compared against i-CAM entries, which are most likely to match. Thus, finding a match in the i-CAM 104 saves power, CAM power that is otherwise expended searching n-CAM 102 .
  • the n-CAM 102 is a circular first in first out (FIFO), such that as a new tag is loaded, the most stale tag (i.e., with the longest time since last use) is shifted out or unloaded and a corresponding block is released in HC 106 .
  • the i-CAM 104 is organized by Least Recently Used (LRU) and, preferably, is also a circular FIFO with entries mapping one-to-one with HB 108 entries.
  • the HB 108 is a fully associative structure, associating HB 108 entries with blocks in HC 106 .
  • the CSB 110 is a staging storage buffer for one or more cache lines from a higher level in the memory hierarchy (not shown), e.g., from a level 2 cache or main memory. Each incoming cache line 114 is loaded into the CSB 110 and is only promoted into the HC 106 following a cache miss 116 in the level one structure that necessitates uploading the new block into the CSB 110 . Once a cache line is promoted from the CSB 110 to the HC 106 , any subsequent access to that cache line, a copy of that cache line is promoted from the HC 106 to the HB 108 and, coincidentally, a copy of the corresponding tag is passed from the n-CAM 102 to the i-CAM 104 .
  • a copy of a cache line being moved into the HB 108 from the HC 106 may also remain in the HC 106 . However, if a copy is left in the HC 106 , then, every time a cache line is written into in the HB 108 , the HC 106 must be searched to find and invalidate the corresponding copy.
  • a 2 stage access of the CAMRAM cache 100 ensures power consumed is minimized.
  • each stage 102 , 104 can be handled in a single clock cycle and accessing data in cache 100 requires no more than 2 clock cycles.
  • both the i-CAM 104 and the CSB 110 compare a tag against their contents for match and, simultaneously, the HB 108 is prepared (pre-charged) for access.
  • the tag is found, i.e., a hit 118
  • data from the corresponding storage location i.e., in HB 108 or CSB 110
  • SBB 112 data from the corresponding storage location, i.e., in HB 108 or CSB 110
  • the n-CAM 102 searches for the tag and HC 106 is precharged. If the tag is found in the n-CAM 102 , the match line 124 is asserted. Data from the HC 106 is latched in the HB 108 , and passed through the SBB to output 120 .
  • a miss 116 indication is provided indicating that the data is elsewhere, e.g., in L2 cache or main memory.
  • an access hit in the CSB 110 or HB 108 in the first cycle costs one clock cycle, while an access hit in the HC 106 occurs in a second clock cycle and costs 1 additional cycle.
  • FIG. 4 shows an example of a multibank CAMRAM cache 130 embodiment with multiple banks 132 substantially similar to the single bank embodiment 100 of FIG. 3 with like elements labeled identically.
  • the CSB includes a CSB line 134 in each bank 132 .
  • a single SBB 136 serves all banks 132 , receiving individual outputs 138 , 140 from each bank HC 106 and HB 108 .
  • An incoming tag 142 is directed by a bank identifier 144 to a particular bank 132 , with associated data at an HB 108 or provided from HC 106 to HB 108 as described above for the single bank embodiment 100 . It has been shown that, frequently, consecutive cache accesses map to the same cache bank 132 and also end up in the same cache line; exhibiting an inherent spatial locality in memory access behavior that affords significant power savings in a preferred embodiment CAMRAM cache 100 or 130 .
  • level one cache misses cause the cache line to reload into CSB line 134 .
  • Cache line data accumulates in CSB lines 134 , over multiple cycles.
  • Cache line data are passed into the HC 106 during a dead cycle in a subsequent L1 cache miss, i.e., phase 1 and phase 2.
  • Back to back accesses to incoming cache lines are satisfied from CSB lines 132 .
  • If a cache miss is caused by a store instruction the data is written into the CSB line 132 .
  • An n-CAM hit causes a copy of respective cache line to move from the HC 106 into the HB 108 and the corresponding HB 108 entry is invalidated. If the HB 108 is full, the HB 108 LRU entry is castout through the SBB 136 .
  • FIG. 5 shows a flow diagram 150 for a load access (read) in a preferred multibank CAMRAM cache, with reference to the CAMRAM cache 130 example in FIG. 4 .
  • the CAMRAM cache 130 receives a load request directed to a memory location that may be in cache 130 .
  • the bank identifier 144 selects a bank 132 .
  • the CSB line 134 and the i-CAM 104 are searched for the tag 142 .
  • the HB 108 is precharged.
  • step 164 the second phase begins searching the n-CAM 102 for the tag. If the tag is found in the n-CAM 102 in step 166 , then in step 168 , the associated data is checked to determine whether it is a block or sub-block.
  • step 170 the sub-block is passed to the HB 108 and a sub-block (Sb) validity bit (not shown) is set.
  • the validity bit is not set for other sub-blocks mapping to that cache line. So, accesses to any of the other those sub-blocks generates further misses, which causes the corresponding data to be loaded into the frame of the already allocated line in the HB and setting the respective validity bit. Otherwise, the full block is being loaded, obviating the need to set validity bits and in step 170 f the block is passed to the HB 108 .
  • the victim block may have been previously modified and, if so, it is passed to the SBB 136 .
  • step 172 a miss indication 116 is returned and, if the CSB line 134 is occupied, its contents are passed to HC 106 .
  • step 174 when the requested block is returned, e.g., from the L2cache, the block is stored in the CSB line 134 and, returning to step 152 , the block is passed through the CAMRAM cache 132 in a next first pass, following first pass steps 154 - 160 .
  • FIG. 6 shows a flow diagram 180 for a store access (write) to a preferred multibank CAMRAM cache, again with reference to the CAMRAM cache 130 example in FIG. 4 .
  • the CAMRAM cache 130 receives a store request directed to a memory location that may be in cache 130 .
  • the bank identifier 144 selects a bank 132 .
  • the CSB line 134 and the i-CAM 104 are searched for the tag 142 .
  • the HB 108 is precharged.
  • step 190 the data is stored in the CSB line 134 , which is marked as dirty to end the first phase and the search. Otherwise, if the tag is found in the i-cache 104 in step 192 , then in step 194 , the data is stored in the associated HB 108 , which also is marked as dirty ending the first phase and the search. However, if the tag is not found in either the CSB line 134 in step 188 or the i-cache 104 in step 192 , the tag was not found in the first phase. Instead, in step 196 the second phase begins searching the n-CAM 102 for the tag.
  • step 200 the associated data is checked whether it is a block or sub-block. If it is a sub-block, then in step 202 s the sub-block is passed to the HB 108 and a sub-block (Sb) validity bit (not shown) is set. Otherwise in step 202 f the block is passed to the HB 108 .
  • the victim block may have been previously modified and, if so, it is passed to the SBB 136 .
  • step 204 a miss indication 116 is returned and, if the CSB line 134 is occupied, its contents are passed to HC 106 .
  • step 206 when the requested block is returned, e.g., from the L2 cache, the block is stored in the CSB line 134 and, returning to step 182 , the block is passed through the CAMRAM cache 130 in a next first pass, following first pass steps 184 - 194 .
  • the present invention leverages the inherent spatial and temporal locality behavior patterns of program code and data elements in applications, in particular in minimizing cache power consumption and, correspondingly, overall system power consumption. Further, the present invention distributes cached data into data cache structures so as to take advantage of the high associativity of modular code in combination with a larger holding/holding-victim cache, while incurring less performance and power consumption penalties.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Memory System Of A Hierarchy Structure (AREA)

Abstract

A cache memory that may include a content addressable memory, random access memory (CAMRAM) cache and method for managing a cache to reduce cache energy consumption. A cache buffer receives incoming data and buffers a storage array. The cache buffer holds a number of most recently accessed data blocks. In any access, cache buffer locations are checked before checking the storage array.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention is related to cache memories and more particularly to storing and accessing data in a cache memory for reduced energy consumption.
2. Background Description
Random access memories (RAMs) are well known in the art. A typical RAM has a memory array wherein every location is addressable and freely accessible by providing the correct corresponding address. Dynamic RAMs (DRAMs) are dense RAMs with a very small memory cell. High performance Static RAMs (SRAMs) are somewhat less dense (and generally more expensive per bit) than DRAMs, but expend more power in each access to achieve speed, i.e., provide better access times than DRAMs at the cost of higher power. Content addressable memories (CAMs), which also are well known in the art, relate memory locations to detectable values (i.e., location content) and have two modes of operation. In a storage mode of operation the CAM accepts data for particular locations (e.g., reading/writing to CAM locations), similar to loading a RAM or loading data in a register file. In a second content addressable or search mode, CAM storage locations are identified by and selected by what the locations contain. A particular identifying value, typically called a Comparand is provided, and the array is searched for a match by comparing array contents to the Comparand.
In a typical data processing system, the bulk of the memory is DRAM in main memory with faster SRAM in cache memory, closer to the processor or microprocessor. Caching is known as an effective technique for increasing microprocessor performance. Typical cache memories are organized with data stored in blocks and, data and tag information in a cache line for each cached data block. Each data block is identified by one of n tags, where each tag may be a virtual index into the cache. The tag, normally, includes the upper bits of a virtual address in combination with an address space identifier that is unique to a particular process. Locating a block in cache requires searching cache line data for the virtual address, i.e., the tag, which may be located in one and only one cache location. So, unfortunately, caching is also a major contributor to microprocessor system energy consumption.
Consequently, because finding a virtual address in RAM requires checking cache lines sequentially, until the virtual address is located; CAMs work well for cache memory applications, especially for finding a particular tag associated with a selected virtual memory address. In particular, an n-way associative cache memory does n tag and data checks in CAM in parallel and, provided the selected block is in cache, quickly locates the tag for the selected block and ignores the rest.
Accordingly as illustrated in FIG. 1, in what is known as a CAMRAM cache 50, tags 52 are stored in CAM 54 and associated data 56 is stored in a bank store (BS) 58, typically SRAM. In this example the CAMRAM 50 is an m (4 in this example) bank 60 cache. Each bank 60 is identified by a bank tag 62. If the incoming tag 52 matches one of the n entries in the CAM 54, that match 64 selects a corresponding data block in BS 58, which is made available for access 66, e.g., as output or for a cached store. Otherwise, a miss 68 is returned and the incoming request is directed to data located elsewhere, e.g., in main memory.
Standard cache memories store data and tag information in the RAM of a cache line. The hardware finds the data based on the virtual address, reads the data and checks the tag against the value stored in the line. The tag for a virtually indexed cache includes the upper bits of the virtual address and an address space identifier, which is unique to a process. An n-way associative cache memory does n tag and data checks in parallel, throwing out the value of all but one of them. While associativity is good and lowers cache miss rates while improving microprocessor performance, the redundant work it requires has a high energy cost. Direct-mapped caches, with associativity of 1, only read one tag and one data word/block and have lower hit energy. However, they have much larger miss rates due to conflicts and since the energy cost per miss is higher, they tend to have larger total memory access energy. Techniques like way-predicting caches can provide associativity at lower hit energy by only checking one way in an n-way set associative cache, but tend to incur energy and delay penalties to access the way-prediction table on way hits and additional energy and performance penalties if predictions are incorrect. Caches are also often split into subbanks, which handle certain address ranges. Bank addresses are direct mapped using the appropriate virtual address bits.
CAMRAM caching facilitates higher associativity and can reduce power consumption because of its sequential tag and data access. During a CAMRAM access, the search tag of the incoming address is broadcast to the tags depository i.e., the CAM. A matching tag (if any) locates the blocks in cache RAM that is requested for access, i.e., requested for a read operation or cached for storage in a store operation. M. Zhang and K Asanovich, “Highly-Associative Caches for Low-Power Processors,” Kool Chips Workshop, 33rd Int'l Symposium on Microarchitecture, (2000) describes how a 32-way CAM-tag search uses abut the same power as a 2-way set associative RAM-tag search. For additional power reduction, CAM-tag caches are often subbanked with a multi phased access. Typically, the CAM-tag compare is the first access phase, where each CAM cell compares its stored value in place with an arriving address. If there is a match in the first phase, the actual data read or write to cache occurs in the next phase.
Unfortunately, CAM-tag caches still use a significant amount of power finding the associated data in the first phase because the arriving address is broadcast to all of the CAM bank locations. Typically, more than half of CAMRAM cache power is consumed in the CAM-tag checking phase. Consequently, CAMRAM power is directly related to the number of bank entries, i.e., the larger the bank, the more power required. For an energy-efficient cache design, therefore, the designer must find the proper mix of associativity, size, structure configuration, and partitioning to achieve an acceptable energy consumption level. Achieving such a mix without proper regards to the inherent code and data behavior of targeted workloads has been difficult.
Thus, there is a need to reduce the number of tag checks per access and further, to reduce cache memory power consumption.
SUMMARY OF THE INVENTION
It is a purpose of the invention to reduce cache power consumption;
It is another purpose of the invention to reduce cache power consumption while maintaining cache performance;
It is yet another purpose of the invention to quickly determine whether memory being accessed is in cache while reducing cache power consumption.
The present invention relates to a cache memory, content addressable memory, random access memory (CAMRAM) cache and method of managing a cache. A cache buffer receives incoming data and buffers a storage array. The cache buffer holds a number of most recently accessed data blocks. In any access, cache buffer locations are checked before checking the storage array.
BRIEF DESCRIPTION OF THE DRAWINGS
The foregoing and other objects, aspects and advantages will be better understood from the following detailed description of a preferred embodiment of the invention with reference to the drawings, in which:
FIG. 1 shows an example of a state of the art CAMRAM;
FIGS. 2A-B show data flow examples of cache accesses according to a preferred embodiment of the present invention;
FIG. 3 shows an example of a preferred embodiment single bank cache memory according to the present invention;
FIG. 4 shows an example of a multibank CAMRAM cache 130 embodiment;
FIG. 5 shows a flow diagram for a load access (read) in a preferred multibank CAMRAM cache;
FIG. 6 shows a flow diagram for a store access (write) to a preferred multibank CAMRAM cache.
DESCRIPTION OF PREFERRED EMBODIMENTS
Turning now to the drawings and, more particularly, FIGS. 2A-B show data flow examples 80A-B of cache accesses according to a preferred embodiment of the present invention with like numbered elements labeled identically. In FIG. 2A, a line in 82, e.g., from main memory or level 2(L2) cache (not shown), is first brought into a staging buffer 84. A CPU data request is first presented in parallel to the staging buffer 84 and the hollow buffer 88 (which holds a number (i) of the most recently accessed lines from the holding cache); and, when the item is not found the request is then presented to the holding cache 86. When a data request is to data not found in either of the buffers 84, 88, but is in the holding cache 86, that data 90 is brought into the hollow buffer 88. A cache miss occurs when the requested data is not found in cache buffers 84, 88 or the holding cache 86. Upon a cache miss, the data from the staging buffer 84 is promoted into the holding cache 86 and the incoming missed data block 82 is brought into the staging buffer 84. A cache line is replaced 92 from the cache system from either the holding cache 86 or the hollow buffer 88. The example of FIG. 2B is substantially similar to that of FIG. 2A, except that data that must be replaced 92 from the hollow buffer is instead placed back 96 into the holding cache 86 making it the holding-victim cache 94. In example 2B, therefore, a cache line is replaced 92 from the cache system only from the holding-victim cache 94.
Thus, the buffers 84, 88 hold the most active cache lines and buffer the holding cache 86, which is dormant unless a data request is directed to data that is not found in the buffers 84, 88. So, for any cache access, first the cache buffers, staging buffer 84 and hollow buffer 88, are checked. Only if the target data are not found in this first check, are the remaining holding cache 86 locations checked. Accordingly, since analysis has shown that more than 85% of all data accesses are found in this initial search, limiting the initial search to the cache buffer 84, 88 substantially reduces cache power consumption. It should be noted that although described herein below with reference to content addressable memory (CAM) random access memory (RAM) or CAMRAM cache, this is for example only and not intended as a limitation. The present invention has application to any suitable memory architecture wherein at any one time, a few locations may be accessed more frequently than the remaining locations.
FIG. 3 shows a CAMRAM example of a preferred embodiment cache memory 100 according to the present invention, a single bank cache memory in this example. In this example, the cache memory is a CAMRAM cache with a two phase access for improved (lower) power consumption. Tags are stored in an n-location CAM or n-CAM 102 and an i-location CAM or i-CAM 104, where n>i and although, i may have any value, preferably, i=8. Similarly, data storage includes an n block deep holding cache (HC) 106 or bank store and a shallow or hollow output buffer, i.e., an i block deep hollow buffer (HB) 108. It should be noted that a block may be a single word or multiple words and, further, includes any portion of a block or subblock. The n tags in the n-CAM 102 are each associated with corresponding blocks in HC 106 and i-CAM tags being associated with corresponding blocks in HB 108. Preferably, the HB 108 is a fast and low power memory array, e.g., an i-stage register file, that is accessible externally, i.e., both readable and writeable. Incoming cache lines are stored as lines in a cache staging buffer (CSB) 110 which serves as an input staging buffer for external data being passed, e.g., to/from a second level (L2) cache and data from a cache hit is passed to/from a stored bank buffer (SBB) 112. Only the CSB 110 can write to the HC 106. However, the CSB 110 can be both read from and written into by the microprocessor, and written into by a higher level memory.
Power is reduced in CAMRAM cache 100 over prior art CAMRAM caches without significant performance degradation because the buffers contain the most likely target locations and the initial search is directed to those location. In particular, the i most recently accessed cache data blocks, which have the highest likelihood of being requested in immediately subsequent accesses, are held in the HB 108. So, in any immediately subsequent access, the incoming tag is compared against i-CAM entries, which are most likely to match. Thus, finding a match in the i-CAM 104 saves power, CAM power that is otherwise expended searching n-CAM 102.
Preferably, the n-CAM 102 is a circular first in first out (FIFO), such that as a new tag is loaded, the most stale tag (i.e., with the longest time since last use) is shifted out or unloaded and a corresponding block is released in HC 106. The i-CAM 104 is organized by Least Recently Used (LRU) and, preferably, is also a circular FIFO with entries mapping one-to-one with HB 108 entries. Preferably, the HB 108 is a fully associative structure, associating HB 108 entries with blocks in HC 106. The CSB 110 is a staging storage buffer for one or more cache lines from a higher level in the memory hierarchy (not shown), e.g., from a level 2 cache or main memory. Each incoming cache line 114 is loaded into the CSB 110 and is only promoted into the HC 106 following a cache miss 116 in the level one structure that necessitates uploading the new block into the CSB 110. Once a cache line is promoted from the CSB 110 to the HC 106, any subsequent access to that cache line, a copy of that cache line is promoted from the HC 106 to the HB 108 and, coincidentally, a copy of the corresponding tag is passed from the n-CAM 102 to the i-CAM 104. A copy of a cache line being moved into the HB 108 from the HC 106 may also remain in the HC 106. However, if a copy is left in the HC 106, then, every time a cache line is written into in the HB 108, the HC 106 must be searched to find and invalidate the corresponding copy.
Accordingly, a 2 stage access of the CAMRAM cache 100 (an i-CAM 104 search followed by an n-CAM 102 search) ensures power consumed is minimized. Typically, each stage 102, 104 can be handled in a single clock cycle and accessing data in cache 100 requires no more than 2 clock cycles. In the first clock cycle or phase, both the i-CAM 104 and the CSB 110 compare a tag against their contents for match and, simultaneously, the HB 108 is prepared (pre-charged) for access. If the tag is found, i.e., a hit 118, data from the corresponding storage location, i.e., in HB 108 or CSB 110, is latched in SBB 112 and provided as CAMRAM cache output 118. Otherwise, if the tag is not found in either of the i-CAM 104 and the CSB 110, i.e., a miss 122, then, the n-CAM 102 searches for the tag and HC 106 is precharged. If the tag is found in the n-CAM 102, the match line 124 is asserted. Data from the HC 106 is latched in the HB 108, and passed through the SBB to output 120. Otherwise, a miss 116 indication is provided indicating that the data is elsewhere, e.g., in L2 cache or main memory. Thus, an access hit in the CSB 110 or HB 108 in the first cycle costs one clock cycle, while an access hit in the HC 106 occurs in a second clock cycle and costs 1 additional cycle.
FIG. 4 shows an example of a multibank CAMRAM cache 130 embodiment with multiple banks 132 substantially similar to the single bank embodiment 100 of FIG. 3 with like elements labeled identically. In this embodiment, the CSB includes a CSB line 134 in each bank 132. Also, a single SBB 136 serves all banks 132, receiving individual outputs 138, 140 from each bank HC 106 and HB 108. An incoming tag 142 is directed by a bank identifier 144 to a particular bank 132, with associated data at an HB 108 or provided from HC 106 to HB 108 as described above for the single bank embodiment 100. It has been shown that, frequently, consecutive cache accesses map to the same cache bank 132 and also end up in the same cache line; exhibiting an inherent spatial locality in memory access behavior that affords significant power savings in a preferred embodiment CAMRAM cache 100 or 130.
In summary, level one cache misses cause the cache line to reload into CSB line 134. Cache line data accumulates in CSB lines 134, over multiple cycles. Cache line data are passed into the HC 106 during a dead cycle in a subsequent L1 cache miss, i.e., phase 1 and phase 2. Back to back accesses to incoming cache lines are satisfied from CSB lines 132. If a cache miss is caused by a store instruction, the data is written into the CSB line 132. An n-CAM hit causes a copy of respective cache line to move from the HC 106 into the HB 108 and the corresponding HB 108 entry is invalidated. If the HB 108 is full, the HB 108 LRU entry is castout through the SBB 136.
FIG. 5 shows a flow diagram 150 for a load access (read) in a preferred multibank CAMRAM cache, with reference to the CAMRAM cache 130 example in FIG. 4. First, in step 152 the CAMRAM cache 130 receives a load request directed to a memory location that may be in cache 130. In step 154 the bank identifier 144 selects a bank 132. In step 156, using the target tag 142, the CSB line 134 and the i-CAM 104 are searched for the tag 142. Coincidentally, the HB 108 is precharged. If the tag is found in the CSB line 134 in step 158 or in the i-CAM 104 in step 160, then the first phase and the search completes in step 162 when the result is returned to the particular register file/ functional unit (RF/FU). Otherwise, if the tag is not found in the CSB line 134 in step 158 or in the i-CAM 104 in step 160, the tag was not found in the first phase. Instead, in step 164 the second phase begins searching the n-CAM 102 for the tag. If the tag is found in the n-CAM 102 in step 166, then in step 168, the associated data is checked to determine whether it is a block or sub-block. If it is a sub-block, then in step 170 s the sub-block is passed to the HB 108 and a sub-block (Sb) validity bit (not shown) is set. The validity bit is not set for other sub-blocks mapping to that cache line. So, accesses to any of the other those sub-blocks generates further misses, which causes the corresponding data to be loaded into the frame of the already allocated line in the HB and setting the respective validity bit. Otherwise, the full block is being loaded, obviating the need to set validity bits and in step 170 f the block is passed to the HB 108. The victim block may have been previously modified and, if so, it is passed to the SBB 136. If, however, the tag is not found in the n-CAM 102 in step 166, i.e., in the second phase; then, in step 172 a miss indication 116 is returned and, if the CSB line 134 is occupied, its contents are passed to HC 106. Finally, in step 174 when the requested block is returned, e.g., from the L2cache, the block is stored in the CSB line 134 and, returning to step 152, the block is passed through the CAMRAM cache 132 in a next first pass, following first pass steps 154-160.
FIG. 6 shows a flow diagram 180 for a store access (write) to a preferred multibank CAMRAM cache, again with reference to the CAMRAM cache 130 example in FIG. 4. First, in step 182 the CAMRAM cache 130 receives a store request directed to a memory location that may be in cache 130. In step 184 the bank identifier 144 selects a bank 132. In step 186, using the target tag 142, the CSB line 134 and the i-CAM 104 are searched for the tag 142. Coincidentally, the HB 108 is precharged. If the tag is found in the CSB line 134 in step 188, then, in step 190 the data is stored in the CSB line 134, which is marked as dirty to end the first phase and the search. Otherwise, if the tag is found in the i-cache 104 in step 192, then in step 194, the data is stored in the associated HB 108, which also is marked as dirty ending the first phase and the search. However, if the tag is not found in either the CSB line 134 in step 188 or the i-cache 104 in step 192, the tag was not found in the first phase. Instead, in step 196 the second phase begins searching the n-CAM 102 for the tag. If the tag is found in the n-CAM 102 in step 198, then in step 200, the associated data is checked whether it is a block or sub-block. If it is a sub-block, then in step 202 s the sub-block is passed to the HB 108 and a sub-block (Sb) validity bit (not shown) is set. Otherwise in step 202 f the block is passed to the HB 108. The victim block may have been previously modified and, if so, it is passed to the SBB 136. However, if the tag is not found in the n-CAM 102 in step 198, i.e., in the second phase; then in step 204 a miss indication 116 is returned and, if the CSB line 134 is occupied, its contents are passed to HC 106. Finally, in step 206 when the requested block is returned, e.g., from the L2 cache, the block is stored in the CSB line 134 and, returning to step 182, the block is passed through the CAMRAM cache 130 in a next first pass, following first pass steps 184-194.
Accordingly, the present invention leverages the inherent spatial and temporal locality behavior patterns of program code and data elements in applications, in particular in minimizing cache power consumption and, correspondingly, overall system power consumption. Further, the present invention distributes cached data into data cache structures so as to take advantage of the high associativity of modular code in combination with a larger holding/holding-victim cache, while incurring less performance and power consumption penalties.
While the invention has been described in terms of preferred embodiments, those skilled in the art will recognize that the invention can be practiced with modification within the spirit and scope of the appended claims.

Claims (23)

1. A cache memory comprising:
a cache buffer containing most recently accessed data;
a storage array comprising a plurality of cache memory locations and selectively receiving data from said cache buffer, selectively received said data being stored in ones of said memory locations; and
a tag memory storing tags associated with data in said storage array and selected data in said cache buffer, wherein cache power is substantially less for accessing said data in said cache buffer than for accessing data in said storage array.
2. A cache memory as in claim 1, wherein cache input data selectively includes executable commands.
3. A cache memory as in claim 1, wherein said cache buffer comprises:
a cache input buffer receiving cache input data.
4. A cache memory as in claim 3, wherein said cache buffer further comprises:
an output buffer, ones of said tags in said tag memory associated with said most recently accessed data in said output buffer.
5. A cache memory as in claim 4, wherein said tag memory comprises:
a first content addressable memory (CAM) containing tags associated with data stored in said storage array; and
a second CAM containing tags associated with said most recently accessed data.
6. A cache memory as in claim 5, wherein a tag for requested data is checked against tags in said second CAM and said cache input buffer before checking tags in said first CAM.
7. A cache memory as in claim 5, wherein each of said first CAM and said second CAM are a circulating first in first out register (FIFO).
8. A cache memory as in claim 4, wherein each said storage array is a static random access (SRAM) array.
9. A content addressable memory (CAM) random access memory (RAM) cache comprising a plurality of CAMRAM banks, each of said CAMRAM banks comprising:
a cache buffer containing most recently accessed data and receiving cache input data, said cache input data selectively including executable commands, wherein said cache buffer comprises:
an input buffer line receiving a cache input data line, and
an output buffer containing said most recently accessed data, ones of said tags in said CAM being associated with said most recently accessed data;
a bank store comprising a plurality of cache memory locations and selectively receiving data from said cache buffer, selectively received said data being stored in ones of said memory locations; and
a CAM storing tags associated with data in said storage array and selected data in said cache buffer, wherein said CAM comprises:
an n-CAM having n tag locations, each n-CAM tag location being associated with one of n storage locations in said bank store, and
an i-CAM containing i tag locations, wherein n>i and each i-CAM tag location is associated with a location in said output buffer, wherein each of said n-CAM and said i-CAM are a circulating first in first out register (FIFO).
10. A CAMRAM as in claim 9, further comprising a cache storage buffer, each said input buffer line in said plurality of CAMRAM banks being a line in said cache storage buffer.
11. A CAMRAM as in claim 9, further comprising means for checking a tag for requested data against tags in said i CAM and said cache input buffer independent of tags in said n CAM.
12. A CAMRAM as in claim 11, wherein said checking means only checks for said tag in said n CAM, when said tag is not found in said i CAM or in said cache input buffer.
13. A CAMRAM as in claim 12, wherein cache power is substantially less for accessing said data in said cache buffer than for accessing data in said bank store.
14. A CAMRAM as in claim 9, wherein said bank store is a static random access (SRAM) array.
15. A method of managing data in a cache, said method comprising the steps of:
a) providing incoming data to an input buffer;
b) selectively loading data from said input buffer into a storage array;
c) selectively loading accessed data from said storage array to an output buffer, a number of most recently accessed data blocks being held in said output buffer;
d) selectively providing data from each of said input buffer, said storage array and said output buffer responsive to an access request;
e) receiving an access request for data;
f) checking said input data buffer for data requested for access; and
g) checking said output buffer for said data requested for access, wherein said output buffer is checked in step (g) coincident with checking said input buffer in step (f).
16. A method of managing data as in claim 15, wherein said access request is a store request and said method further comprises the steps of:
h) storing said data in said output buffer; and
i) marking said stored data as dirty.
17. A method of managing data as in claim 15, wherein whenever said data requested for access is not found in said output buffer or said input buffer, said method further comprises the steps of:
h) checking said storage array for said data requested for access.
18. A method of managing data as in claim 17, wherein whenever said data requested for access is found in said storage array, said method further comprises the steps of:
i) loading said data requested for access into said output buffer; and
j) providing said data requested for access as an output.
19. A method of managing data as in claim 17, wherein whenever said data requested for access is not found in said storage array, said method further comprises the steps of:
i) sending a miss request;
j) loading said input buffer; and
k) providing said data from said input buffer as an output.
20. A method of managing data as in claim 19, wherein whenever said input buffer contains data other than said data requested for access, said sending step (h) further comprises loading other said data from input buffer to said output buffer.
21. A method of managing data as in claim 17, wherein data in each of said input buffer, said storage array and said output buffer are identified by tags, said tags being checked in checking steps (f), (g) and (h).
22. A cache memory comprising:
a cache buffer containing most recently accessed data, wherein said cache buffer further comprises an output buffer, ones of said tags in said tag memory associated with said most recently accessed data in said output buffer;
a storage array comprising a plurality of cache memory locations and selectively receiving data from said cache buffer, selectively received said data being stored in ones of said memory locations; and
a tag memory storing tags associated with data in said storage array and selected data in said cache buffer, wherein said tag memory comprises:
a first content addressable memory (CAM) containing tags associated with data stored in said storage array, and
a second CAM containing tags associated with said most recently accessed data, wherein each of said first CAM and said second CAM are a circulating first in first out register (FIFO).
23. A content addressable memory (CAM) random access memory (RAM) cache comprising a plurality of CAMRAM banks, each of said CAMRAM banks comprising:
a cache buffer containing most recently accessed data and receiving cache input data, said cache input data selectively including executable commands, wherein said cache buffer comprises:
an input buffer line receiving a cache input data line, and
an output buffer containing said most recently accessed data, ones of said tags in said CAM being associated with said most recently accessed data;
a bank store comprising a plurality of cache memory locations and selectively receiving data from said cache buffer, selectively received said data being stored in ones of said memory locations; and
a CAM storing tags associated with data in said storage array and selected data in said cache buffer, wherein said CAM comprises:
an n-CAM having n tag locations, each n-CAM tag location being associated with one of n storage locations in said bank store,
an i-CAM containing i tag locations, wherein n>i and each i-CAM tag location is associated with a location in said output buffer; and
means for checking a tag for requested data against tags in said i CAM and said cache input buffer independent of tags in said n CAM, wherein said checking means only checks for said tag in said n CAM, when said tag is not found in said i CAM or in said cache input buffer and wherein cache power is substantially less for accessing said data in said cache buffer than for accessing data in said bank store.
US10/644,210 2003-08-20 2003-08-20 Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof Expired - Fee Related US7552277B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/644,210 US7552277B2 (en) 2003-08-20 2003-08-20 Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/644,210 US7552277B2 (en) 2003-08-20 2003-08-20 Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof

Publications (2)

Publication Number Publication Date
US20050044317A1 US20050044317A1 (en) 2005-02-24
US7552277B2 true US7552277B2 (en) 2009-06-23

Family

ID=34194030

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/644,210 Expired - Fee Related US7552277B2 (en) 2003-08-20 2003-08-20 Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof

Country Status (1)

Country Link
US (1) US7552277B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013101060A2 (en) * 2011-12-29 2013-07-04 Intel Corporation Efficient support of sparse data structure access

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2859292A1 (en) * 2003-08-27 2005-03-04 St Microelectronics Sa DEVICE FOR CONTROLLING ACCESS TO A SECURE MEMORY, COMPRISING A SYNCHRONOUS CIRCUIT FOR SEARCHING FOR ATTRIBUTES
FR2859307A1 (en) * 2003-08-27 2005-03-04 St Microelectronics Sa ASYNCHRONOUS PLAYBACK MEMORY AND DEVICE FOR CONTROLLING ACCESS TO A DATA MEMORY COMPRISING SUCH CACHE MEMORY
ITMI20031673A1 (en) * 2003-08-28 2005-02-28 Nuovo Pignone Spa FIXING SYSTEM OF A FLAME TUBE OR "LINER".
US20070113031A1 (en) * 2005-11-16 2007-05-17 International Business Machines Corporation Memory management system and method for storing and retrieving messages
US20080104323A1 (en) * 2006-10-26 2008-05-01 Colglazier Daniel J Method for identifying, tracking, and storing hot cache lines in an smp environment
US9176856B2 (en) 2013-07-08 2015-11-03 Arm Limited Data store and method of allocating data to the data store
US9946588B2 (en) * 2014-12-17 2018-04-17 International Business Machines Corporation Structure for reducing power consumption for memory device
US11442867B2 (en) * 2018-12-20 2022-09-13 Micron Technology, Inc. Using a second content-addressable memory to manage memory burst accesses in memory sub-systems
US11106609B2 (en) 2019-02-28 2021-08-31 Micron Technology, Inc. Priority scheduling in queues to access cache data in a memory sub-system
US10970222B2 (en) 2019-02-28 2021-04-06 Micron Technology, Inc. Eviction of a cache line based on a modification of a sector of the cache line
US10908821B2 (en) * 2019-02-28 2021-02-02 Micron Technology, Inc. Use of outstanding command queues for separate read-only cache and write-read cache in a memory sub-system
US11288199B2 (en) 2019-02-28 2022-03-29 Micron Technology, Inc. Separate read-only cache and write-read cache in a memory sub-system
US20210157738A1 (en) * 2019-11-26 2021-05-27 International Business Machines Corporation Recoverable user cache within recoverable application memory within volatile memory
US11775216B2 (en) * 2021-08-30 2023-10-03 Micron Technology, Inc. Media access operation command management using media buffers

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4442487A (en) * 1981-12-31 1984-04-10 International Business Machines Corporation Three level memory hierarchy using write and share flags
US5802568A (en) * 1996-06-06 1998-09-01 Sun Microsystems, Inc. Simplified least-recently-used entry replacement in associative cache memories and translation lookaside buffers
US20020159283A1 (en) * 2001-04-26 2002-10-31 International Business Machines Corporation Method and structure for a CAMRAM cache memory
US6625715B1 (en) * 1999-12-30 2003-09-23 Intel Corporation System and method for translation buffer accommodating multiple page sizes
US20040010675A1 (en) * 2002-07-09 2004-01-15 Moritz Csaba Andras Statically speculative memory accessing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4442487A (en) * 1981-12-31 1984-04-10 International Business Machines Corporation Three level memory hierarchy using write and share flags
US5802568A (en) * 1996-06-06 1998-09-01 Sun Microsystems, Inc. Simplified least-recently-used entry replacement in associative cache memories and translation lookaside buffers
US6625715B1 (en) * 1999-12-30 2003-09-23 Intel Corporation System and method for translation buffer accommodating multiple page sizes
US20020159283A1 (en) * 2001-04-26 2002-10-31 International Business Machines Corporation Method and structure for a CAMRAM cache memory
US20040010675A1 (en) * 2002-07-09 2004-01-15 Moritz Csaba Andras Statically speculative memory accessing

Non-Patent Citations (10)

* Cited by examiner, † Cited by third party
Title
Edward S. Tam et al. "Active Management of Data Caches by Exploiting Reuse Information" IEEE Transactions on Computers, vol. 48, No. 11, pp. 1244-1259, Nov. 1999.
Edward S. Tam et al. "Evaluating the Performance of Active Cache Management Schemes" Proceedings of the 1998 International Conference on Computer Design, Oct. 1998.
J. A. Rivers et al. "On effective data supply for multi-issue processors." In Proceedings of ICCD'97, pp. 519-528, Oct. 1997.
J. A. Rivers et al. "Reducing conflicts in direct-mapped caches with a temporality-based design." In Proceedings of the 1996 ICPP, pp. 154-163, Aug. 1996.
J. A. Rivers et al. "Utilizing reuse information in data cache management." In Proceedings of ICIS'98, pp. 449-456, 1998.
J. A. Rivers, Performance Aspects of High-Bandwidth Multi-Lateral Cache Organizations, Apr. 1, 1998.
J.A. Rivers et al., "On High-Bandwidth Data Cache Design for Multi-Issue Processors," Proc. 30th Int'l Symp. Microarchitecture, pp. 46-56, Dec. 1997.
J.A. Rivers et al., "Performance Issues in Integrating Temporality-Based Caching" Elsevier, Aug. 26, 1996.
Juce A. Rivers, Performanceof High-Bandwidth Multi-Lateral Cache Organizations, Apr. 1, 1998, pp. 143-147.
M Zhang and K. Asanovich, Highly Associative Caches for Low-Power Processors, Kool Chips Workshop, 33rd Int'l Symposium on Microarchitecture, , Published in: US.

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013101060A2 (en) * 2011-12-29 2013-07-04 Intel Corporation Efficient support of sparse data structure access
WO2013101060A3 (en) * 2011-12-29 2014-04-10 Intel Corporation Efficient support of sparse data structure access
US9037804B2 (en) 2011-12-29 2015-05-19 Intel Corporation Efficient support of sparse data structure access

Also Published As

Publication number Publication date
US20050044317A1 (en) 2005-02-24

Similar Documents

Publication Publication Date Title
US8984254B2 (en) Techniques for utilizing translation lookaside buffer entry numbers to improve processor performance
CN107111455B (en) Electronic processor architecture and method of caching data
US10025720B2 (en) Cache organization and method
USRE45078E1 (en) Highly efficient design of storage array utilizing multiple pointers to indicate valid and invalid lines for use in first and second cache spaces and memory subsystems
US5751990A (en) Abridged virtual address cache directory
US6912623B2 (en) Method and apparatus for multithreaded cache with simplified implementation of cache replacement policy
US7552277B2 (en) Distributed buffer integrated cache memory organization and method for reducing energy consumption thereof
US6990557B2 (en) Method and apparatus for multithreaded cache with cache eviction based on thread identifier
JP4137641B2 (en) Cache way prediction based on instruction base register
CA2020275C (en) Apparatus and method for reading, writing, and refreshing memory with direct virtual or physical access
US6356990B1 (en) Set-associative cache memory having a built-in set prediction array
US6427188B1 (en) Method and system for early tag accesses for lower-level caches in parallel with first-level cache
US7069388B1 (en) Cache memory data replacement strategy
US5715427A (en) Semi-associative cache with MRU/LRU replacement
US7822911B2 (en) Memory device and method with on-board cache system for facilitating interface with multiple processors, and computer system using same
US7039768B2 (en) Cache predictor for simultaneous multi-threaded processor system supporting multiple transactions
JP2004530995A (en) Fast and accurate cashway selection
US6643733B2 (en) Prioritized content addressable memory
US20100011165A1 (en) Cache management systems and methods
US6240489B1 (en) Method for implementing a pseudo least recent used (LRU) mechanism in a four-way cache memory within a data processing system
US6145057A (en) Precise method and system for selecting an alternative cache entry for replacement in response to a conflict between cache operation requests
US5897651A (en) Information handling system including a direct access set associative cache and method for accessing same
US20040243764A1 (en) Tag array access reduction in a cache memory
US6889290B2 (en) Memory management apparatus and method
US6049849A (en) Imprecise method and system for selecting an alternative cache entry for replacement in response to a conflict between cache operation requests

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RIVERS, JUDE A.;REEL/FRAME:014447/0820

Effective date: 20030819

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

REMI Maintenance fee reminder mailed
FPAY Fee payment

Year of fee payment: 4

SULP Surcharge for late payment
FPAY Fee payment

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20210623