PRIORITY CLAIM
The present application is a continuation of and claims priority from U.S. patent application Ser. No. 12/358,873, filed on Jan. 23, 2009 titled “Optimizing a Cache Back Invalidation Policy,” which is incorporated by reference herein in its entirety and for all purposes.
BACKGROUND
1. Technical Field
The present invention generally relates to cache memory systems and in particular to replacement policies of cache coherence systems.
2. Description of the Related Art
Caches are typically organized in a set-associative fashion, with each cache divided into a certain number of “sets”. Each set holds one or more cache-lines arranged in one or more “ways”. The “ways” and “sets” of a cache are depicted in the illustration of FIG. 4. Each set has Least Recently Used (LRU) bits associated with the set that describes the level of recent use of each cache-line within a set. These LRU bits are used to make replacement decisions when removing a line from a set, in order to make space for a new line. In addition, each cache-line may have other associated state bits. For example, if the cache participates in the coherence protocol (for example, the Modified Exclusive Shared Invalid (MESI) protocol), the cache maintains the coherence protocol state bits per cache-line.
Caches are, often, hierarchically organized into multiple levels which include Level 1 caches (L1) and Level 2 caches (L2). Typically, in a multi-processor system (or a processor designed with the intent of being usable in a multi-processor system), “inclusion” is maintained between cache levels. Inclusion means that all the data stored in a higher level cache is also present in a lower level cache. For example, all data in the (higher level) L1 caches are present in the (lower level) L2 cache that the L1s share. The main motivation for inclusion is that inclusion makes cache coherency easier to manage. A request from the bus may initiate an inspection/search of a lower level cache only and be assured that the search has not missed any data that the higher level caches may contain.
In the example of an L2 shared by a few L1 caches, inclusiveness imposes that upon replacement of a cache line in the L2, the line is consequently invalidated in the L1s that share the L2. The invalidation of the cache-line in the L1 is called back-invalidation. Since the size of an L2 is typically much larger than the combined sizes of all the L1s that share the L2, most back-invalidation messages may end up searching the L1 for a line that the L1 does not have.
In general, the traditional cache replacement policies results in performance degrading back-invalidation caused by the policies' eviction of hot cache-lines from caches closer to cores (e.g. L1) due to back-invalidation from inclusive caches that are further from the cores (e.g. L2).
SUMMARY
Disclosed are a method, a system and a computer program product for enhancing a cache back invalidation policy by utilizing least recently used (LRU) bits and presence bits in selecting cache-lines for eviction. A cache back invalidation (CBI) utility evicts cache-lines by using presence bits to avoid replacing a cache-line in a lower level cache that is also present in a higher level cache. Furthermore, the CBI utility selects the cache-line for eviction from an LRU group. The CBI utility ensures that dormant cache-lines in the higher level caches do not retain corresponding presence bits set in the lower level caches by unsetting the presence bits in the lower level cache when a line is replaced in the higher level cache. Additionally, when a processor core becomes idle, the CBI utility invalidates the corresponding higher level cache by unsetting the corresponding presence bits in the lower level cache.
The above as well as additional objectives, features, and advantages of the present invention will become apparent in the following detailed written description.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention itself, as well as advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
FIG. 1 provides a block diagram representation of a data processing system within which the invention is practices, according to one embodiment of the invention;
FIG. 2 illustrates an example of a traditional cache replacement policy in a processor memory sub-system, according to the prior art;
FIG. 3 illustrates the use of presence bits in a cache hierarchy, according to one embodiment of the invention;
FIG. 4 illustrates the use of LRU and Pseudo LRU bits within a processor cache, according to one embodiment of the invention;
FIG. 5 illustrates the modification to LRU bits in response to a hit on a more recently used cache-line of a way pair, according to one embodiment of the invention;
FIG. 6 illustrates the modification to LRU bits in response to a cache miss, according to one embodiment of the invention;
FIG. 7 illustrates the modification to LRU bits in response to a hit on a less recently used cache-line of a way pair, according to one embodiment of the invention;
FIG. 8 illustrates a cache replacement policy which uses presence bits in the lower level cache to reduce back-invalidate traffic in a processor and memory-subsystem, according to one embodiment of the invention;
FIG. 9 illustrates the use of Pseudo LRU bits and presence bits in selecting cache-lines for eviction, according to one embodiment of the invention; and
FIG. 10 is a flow chart illustrating the process of enhancing cache replacement and back-invalidation policies, according to one embodiment of the invention.
DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENTS
The illustrative embodiments provide a method, a system and a computer program product for enhancing a cache back invalidation policy by utilizing least recently used (LRU) bits and presence bits in selecting cache-lines for eviction. A cache back invalidation (CBI) utility evicts cache-lines by using presence bits to avoid replacing a cache-line in a lower level cache that is also present in a higher level cache. Furthermore, the CBI utility selects the cache-line for eviction from an LRU group. The CBI utility ensures that dormant cache-lines in the higher level caches do not retain corresponding presence bits set in the lower level caches by unsetting the presence bits in the lower level cache when a line is replaced in the higher level cache. Additionally, when a processor core becomes idle, the CBI utility invalidates the corresponding higher level cache by unsetting the corresponding presence bits in the lower level cache.
In the following detailed description of exemplary embodiments of the invention, specific exemplary embodiments in which the invention may be practiced are described in sufficient detail to enable those skilled in the art to practice the invention, and it is to be understood that other embodiments may be utilized and that logical, architectural, programmatic, mechanical, electrical and other changes may be made without departing from the spirit or scope of the present invention. The following detailed description is, therefore, not to be taken in a limiting sense, and the scope of the present invention is defined by the appended claims and equivalents thereof.
Within the descriptions of the figures, similar elements are provided similar names and reference numerals as those of the previous figure(s). Where a later figure utilizes the element in a different context or with different functionality, the element is provided a different leading numeral representative of the figure number. The specific numerals assigned to the elements are provided solely to aid in the description and not meant to imply any limitations (structural or functional or otherwise) on the described embodiment.
It is understood that the use of specific component, device and/or parameter names (such as those of the executing utility/logic described herein) are for example only and not meant to imply any limitations on the invention. The invention may thus be implemented with different nomenclature/terminology utilized to describe the components/devices/parameters herein, without limitation. Each term utilized herein is to be given its broadest interpretation given the context in which that terms is utilized.
With reference now to the figures, and beginning with FIG. 1, there is depicted a block diagram representation of an example data processing system (DPS), as utilized within one embodiment. DPS may be a computer, a portable device, such as a personal digital assistant (PDA), a smart phone, and/or other types of electronic devices that may generally be considered processing devices. As illustrated, DPS 100 comprises at least one processor or central processing units (CPUs) connected to system memory 106 via system interconnect/bus 102. In particular, CPU 101 is illustrated with a two (2) processing cores, Core-0 103 and Core-1 104. CPU 101 further comprises at least one higher level cache of which L1 105 is illustrated. Illustrated within CPU 101 is lower level cache L2 109. Cache controller 112 is also connected to system bus 102. Also connected to system bus 102 is input/output (I/O) controller 115, which provides connectivity and control for input devices, of which pointing device (or mouse) 116 and keyboard 117 are illustrated. I/O controller 120 also provides connectivity and control for output devices, of which display 118 is illustrated. Additionally, a multimedia drive 119 (e.g., compact disk read/write (CDRW) or digital video disk (DVD) drive) and USB (universal serial bus) port 121 are illustrated, coupled to I/O controller 115. Multimedia drive 119 and USB port 121 enable insertion of a removable storage device (e.g., optical disk or “thumb” drive) on which data/instructions/code may be stored and/or from which data/instructions/code may be retrieved. DPS 100 also comprises storage 107, within/from which data/instructions/code may also be stored/retrieved.
DPS 100 is also illustrated with a network interface device (NID) 125, by which DPS 100 may connect to one or more access/external networks 130, of which the Internet is provided as one example. In this implementation, the Internet represents/is a worldwide collection of networks and gateways that utilize the Transmission Control Protocol/Internet Protocol (TCP/IP) suite of protocols to communicate with one another. NID 125 may be configured to operate via wired/or wireless connection to an access point of the network. Network 130 may be an external network such as the Internet or wide area network (WAN), or an internal network such as an Ethernet (local area network—LAN) or a Virtual Private Network (VPN). Connection to the external network 130 may be established with one or more servers 133, which may also provide data/instructions/code for execution on DPS 100, in one embodiment.
In addition to the above described hardware components of DPS 100, various features of the invention are completed/supported via software (or firmware) code or logic stored within memory 106 or other storage (e.g., storage 107) and executed by CPU 101. Thus, for example, illustrated within memory 106 are a number of software/firmware/logic components, including operating system (OS) 108 (e.g., Microsoft Windows®, a trademark of Microsoft Corp, or GNU®/Linux®, registered trademarks of the Free Software Foundation and The Linux Mark Institute), applications 114, Coherence Protocol 111, Pseudo-LRU Policy 113, Enhanced Cache Replacement Policy (CRP) 124 and Cache Back-Invalidation (CBI) utility 110 (which executes on CPU 101 or cache controller 112). In actual implementation, CBI utility 110 may be combined with Coherence Protocol 111, Pseudo-LRU Policy 113 and Enhanced Cache Replacement Policy (CRP) 124 to provide a single executable component, collectively providing the various functions of each individual software component when the corresponding combined code is executed by the CPU 101. For simplicity, CBI utility 110 is illustrated and described as a stand alone or separate software/firmware component, which provides specific functions, as described below.
In one embodiment, servers 133 includes a software deploying server, and DPS 100 communicates with the software deploying server (133) via network (e.g., Internet 130) using network interface device 125. Then, CBI utility 110 may be deployed from/on the network, via software deploying server 133. With this configuration, software deploying server performs all of the functions associated with the execution of CBI utility 110. Accordingly, DPS 100 is not required to utilize internal computing resources of DPS 100 to execute CBI utility 110.
CPU 101 executes CBI utility 110 as well as OS 108, which supports the user interface features of CBI utility 110. In the described embodiment, CBI utility 110 generates/provides several graphical user interfaces (GUI) to enable user interaction with, or manipulation of, the functional features of CBI utility 110. Certain functions supported and/or implemented by CBI utility generate processing logic executed by processor and/or device hardware to complete the implementation of that function. For simplicity of the description, the collective body of code that enables these various features is referred to herein as CBI utility 110. Among the software code/instructions/logic provided by CBI utility 110, and which are specific to the invention, are: (a) code/logic for responding to data requests using an enhanced/optimized cache replacement policy which initiates a search for data in a higher level cache; (b) code/logic for selecting for eviction a cache-line from a lower level cache by using one or more of presence bits, presence indicators and LRU bits, according to the optimized cache replacement policy; (c) code/logic for updating LRU bits, according to the optimized cache replacement policy; and (d) code/logic for executing enhanced cache back invalidation. According to the illustrative embodiment, when CPU 101 executes CBI utility 110, DPS 100 initiates a series of functional processes that enable the above functional features as well as additional features/functionality. These features/functionality are described in greater detail below within the description of FIGS. 3-10.
Those of ordinary skill in the art will appreciate that the hardware components and basic configuration depicted in FIG. 1 may vary. The illustrative components within DPS 100 are not intended to be exhaustive, but rather are representative to highlight essential components that are utilized to implement the present invention. For example, other devices/components may be used in addition to or in place of the hardware depicted. The depicted example is not meant to imply architectural or other limitations with respect to the presently described embodiments and/or the general invention. The data processing system depicted in FIG. 1 may be, for example, an IBM eServer pSeries system, a product of International Business Machines Corporation in Armonk, N.Y., running the Advanced Interactive Executive (AIX) operating system or LINUX operating system.
FIG. 2 illustrates a cache replacement policy in a processor and memory-subsystem, according to the prior art. System 200 comprises processor 202 which further comprises two cores, Core-0 203 and Core-1 204. Illustrated within Processor 202 is first L1 205, the independent higher level cache of Core-0 203. Second L1 207 represents the independent higher level cache of Core-1 204. Also illustrated within system 200 is lower level (L2) cache 209. L2 cache 209 is shared by Core-0 203 and Core-1 204. Processor 202 connects to the memory subsystem (comprising memory controller 216 and memory 206 to which access is controlled by memory controller 216) via a memory bus.
In traditional cache eviction policies, selection of the cache-line for eviction is based upon identifying the LRU line. System 200 illustrates an example of processor 201 with two cores, Core-0 203 and Core-1 204. System 200 has L2 209 that is shared by a pair of L1s, first L1 205 and second L1 207, respectively, and may be used to highlight performance issues with traditional cache eviction policies.
When one of the cores, say core-0 203, runs a workload that is cache-intensive and core-0's data set fits in L1 205, the data remains “hot” (i.e., is actively fetched) in L1 205. However, the data starts “cooling” (i.e., is infrequently fetched) in L2 209 because the LRU bits in L2 209 are never updated (because the data is always hit in core-0's L1 205). As a result, the chances of eviction of core-0's hot data from L2 209 are increased. In addition, the chance of eviction of the same/replicated data from the L1 205 (due to back-invalidation) is also increased because core0's data in L2 209 is not recent. When core-1 204 runs a workload that has a large working set, continuous fetching of new lines into L2 209 and core-1's (second) L1 207 occurs. As a result, core-1's data becomes more recent in L2 than core-0's data, eventually causing eviction of core-0's data from L2 209 and core-0's L1 205 due to back-invalidation. So, even though the data is heavily reused by core-0 203, core-0's data gets evicted because core-1 204 “hogs” L2 209. Consequently, the workload on core-0 203 pays a significant performance penalty because core-0 203 experiences a lot of cache misses even though core-0 203 is actually very cache friendly.
In general, the traditional cache replacement policies results in performance degrading back-invalidation caused by the policies' eviction of hot cache-lines from caches closer to cores (e.g. L1 205) due to back-invalidation from inclusive caches that are further from the cores (e.g. L2 209).
FIG. 3 illustrates the use of presence bits in a cache hierarchy, according to one embodiment of the invention. FIG. 3 comprises DPS 301 and DPS 100. As shown in DPS 100 and DPS 301, caches may be hierarchically organized into multiple levels including Level 1 caches (L1) and Level 2 caches (L2). In a multi-processor system (or a processor designed with the intent of being usable in a multi-processor system), inclusion is maintained between cache levels. This means that all the data stored in a higher level cache is also present in a lower level cache. For example, all data in the L1 caches is present in the L2 cache that the L1s share.
The main motivation for inclusion is that is makes cache coherency easier to manage. A request from the bus only needs to check a lower level cache and be assured that the check has not missed any data that the higher level caches may contain. Let us take the example of an L2 shared by a few L1 caches. Inclusiveness imposes that upon replacement of a cache line in the L2, the line is invalidated in the L1s that share the L2. This is called back-invalidation. To reduce the back-invalidation traffic, L2s maintain “presence bits” to indicate which L1s do have the cache line. Back-invalidation is illustrated in DPS 301. In DPS 301 a single (higher level cache) L1 308 uses L2 314. In this case only 1 presence bit 322 (maintained within cache-line 318) suffices in the L2. In DPS 100, there are two (2) L1s (L1 307, L1 105) sharing L2 109. There are 2 presence bits (324, 326) per cache-line maintained in L2 316. In DPS 100, the highlighted cache-line (312) is present in L1 105 and not in L1 307. Therefore, the presence bits indicate that presence appropriately. When the highlighted cache-line in the L2 needs to be replaced, only a back-invalidate needs to be sent to L1 105, which has the line, as indicated by the presence bits. In particular, first presence bit-2 326 has a value of “0” indicating that no back-invalidate is sent to corresponding L1 307 whereas second presence bit-2 324 has a value of “1” indicating that a back-invalidate is sent to corresponding L1 105. Therefore, using the presence bits only the necessary back-invalidates need to be sent to the L1s. Of course, to keep the presence bits up to date in the L2, upon an L1 replacement the corresponding presence bits are unset in the L2. To keep the impact of this extra traffic to the L2 low and to parallelize the data access due to the L1 miss, and the update of the “presence bits” due to the replacement, presence bits may be maintained in a separate structure in the L2.
FIG. 4 illustrates the use of LRU and Pseudo LRU bits within a processor cache, according to one embodiment of the invention. L2 cache 109 comprises 4 “ways” illustrated by way-A 402, way-B 404, way-C 406 and way-D 408, respectively. L2 cache 109 also comprises LRU bits 407 further comprising LRU-2 410, LRU-1 412 and LRU-0 414.
Pseudo-LRU is an approximation of True LRU used for the eviction of cache-lines. True LRU maintains a true ranking of how recently a cache-line in a cache set is used. True LRU uses “n multiplied by log 2n” bits per set for an n-way cache because each of the n ways uses “log 2n” bits to encode the corresponding ranking (which ranges in value from 0 through n−1). Pseudo-LRU uses “n−1” bits per set for an n-way cache. As illustrated, L2 cache 109 is a single cache set in 4-way cache. In L2 cache 109, 3 (i.e., n=4−1=3) bits are used per set to maintain “recency”, i.e., a measure of recent use of a cache-line. A cache consists of a number of sets which is indicative of the cache size.
In a Pseudo-LRU policy/scheme with 4-ways, the scheme indicates 3 possibilities. Of the way pairs A-B and C-D, the policy indicates which way pair is more recently used. Of ways A 402 and B 404, the policy indicates which way is more recently used. Of ways C 406 and D 408, the policy indicates which way is more recently used. Thus, in this example of FIG. 4, Pseudo-LRU policy/scheme uses only 3 bits which are described as follows: LRU1 412 “remembers”/indicates which way-pair between pair A-B and pair C-D is LRU; LRU2 410 remembers which way between ways A 402 and B 404 is LRU; and LRU0 414 remembers which way between ways C 406 and D 408 is LRU.
LRUO and LRU2 form, what may be termed, level-1 bits of Pseudo-LRU and LRU1 forms level-2 bits. Similarly, an 8-way cache requires 7 pseudo-LRU bits, with 4 bits in level-1, 2 bits in level-2 and 1 in level-3. The Pseudo-LRU concept similarly extends to higher levels of associativity, for example, to the level of associativities of a 32-way cache and a 64-way cache.
The following three illustrations (FIG. 5, FIG. 6 and FIG. 7) show modifications to the LRU bits in response to cache hits and cache misses.
FIG. 5 illustrates the modification to LRU bits in response to a hit on a more recently used cache-line of a way pair, according to one embodiment of the invention. In L2 109 of FIG. 5, arrows are used to indicate the meaning/value of the LRU bits. The arrows point towards less recently used regions of the cache set. In particular, the arrows (i.e., following the path of the arrows illustrated with a distinctively bolder appearance and depicted by LRU1 504 and LRU2 508, respectively) lead to way B 404, indicating that way B 404 is the least recently used way according to pseudo-LRU bits 407. When a cache hit is detected in way A 402, there is no change to LRU bits 407 because way B 404 continues to be the least recently used way.
In one embodiment, CBI utility 110 modifies/updates (one or more of) LRU bits 407 following the replacement of an evicted cache-line (after a cache miss) similar to the method in which updates are performed following a cache hit in the same cache-line. In either case (i.e., replacement of an evicted cache-line or a cache hit on the same cache-line, which cache-line represents, for example, the more recently used cache-line of an LRU pair), the cache-line represents the most recently used cache-line (following cache-line replacement or cache hit). Thus, the LRU cache-line within the LRU group continues to be the least recently used way and there is no change to LRU bits 407.
FIG. 6 illustrates the modification to LRU bits in response to a cache miss, according to one embodiment of the invention. L2 109 of FIG. 6 illustrates a modified set of pseudo-LRU bits following a cache miss in L2 109 of FIG. 5. Since way B 404 contains the LRU cache-line during the cache miss, the cache-line in way B 404 is evicted and replaced. As a result, way B 404 currently contains a line that is recently brought in, and therefore becomes the most recently used way. Hence LRU bits 407 are rearranged. All bits that lead to way B 404 are flipped, as illustrated by arrows 602 and 604, respectively. The intent is here is that since way B 404 is now the most recently used way, LRU2 410 points to way A 402. The value of LRU1 412 is flipped/changed to point to the less recently used way pair.
FIG. 7 illustrates the modification to LRU bits in response to a hit on a less recently used cache-line of a way pair, according to one embodiment of the invention. L2 109 of FIG. 7 illustrates a modified set of pseudo-LRU bits following a cache hit on way C 406 in L2 109 of FIG. 5. Following the cache hit in way C 406, way B 404 continues to be the least recently used way. Thus, in L2 109 of FIG. 7, LRU1 412 and LRU2 410, which lead to way B 404, are left unchanged. However, between ways C 406 and D 408, way C 406 is currently more recently used and hence the value of LRU0 414 is flipped/changed. Now LRU0 414 points to way D 408 which is indeed the less recently used way amongst way C 406 and D 408.
FIG. 8 illustrates a cache replacement policy which uses presence bits in the lower level cache to reduce back-invalidate traffic in a processor and memory-subsystem, according to one embodiment of the invention. DPS 100 comprises processor 101 which further comprises two cores, Core-0 103 and Core-1 104. Illustrated with Processor 101 is first L1 105, the independent higher level cache of Core-0 103. Second L1 309 represents the independent higher level cache of Core-1 104. Also illustrated within DPS 100 is lower level (L2) cache 109. L2 cache 109 is shared by Core-0 103 and Core-1 104. DPS 100 also includes cache controller 112. Processor 101 is connected to the memory subsystem (comprising memory controller 804 and memory 106 to which access is controlled by memory controller 804). Memory 106 includes CBI utility 110.
The optimized cache replacement/eviction policy may be described with reference to DPS 100. The optimized replacement policy also applies to systems (having L3 caches, for example) which may include cache levels which are lower than the L2 cache level. The optimized replacement policy is described in relation to the LRU policy.
When a core (e.g., Core-0 103) initiates a read request for data, CBI utility 110 executes the data search as outlined in the following enumerated steps.
(1) CBI utility 110 searches in L1 105. If data is available or “hits” in L1 105, then
-
- (A) The data is read into the core's internal registers.
- (B) The LRU bits and state bits for the cache-line are updated in L1 105.
- (C) CBI utility 110 concludes the data search in L1 105.
(2) If a cache miss occurs in L1 105, CBI utility 110 searches for data in L2 109. If data is available in L2 109, then
-
- (A) Data is copied into the corresponding L1 105 by evicting the LRU cache-line.
- (B) Data is copied into the core's internal registers for further use.
- (C) LRU bits and state bits for the cache-line in are updated in L1 105 and L2 109.
- (D) The presence bits are updated in L2 109 for the cache-line that moved into L1 105.
- (E) The presence bits are also updated for the cache-line evicted from L1 105 to indicate that the cache-line is no longer present in L1 105.
- (F) CBI utility 110 concludes the data search in L2 109.
Further discussion of the optimized cache replacement policy is enhanced with the illustration of FIG. 9.
FIG. 9 illustrates the use of Pseudo LRU bits and presence bits in selecting cache-lines for eviction, according to one embodiment of the invention. L2 Cache 109 previously defined in FIGS. 4-7 is further illustrated with presence bits 910. P0 912 is one element of presence bits 910. Also illustrated in L2 cache 109 is way pair A-B/LRU group 906 and way pair C-D/MRU group 908.
CBI utility 110 utilizes presence bits 910 to indicate which higher level caches a cache line is present in. Unlike typical caches where only the LRU bits are used for selection of cache-lines for eviction, CBI utility 110 employs both LRU bits 407 and presence bits 910 in the process of selecting a cache-line for eviction. Conversely, in traditional LRU policy, the cache-line pointed to by LRU bits as being least recently used is evicted even if the corresponding presence bit indicates that the cache-line is present in a higher level (L1) cache.
In L2 cache 109, every cache-line is associated with Pseudo-LRU bits 407 and p presence bits (e.g., presence bits 910), where p is the number of processors. In the example of FIG. 9, “p” has a value of 4 in L2 cache 109. A presence bit of 1 indicates that the cache-line is present in the corresponding L1.
By incorporating presence bits 910 into eviction decisions, CBI utility 110 provides an optimized cache eviction policy. Pseudo-LRU partitions the “n” ways in a set into 2 groups: a more recently used (MRU) group (e.g., MRU 908); and a less recently used (LRU) group (e.g., LRU 906). The LRU group comprises all the ways in the set that are pointed to by the LRU bits (e.g., A 402, B 404). All other ways comprise the MRU group (e.g., C 406, D 408). For the purpose of cache eviction, CBI utility 110 selects from cache-lines that are in LRU group 906.
Referring again to the optimized cache replacement policy of FIG. 8, CBI utility 110 responds to a second cache miss in L2 109 which second cache miss is preceded by a first cache miss in L1 105 by utilizing presence bits (e.g., presence bits 910) in the eviction decision and incorporating the pseudo-LRU policy. CBI utility 110 continues the execution of the optimized cache eviction policy as described in the following (continuation of) enumerated steps:
(3) In response to a cache miss in L1 105 followed by a cache miss in L2 109:
-
- (A) The selection of the line to be evicted in the L2 set is made based on the following:
- (I) CBI utility 110 determines whether the LRU cache-line has the associated presence bit set (e.g., P0 912).
- (II) If the LRU cache-line (e.g., B 404) does not have the associated presence bit set:
- (IIa) CBI utility 110 selects the LRU cache-line for eviction; and
- (IIb) CBI utility 110 updates/modifies the LRU bits in response to a cache miss as described for L2 109 in FIG. 6.
- (III) If CBI utility 110 previously determined that the LRU cache-line has the associated bit set, CBI utility 110 determines whether all (remaining) cache-lines in the LRU group have the corresponding presence bits set.
- (IV) If all cache-lines in LRU group 906 have the corresponding presence bits set:
- (IVa) CBI utility 110 selects the LRU cache-line for eviction; and
- (IVb) CBI utility 110 updates/modifies the LRU bits in response to a cache miss as described for L2 109 in FIG. 6.
- (V) If CBI utility 110 previously determined that the LRU cache-line has the associated bit set and all (remaining) cache-lines in the LRU group do not have the corresponding presence bits set:
- (Va) CBI utility 110 randomly selects a cache-line from LRU group 906 that does not have a corresponding presence bit set; and
- (Vb) CBI utility 110 updates/modifies the LRU bits in response to a cache hit on a less recently used cache-line of a way pair as described for L2 109 in FIG. 7.
- (B) CBI utility 110 inspects the presence bit for the evicted line in L2 109.
- (I) If the evicted cache-line is present in any L1, the evicted cache-line is invalidated in the L1(s) to maintain the inclusive policy.
- (II) If the evicted cache-line is not present in any L1, then no back-invalidates are sent to the L1(s). A cache-line is evicted from the requesting core's L1 using LRU policy of L1, to make space for the incoming line.
- (C) Once a cache-line is evicted from both L1 105 and L2 109, the new line is brought into L1 105, L2 109 and internal core registers.
- (D) The presence bits, LRU bits and state bits are appropriately set in L1 and L2 109.
FIG. 10 is a flow chart illustrating the process of enhancing cache replacement and back-invalidation policies, according to one embodiment of the invention. FIG. 10 is a flow chart illustrating the method by which the above process of the illustrative embodiment is completed. Although the method illustrated in FIG. 10 may be described with reference to components shown in FIGS. 1, 3-9, it should be understood that this is merely for convenience and alternative components and/or configurations thereof can be employed when implementing the various methods. Key portions of the methods may be completed by CBI utility 110 executing on processor 101 within DPS 100 (FIG. 1) and controlling specific operations of/on DPS 100, and the methods are thus described from the perspective of either/both CBI utility 110 and DPS 100.
The process of FIG. 10 begins at initiator block 1002 and proceeds to block 1004, at which CBI utility 110 detects a read request. At block 1006, CBI utility 110 searches for data in L1 105. CBI utility 110 detects a cache miss in L1 105, as shown at block 1008. At block 1010, CBI utility 110 searches for data in L2 109. CBI utility 110 detects a cache miss in L2 109, as shown at block 1012. CBI utility 110 selects a cache-line for eviction from LRU group 906 based on the values of presence bits 910, as shown at block 1014. At block 1016, CBI utility 110 updates LRU bits 407 appropriately. CBI utility 110 initiates eviction of an L1 cache-line, as shown at block 1018. At block 1020, CBI utility 110 initiates appropriate back-invalidations. A new cache-line is brought in to L1(s) and L2 109, as shown at block 1022. The presence bits, LRU bits and state bits are appropriately set in L1(s) and L2 109, as shown at block 1024. The process ends at block 1026.
Although the invention is described with respect to a level 2 cache, the invention applies to any cache level which maintains presence bits to indicate the presence of a cache-line at a higher level of cache. Though described with respect to using pseudo-LRU replacement policy as the base replacement policy, the invention may be similarly extended to any other replacement policy (even non-LRU based policies). The invention uses the presence bits to avoid replacing a line that is still in a higher level cache. Though the invention refers to selection of a replacement candidate from the LRU group (less recently used lines) as one embodiment, in another embodiment, a replacement candidate may be selected from a group of cache-lines within the LRU to MRU order, as long as the cache-line does not have any presence bits set.
The invention places a higher priority on a line in a lower level cache (L2) which is stored in any of the higher level caches (L1s) than a line which is not. Further, given that a line present in any higher level of cache is more important than a line that is not present in any of the higher level caches, the invention ensures that stale/dormant lines in the higher level caches do not continue to keep the corresponding presence bits set in the lower level caches. Therefore, the processor cores firstly unset the presence bits in the lower level cache when a line is replaced in the higher level cache. Secondly, when a processor core goes idle (e.g., completes running a particular program), the core invalidate all the corresponding higher level cache (for example, the L1 cache). Thus, the corresponding presence bits in the lower level cache (for example L2) are unset, preventing those stale lines in the higher level cache from unnecessarily retaining space in the lower level cache.
Though the presence bits are described as being encoded as one full bit-vector, i.e., one bit per higher level cache that shares the lower level cache (which maintains the presence bits), the invention applies to any other type of presence indicators. For example, to reduce the overhead of presence bits, coarser vectors may be used in designs embodying the invention. With coarser vectors, a single presence bit indicates the presence of the line in, potentially, more than one higher level cache. Although the unsetting of presence bits in the lower level cache on a replacement in the higher level cache is more complex with coarser vectors, the invention, which focuses on the application of presence bits/indicators to eviction decisions, may be applied to designs with coarser vectors as well.
In the flow charts above, one or more of the methods are embodied in a computer readable medium containing computer readable code such that a series of steps are performed when the computer readable code is executed (by a processing unit) on a computing device. In some implementations, certain processes of the methods are combined, performed simultaneously or in a different order, or perhaps omitted, without deviating from the spirit and scope of the invention. Thus, while the method processes are described and illustrated in a particular sequence, use of a specific sequence of processes is not meant to imply any limitations on the invention. Changes may be made with regards to the sequence of processes without departing from the spirit or scope of the present invention. Use of a particular sequence is therefore, not to be taken in a limiting sense, and the scope of the present invention extends to the appended claims and equivalents thereof.
As will be appreciated by one skilled in the art, the present invention may be embodied as a method, system, and/or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module,” “logic”, or “system.” Furthermore, the present invention may take the form of a computer program product on a computer-usable storage medium having computer-usable program code embodied in or on the medium.
As will be further appreciated, the processes in embodiments of the present invention may be implemented using any combination of software, firmware, microcode, or hardware. As a preparatory step to practicing the invention in software, the programming code (whether software or firmware) will typically be stored in one or more machine readable storage mediums such as fixed (hard) drives, diskettes, magnetic disks, optical disks, magnetic tape, semiconductor memories such as RAMs, ROMs, PROMs, etc., thereby making an article of manufacture in accordance with the invention. The article of manufacture containing the programming code is used by either executing the code directly from the storage device, by copying the code from the storage device into another storage device such as a hard disk, RAM, etc., or by transmitting the code for remote execution using transmission type media such as digital and analog communication links. The medium may be electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system (or apparatus or device) or a propagation medium. Further, the medium may be any apparatus that may contain, store, communicate, propagate, or transport the program for use by or in connection with the execution system, apparatus, or device. The methods of the invention may be practiced by combining one or more machine-readable storage devices containing the code according to the described embodiment(s) with appropriate processing hardware to execute the code contained therein. An apparatus for practicing the invention could be one or more processing devices and storage systems containing or having network access (via servers) to program(s) coded in accordance with the invention. In general, the term computer, computer system, or data processing system can be broadly defined to encompass any device having a processor (or processing unit) which executes instructions/code from a memory medium.
Thus, it is important that while an illustrative embodiment of the present invention is described in the context of a fully functional computer (server) system with installed (or executed) software, those skilled in the art will appreciate that the software aspects of an illustrative embodiment of the present invention are capable of being distributed as a program product in a variety of forms, and that an illustrative embodiment of the present invention applies equally regardless of the particular type of media used to actually carry out the distribution. By way of example, a non exclusive list of types of media, includes recordable type (tangible) media such as floppy disks, thumb drives, hard disk drives, CD ROMs, DVDs, and transmission type media such as digital and analogue communication links.
While the invention has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the invention. In addition, many modifications may be made to adapt a particular system, device or component thereof to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to the particular embodiments disclosed for carrying out this invention, but that the invention will include all embodiments falling within the scope of the appended claims. Moreover, the use of the terms first, second, etc. do not denote any order or importance, but rather the terms first, second, etc. are used to distinguish one element from another.