US7343598B2 - Cache-conscious coallocation of hot data streams - Google Patents
Cache-conscious coallocation of hot data streams Download PDFInfo
- Publication number
- US7343598B2 US7343598B2 US10/737,205 US73720503A US7343598B2 US 7343598 B2 US7343598 B2 US 7343598B2 US 73720503 A US73720503 A US 73720503A US 7343598 B2 US7343598 B2 US 7343598B2
- Authority
- US
- United States
- Prior art keywords
- coallocation
- solution
- computer
- objects
- heap
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
- G06F12/02—Addressing or allocation; Relocation
- G06F12/08—Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
- G06F12/0802—Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F12/00—Accessing, addressing or allocating within memory systems or architectures
- G06F12/02—Addressing or allocation; Relocation
- G06F12/0223—User address space allocation, e.g. contiguous or non contiguous base addressing
- G06F12/023—Free address space management
Definitions
- the invention relates generally to cache memory performance. More specifically, it relates to improving cache memory performance by coallocating objects and/or object fields from hot data streams.
- a processor is the part of a computer which controls the other parts.
- a processor typically includes units for fetching instructions and processing them to produce signals which control the other parts of the computer.
- the speed of a processor affects how quickly the processor performs operations.
- Computer memory in a general sense, is a medium or device that can hold data in a format that the processor can use.
- the data in computer memory usually include computer instructions as well as data operated upon according to the computer instructions.
- Data stored in memory is often organized into data objects, which may in turn include attributes or fields.
- Data objects can correspond to a wide variety of data items or data structures, such as program variables and constants, arrays, records, or other aggregate data structures.
- One way to improve the performance of software on computer systems is to install large amounts of fast memory.
- installing large amounts of fast memory is expensive—generally, the faster the memory, the more expensive it is.
- larger memory is generally slower to access than smaller memory due to the lookup circuitry. So, even if fast memory is cheap, smaller memory will still be faster than larger memory, and there will still be a memory hierarchy.
- An economical solution to this problem is to have a memory hierarchy within the computer system, which takes advantage of the cost/performance benefit of different memory technologies.
- the hierarchy is based on memories of different speeds and sizes, which are organized into several levels, each smaller, faster, and more expensive per byte than the next level. To work correctly, the levels of the hierarchy usually encompass one another.
- cache memory In the memory hierarchy, cache memory is the name generally given to the first, topmost level of memory encountered by a processor. It is typically the fastest memory in terms of performance, as well as the most expensive per byte. Some processors have separate caches for instructions and other data, where both can be active at the same time. Caches can use various kinds of addressing (e.g., direct mapped, fully associative, set associative). Effective cache memory utilization is an important determinant of overall program performance and may help bridge the processor-memory performance gap. Thus, improving cache memory performance is vital to software performance.
- cache memory is typically small, fast memory holding recently accessed data, designed to speed up subsequent access to the same data.
- a copy is also saved in the cache along with the associated main memory address.
- the system monitors addresses of subsequent reads to see if the data is already in the cache. If the data is already in the cache (called a cache hit), then the data is returned immediately and the main memory read is aborted (or not started). If the data is not cached (called a cache miss), then the data is fetched from main memory and also saved in the cache, which can be time-consuming and can cause processing to wait for the data to be loaded.
- the cache is built from faster memory chips than main memory so a cache hit takes much less time to complete than a normal memory access.
- the cache may be located on the same integrated circuit as the CPU in order to further reduce the access time.
- the cache is often known as “primary” since there may be a larger, slower secondary cache (i.e., a lower level in the hierarchy) outside the CPU chip.
- the hit rate is its hit rate—the fraction of all memory accesses which are satisfied from the cache.
- the hit rate depends on the cache design as well as its size relative to the main memory. The size is limited by the cost of fast memory chips.
- the hit rate also depends on the access pattern of the particular program being run (the sequence of addresses being read, written, etc.).
- caches rely on two properties of the access patterns of most programs: (1) temporal locality—if something is accessed once, it is likely to be accessed again soon; and (2) spatial locality—if one memory location is accessed, then nearby memory locations are likely to be accessed. In order to exploit spatial locality, caches often operate on several words (units of storage in a computer) at a time—a “cache line” or “cache block.” Main memory reads and writes are whole cache blocks.
- a cache block is the smallest unit of memory that can be transferred between the memory and the cache. Rather than reading a single word or byte from main memory at a time, each cache entry usually holds a certain number of words, and a whole block is read and cached at once. This takes advantage of the principle of spatial locality: if one location is read, then nearby locations (particularly, following locations) are likely to be read soon afterward.
- Cache miss rate is the fraction of accesses that are not in cache memory. Miss penalties are the additional time required to service a cache miss (e.g., to perform a normal, non-cache memory access). Cache miss reductions indicate the number or proportion of misses that could be avoided based on a given layout.
- Object frequency profiles typically rely on processor access history to make caching decisions and keep the most frequently used objects in cache.
- object frequency profiles typically cannot find an optimal cache layout solution because some objects may accumulate large reference counts and never become candidates for replacement, even if the objects are no longer active. Aging techniques counterbalance the accumulation effect, but they add additional levels of complication to the profile.
- a pairwise affinity profile maintains information on how many cache misses would be caused if a pair of objects are mapped to the same cache block.
- One drawback is that the profile limits the number of objects it evaluates to two.
- a pairwise affinity profile needs to compute every combination of object pairings and the effects of the object pairings on the overall cache layout. These computations require intensive and complicated heuristics that may ultimately result in failing to find a good solution for many kinds of caches.
- the Seidl article describes allocating heap objects in four pre-defined memory “arenas” based on a predicted hit-miss ratio.
- the four arenas are four different areas of memory, labeled highly referenced, not-highly reference, short lived, and other.
- the Rubin article describes using a search-based learning technique to classify heap objects according to runtime characteristics such as allocation calling context, object size, and other like characteristics.
- the Rubin article also describes, based on this classification, allocating objects in separate heap arenas.
- the techniques in the Seidl and Rubin articles improve virtual memory performance by increasing page utilization.
- the chilimbi 2 article describes ccmalloc, a cache-conscious heap allocator that uses programmer annotations to allocate contemporaneously accessed data objects to be in the same cache block.
- One major drawback to this technique is that it requires programmer intervention—the programmer manually places the annotations in the software, which can be time-consuming.
- Another drawback is that the technique requires that the programmer accurately assess the run-time behavior of the software, which may be difficult to discern for some scenarios and may change in different scenarios.
- Calder describes applying placement techniques developed for instruction caches to data. Specifically, Calder describes a compiler-directed approach that creates an address placement scheme for stack variables, global variables, and heap objects in order to reduce data cache misses. Calder describes calculating a temporal relationship graph (“TRG”).
- TRG temporal relationship graph
- the TRG improves performance for stack objects and global variables but does little to improve the caching performance of heap objects.
- drawbacks to the TRG include the fact that the TRG does not use hot data stream profiles and that it depends on an arbitrary temporal reference window size, which may not improve cache miss rates for some programs. For additional details, see the respective papers.
- a tool such as a coallocation tool receives a temporal data reference profile and detects hot data streams in it.
- the tool analyzes the hot data streams and the profile to determine a coallocation solution for allocations in heap memory. This enables cache-conscious coallocation of elements such as objects or object fields in the hot data streams during subsequent execution of a computer program, for example, by isolating and directing allocations of the hot data streams to areas of memory.
- a tool such as a coallocation tool receives a profile of object field accesses and determines a coallocation solution from the profile.
- the coallocation solution increases the locality of object fields in memory to improve cache performance, especially since object fields are small and more likely to fit with other object fields in a single cache block.
- a tool such as a coallocation tool receives a comprehensive temporal data access profile for a computer program and detects data access patterns in it.
- the tool analyzes the patterns and profile to determine a coallocation solution for allocations in memory.
- the coallocation solution may include object restructuring by reordering fields of objects, splitting objects, and/or merging objects. Such restructuring techniques improve subsequent cache-conscious coallocation.
- the tool also includes an enforcement module for automatically enforcing the coallocation solution during execution of the altered computer program. In this way, the tool simplifies the alteration of the computer program and enforcement of the coallocation solution.
- FIG. 1 is a block diagram of a suitable computing system for implementing described embodiments.
- FIG. 2A is a block diagram of a data reference trace that includes hot data streams.
- FIG. 2B is a block diagram of a data layout in which objects are placed contiguously in allocation order.
- FIG. 2C is a block diagram of a layout after cache-conscious coallocation.
- FIG. 3A is a chart showing the results of hot data stream analysis.
- FIG. 3B is a chart showing accumulated and normalized coallocation set miss reductions.
- FIG. 3C is a chart showing possible partitions of allocation sites, in terms of weighted set packings.
- FIG. 3D is a block diagram of a layout for the top set packing in FIG. 3C .
- FIG. 4 is a flowchart showing a technique for finding a solution for coallocating hot data streams.
- FIG. 5A is a pseudocode listing for an algorithm for coallocating hot data streams.
- FIG. 5B is a pseudocode listing for an algorithm for computing weights for coallocation sets.
- FIG. 6 is a block diagram of a memory layout showing objects too large for effective cache-conscious coallocation at the object level of granularity.
- FIG. 7A is a block diagram showing objects and fields for allocation sites.
- FIG. 7B is a block diagram of a layout at the object level without cache-conscious coallocation.
- FIG. 7C is a block diagram of a layout at the object field level without cache-conscious coallocation.
- FIG. 7D is a block diagram of a layout at the object field level after field reordering.
- FIG. 7E is a block diagram of a layout at the object field level after object splitting and merging, with field reordering.
- FIG. 8 shows a generalized technique for finding a solution for field coallocation.
- FIGS. 9A and 9B are a pseudocode listing for an algorithm for field coallocation.
- FIG. 10A is a chart showing analysis of coallocatable object fields and coallocation sets for hot data streams on a field access trace.
- FIG. 10B is a chart showing the accumulated and normalized cache miss reductions due to object field coallocation.
- FIG. 10C is a chart showing the results of applying WSP for fields, when object-level allocation sites are used.
- FIG. 10D is a chart showing WSP field solutions.
- FIG. 10E is a block diagram of the layout of object fields in accordance with split or merge coallocation.
- FIG. 11 is a block diagram of a coallocation tool.
- FIGS. 12A-12F are charts showing results with respect to several benchmark computer programs.
- the following description is directed to techniques and tools for improving cache memory performance. Described embodiments improve memory system performance by coallocating contemporaneously accessed heap objects.
- the techniques and tools use relatively precise global temporal profile information that is nevertheless both cheap to collect and relatively stable across different program runs.
- the techniques and tools use an efficient polynomial approximation algorithm to process the profile information and produce a good data layout, rather than rely on ad-hoc heuristics.
- the techniques include profile-based analysis for producing an effective memory layout for a program.
- the heap is divided into multiple sequentially allocated arenas, and an analysis tool achieves cache-conscious coallocation of a hot data stream H by isolating and directing allocation sites of H to the same sequentially allocated arena such that intervening allocations coming from other sites are separated.
- the coallocation solution produced by the analysis is enforced by an automatic tool that redirects a program's heap allocations to a run-time coallocation library.
- the techniques and tools extend to coallocation at object field granularity.
- the resulting field coallocation solution generalizes common data restructuring techniques, such as field reordering, object splitting, and object merging.
- use of the techniques and tools yields insight into object restructuring by breaking down the coallocation benefit on a per-technique basis, which provides the opportunity to pick the “sweet spot” for each program.
- the techniques described herein are implemented after compile time but before run time. In other embodiments, the techniques are implemented at design time, compile time, or runtime.
- a hot data stream is a set of two or more data references that repeats relatively frequently.
- Various embodiments focus on objects or object fields in hot data streams since such objects/fields fit into that category of frequent accesses that are only a small fraction of all of the data objects/fields in a program.
- other implementations may extend to other environments and definitions as well.
- a cache-conscious coallocation tool performs various techniques. Although the operations for these techniques are typically described in a particular, sequential order for the sake of presentation, it should be understood that this manner of description encompasses minor rearrangements in the order of operations, unless a particular ordering is required. For example, operations described sequentially may in some cases be rearranged or performed concurrently. Moreover, for the sake of simplicity, flowcharts typically do not show the various ways in which particular techniques can be used in conjunction with other techniques.
- FIG. 1 illustrates an exemplary system for practicing the invention according to one or more embodiments.
- the system includes computing device 100 .
- computing device 100 typically includes at least one processing unit 102 and system memory 104 .
- Processing unit 102 includes single physical processors, multiple processors acting together, virtual processors, and any other device or software program capable of interpreting and executing computer instructions.
- system memory 104 may be volatile (such as RAM), non-volatile (such as ROM, flash memory, etc.), or some combination of the two, in various levels of a memory hierarchy.
- System memory 104 typically stores software and data for an operating system 105 , one or more program modules with their associated data 106 , and a cache-conscious coallocation tool 114 .
- Computing device 100 may also have additional features or functionality.
- computing device 100 may also include additional data storage devices (removable and/or non-removable) such as magnetic disks, optical disks, or tape.
- additional storage is illustrated in FIG. 1 by removable storage 107 and non-removable storage 108 .
- Computer storage media may include volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.
- System memory 104 , removable storage 107 , and non-removable storage 108 are all examples of computer storage media.
- Computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, DVD or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store information and which can be accessed by computing device 100 .
- Computing device 100 may also have input device(s) 112 such as keyboard, mouse, pen, voice input device, touch input device, etc.
- Output device(s) 110 such as a display, speakers, printer, and the like may also be included. All these devices are known in the art and need not be discussed at length here.
- Computing device 100 may also include communications connection(s) 113 that allow the device to communicate with other computing devices 120 , such as over a network.
- Communications connection(s) 113 is an example of communication media, which typically convey computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
- modulated data signal means a signal that has one or more of its characteristics set or changed is such a manner as to encode information in the signal.
- communication media includes wired media such as a wired network or direct wired connections, and wireless media such as acoustic, RF, infrared, and other wireless media.
- the term computer readable media as used herein includes both storage media and communication media.
- Embodiments can be described in the general context of computer-executable instructions, such as those included in program modules, being executed in a computing system on a target real or virtual processor.
- computer programs and program modules include routines, libraries, objects, classes, components, data structures, etc. that perform particular tasks or implement particular abstract data types.
- the functionality of the program modules may be combined or split between program modules as desired in various embodiments.
- Computer-executable instructions for program modules may be executed within a local or distributed computing system.
- the application uses terms like “determine,” “generate,” “adjust,” and “apply” to describe computer operations in a computing system. These terms are high-level abstractions for operations performed by a computer, and should not be confused with acts performed by a human being. The actual computer operations corresponding to these terms vary depending on implementation.
- Memory system performance is improved by coallocating contemporaneously accessed heap objects into memory arenas using a robust and efficient algorithm. This is accomplished using relatively precise global temporal profile information that is nevertheless both cheap to collect and relatively stable across different program runs. In addition, this is accomplished using an efficient polynomial approximation algorithm to process the profile information and produce a good data layout, rather than relying on ad-hoc heuristics.
- this section provides an example of cache-conscious coallocation and a summary of the tools and techniques presented in detail in other sections.
- a heap object is a data object allocated from “heap” memory.
- the “heap” is an area of memory used for dynamic memory allocation, where blocks of memory are typically allocated and freed in an arbitrary order, and the pattern of allocation and size of blocks may not be known until run-time.
- a program has one heap which it may use for several different purposes.
- a heap is required by some languages in which functions can return arbitrary data structures or functions with free variables.
- the functions malloc and free provide access to the heap.
- Other languages have the same or similar functions, or perform dynamic memory allocation in a more automatic manner.
- the function malloc is supported by a standard C library routine for storage allocation.
- Malloc takes the number of bytes to be allocated and returns a pointer to a block of memory that size. Storage is allocated from a heap, which lies after the end of the program and data areas. Following good programming practice, memory allocated with malloc is freed explicitly using the “free” routine, and can then be re-used. Allocation sites are the sites in a program at which the program allocates memory.
- WPS whole program stream
- Cache-conscious coallocation of a hot data stream H can be achieved by isolating and directing allocation sites of H's data members (i.e., the objects that are members of the hot data stream H) to a sequentially allocated arena, such that intervening allocations coming from other sites are separated.
- the heap is divided into multiple sequentially allocated arenas, and an analysis tool determines which allocation sites to direct to the same arena to achieve coallocation, and which allocation sites prevent or hinder coallocation and should be isolated in a separate arena.
- Objects in hot data streams typically incur most of the cache misses, but are small fraction of all data objects.
- FIG. 2A shows a portion of a data reference trace 202 of a computer program 201 .
- FIG. 2A uses the following convention: object x i is allocated at allocation site X (or calling context X) and is the i th global allocation request in the program.
- object b 1 is allocated at allocation site B, and this is the first object allocation in the program.
- the object e 6 is the sixth allocation request in the program, and this object is allocated at allocation site E.
- the trace 202 reflects a part of execution that occurs after the data objects have been allocated, which is why earlier allocated objects appear after later allocated objects in the trace 202 .
- the data reference trace 202 contains a hot data stream H 203 .
- the hot data stream H 203 has the members a 3 , b 5 , and c 10 .
- Allocation sites A, B, and C allocate objects a 3 , b 5 , and c 10 , respectively, along with other objects that are not part of the hot data stream H 203 .
- Coallocating the data objects of the hot data stream H 203 which are frequently accessed contemporaneously during the trace 202 , can significantly reduce cache misses, especially if H fits in a cache block.
- FIG. 2B shows an example data layout 211 in heap memory of system memory, which assumes objects are small and thus placed contiguously according to the order by which the objects were allocated in the program. Each subsequent object is allocated sequentially.
- the object allocated first (b 1 ) is followed by the object allocated second (d 2 ), which is followed by the object allocated third (a 3 ), etc.
- the objects a 3 , b 5 , and c 10 that make up the hot data stream H 203 are not placed contiguously in memory in this layout 211 .
- This layout 211 is likely to be inefficient since, following it, the objects a 3 , b 5 , and c 10 probably will not be in a single cache block.
- the objects a 3 , b 5 , and c 10 are separated by other objects (from allocation sites D and E) in the layout 211 .
- a 3 , b 5 , and c 10 may be in three different cache blocks, or a 3 and b 5 may be in a different block than c 10 , or there may be some other configuration.
- the layout 211 likely will result in cache misses and extra non-cache memory access operations to access the objects in the hot data stream.
- FIG. 2C shows a memory layout in which co-location of the objects in H is achieved.
- allocation requests from sites A, B, and C are isolated and directed to a first sequentially allocated arena 231 , while allocation requests from other sites (i.e., D and E) are separated and directed to a different arena 241 .
- other objects allocated at the sites A, B, and C are also placed in the same heap arena 231 as the hot data stream objects, this is acceptable as far as the objective to co-locate hot data stream objects. (It is not always possible to co-locate hot data stream members in this fashion.
- a profile-based analysis tool produces a cache-conscious coallocation of heap objects that participate in hot data streams, when possible.
- Optimal coallocation can be reduced to a weighted set packing (“WSP”) problem, which is known to be NP hard.
- WSP weighted set packing
- the tool uses the best known polynomial time approximation to the WSP problem to arrive at the object coallocation solution.
- a profile-based analysis tool uses other and/or additional techniques for analysis.
- an instrumentation tool replaces a program's heap allocation requests with calls to a run-time coallocation library to enforce the coallocation solution layout. For example, suppose a computer program includes the following lines:
- coallocating fields is a generalization of the combination of object field reordering, object splitting, and object merging
- this analysis permits estimating the contribution of each individual technique to the field coallocation solution.
- Restrictions on field rearrangements yield a spectrum of coallocation solutions, which tradeoff ease of layout enforcement (especially for non-typesafe languages such as C, C++) versus larger cache miss rate reductions.
- most of the benefits of the general solution arbitrary field rearrangement
- these field coallocation solutions produce execution time improvements of up to 22% on current machines for a suite of memory-performance-limited benchmarks.
- This section describes techniques for coallocating heap objects that participate in hot data streams.
- the section briefly reviews hot data stream construction, introduces terminology and metrics for hot data streams, and discusses a simple model to estimate the benefits of coallocation.
- the section describes when hot data stream objects may be coallocated.
- a cache-conscious object coallocation algorithm is presented.
- a hot data stream is a frequently repeated set of data references (in other words, a frequent data access pattern).
- the chilimbi 1 article describes using a compression algorithm called Sequitur to construct a compact context-free grammar representation of a data reference trace, which could then be efficiently analyzed to detect hot data streams.
- the “heat” of a data stream is its length (number of references in the data stream) multiplied by its frequency (number of time the data stream occurs in a trace).
- a minimal hot data stream has at least the minimal prefix (exceeding length 1) of a hot data stream, with a heat c or more.
- the size of hot data streams detected by the analysis is also adjustable and implementation dependent. For example, the analysis is set to detect minimal hot data stream that contain at least 2 and not more than 20 elements. While longer hot data streams offer more coallocation opportunities, they also increase the computational cost of the coallocation analysis.
- the analysis also computes the normalized heat value for each hot data stream in B, such that a hot data stream H with normalized heat value H c covers H c % of the reduced trace. In other words, H covers (P*H c )% of the original trace.
- hot data streams may actually be sets of data references in arbitrary order, which are still considered and termed hot data streams.
- analysis ignores global variable references.
- FIG. 2B shows a sequence of objects allocated by a program during execution.
- the focus is on “linearizing” the layout to the access pattern and ignores both object and cache block size. The goal is to achieve a robust solution on a variety of machines with different cache architectures rather than tune a layout to a particular cache configuration.
- FIGS. 3A-3D are an elaboration of FIGS. 2A-2C .
- FIG. 3A shows the result of the hot data stream analysis. As shown in the normalized heat column (“ht”), the hot data stream a 3 b 5 c 10 covers 60% of the reduced trace, the data stream d 4 b 15 e 6 covers 30% of the reduced trace, and the data stream e 6 d 4 a 12 a 14 covers 10% of the reduced trace. ( FIG. 3A shows results of analysis of a trace with several hundred data references, a portion of which is shown in FIG. 2A .)
- Improvements in performance may be deduced from the data.
- every stream data access can cause a cache miss.
- coallocating b 5 and c 10 in the same arena could save one cache miss for every occurrence of the stream a 3 b 5 c 10 (as shown in the misses reduced column “mr”), assuming b 5 and c 10 fit in one cache block.
- Coallocating a 3 , b 5 , and c 10 saves two cache misses if those objects fit in one cache block.
- the potential cache miss reduction for two other layouts are (i) the original layout produced by the current heap allocator, and (ii) the layout produced if all objects are allocated contiguously according to allocation order.
- the member objects of a hot data stream H can be coallocated if all intervening allocations between members of H come from allocations sites other than the ones used to allocate members of H.
- the objects b 5 and c 10 that come from allocation sites B and C, respectively
- the intervening allocations of e 6 , d 7 , d 8 , and d 9 come from other allocation sites (namely, D and E).
- d 4 and b 15 which are in a hot data stream d 4 b 15 e 6 , are not coallocatable since d 7 (and also d 8 and d 9 ) is an intervening allocation that comes from D, which allocates d 4 in H.
- objects in H are coallocatable if, for every program object X t ⁇ H, (where x t denotes an object that is allocated at site X by the t th global allocation request in the program): min( T alloc ( H )) ⁇ t ⁇ max( T alloc ( H )) X ⁇ S alloc ( H ) where:
- coallocation set denotes the set of allocation sites of coallocatable objects.
- a coallocation set includes the allocation sites for a particular group of data objects to be coallocated in an arena.
- ⁇ A, B, C ⁇ is the coallocation set for coallocatable objects a 3 , b 5 , and c 10 . Enforcing this coallocation set saves two cache misses for every occurrence of the hot data stream a 3 b 5 c 10 according to the model presented in the preceding sub-section.
- FIG. 4 shows a technique 400 for finding a solution for coallocating hot data streams, with the goal of finding a coallocation strategy that maximizes cache miss reduction.
- An analysis tool performs the technique.
- FIG. 5A shows a corresponding pseudocode listing for an algorithm for coallocating hot data streams in some embodiments.
- an analysis tool uses a different technique for finding a solution for coallocating hot data streams. For example, the tool uses additional and/or different criteria for identifying the allocation context (e.g., with call stack signatures) or measuring coallocation benefits.
- FIG. 5B shows a pseudocode listing for the first phase of the pseudocode of FIG. 5A .
- the analysis tool performs involves an exponential exploration of the possible coallocations. This is tractable since hot data streams are usually small (and recall that, with typical settings, the analysis detects minimal hot data stream with length at least 2 and not exceeding 20).
- hot data streams are usually small (and recall that, with typical settings, the analysis detects minimal hot data stream with length at least 2 and not exceeding 20).
- the technique uses a polynomial algorithm to find the maximal partition of coallocatable objects that come from a coallocation set.
- a weighted miss reduction for coallocation set ⁇ A,E ⁇ only accounts for coallocating the objects e 6 , a 12 , a 14 .
- the tool accumulates the potential miss reduction per coallocation set and then normalizes it with respect to total expected cache misses in the model, as shown in FIG. 3B .
- the tool uses additional and/or different criteria to compute or evaluate coallocation sets.
- the tool computes a partition of the set of allocation sites, such that cache miss reduction is increased.
- the partition problem is an instance of a known NP-hard problem—weighted set packing (“WSP”).
- WSP weighted set packing
- Given a set S of m base elements and a collection C ⁇ C 1 , C 2 , . . . , C n ⁇ of weighted subsets of S with a weight function w, find a subcollection C′ ⁇ C of disjoint sets of maximum total ⁇ C i ′ ⁇ C′ w(C i ′).
- S is the set of hot allocation sites (i.e., the set of sites allocating at least one object that participates in a hot data stream).
- C is the set of coallocation sets, and w(C i ) is the normalized potential cache miss reduction.
- the best approximation algorithm for the WSP problem is that of Halldórsson. This algorithm approximates WSP to within 2 ⁇ square root over (m) ⁇ of the optimal solution in time proportional to the time it takes to sort the weights. Although the approximation algorithm may yield poor results for large values of m, this is not a significant problem for most benchmarks.
- FIG. 3C presents possible partitions of allocation sites, in terms of weighted set packings.
- partition/set packing ⁇ A, B, C ⁇ and ⁇ D, E ⁇
- Applying Halldórsson's WSP algorithm indeed yields this partition.
- FIG. 3D presents the layout of objects using the ⁇ A, B, C ⁇ , ⁇ D, E ⁇ partition.
- the objects of the hot data stream a 3 b 5 c 10 are allocated contiguously.
- the objects d 4 and e 6 which participate in two hot data streams, are also colocated.
- This layout also allocated a 12 and a 14 for “free”—this coallocation could be attributed to ⁇ A, B, C ⁇ as well.
- the tool computes the potential cache miss reduction for other layouts (e.g., the “original layout” and “allocation order layout,” discussed previously). And, in the fourth phase 440 , the tool relates these to R o .
- the resulting number (R o 1 ) reflects the expected cache miss reduction over a given layout. Since in this example, R 1 for both layouts (i.e., the original layout and allocation order layout) is 0, the benefits for coallocating ⁇ A, B, C ⁇ , ⁇ D, E ⁇ are 76% over both layouts.
- the tool uses additional and/or different evaluation criteria, for example, using more, fewer, or even no comparisons.
- FIG. 6 shows another version of the memory layout of FIG. 3D , in which the data objects have different sizes. Certain objects (for example, the object a 3 ) may be larger than a cache block. Or, two objects (for example, the objects b 1 and a 3 ) may be too large to be put in a cache block together.
- a typical cache block is 32 bytes, though processors such as the P4 and Itanium have 128 byte L2 cache blocks.
- FIG. 7A shows an allocation site A allocating a 3-field object, and allocation sites B and C allocating 2-field objects, respectively.
- a 1 , A 2 , and A 3 are corresponding field allocation sites, for independent allocation of the fields of A.
- FIG. 7B shows a memory layout 711 for 4 objects (a 1 , b 2 , a 3 , and c 4 ) allocated at the object-level allocation sites A, B, and C.
- the objects are laid out in the order by which they were allocated.
- the layout 721 is expressed at a field granularity, where a i,j denotes the j th field global allocation request in the program, which is allocated at field allocation site A i (the i th field of an object allocated at A).
- the data object a 3 is allocated by three consecutive field allocation requests coming from the field allocation sites A 1 , A 2 , and A 3 allocating fields a 1,6 , a 2,7 , and a 3,8 respectively.
- R 1 denotes the potential cache miss reduction over a given layout (“original layout” or “allocation order layout” are used), supposing that combinations of field reordering, object splitting, and object merging are permitted in object restructuring.
- techniques are used to coallocate fields while limiting object restructuring to simple mechanisms, such as field reordering, object splitting and object merging.
- the techniques use any one or more of four field coallocation solutions categorized as follows, depending on the restriction of the object restructuring mechanisms used:
- coallocation techniques consider coallocation opportunities at the field granularity
- solutions are enforceable at the object allocation site level. This eliminates the need to transform an object allocation into multiple field allocations, but may require object restructuring to split an object or merge objects for allocation of the appropriate data. Implementing the resulting coallocation solutions is thus much simpler in many cases, at the cost of some lost opportunities for field coallocation.
- the following restricted coallocation criteria are considered:
- Fields in a hot data stream are “field reordering coallocatable” if field reordering suffices to guarantee coallocation of the fields.
- the fields a 1,1 and a 3,3 are field reordering coallocatable, since reordering the fields of A (i.e., by placing the first and the third field together) ensures coallocation.
- FIG. 7D demonstrates the layout 731 after this field reordering.
- Fields in a hot data stream are “object split coallocatable” if object splitting (either with or without field reordering) suffices to ensure coallocation of these fields.
- object splitting either with or without field reordering
- FIG. 7C the fields a 3,3 and a 1,6 are object split coallocatable, since splitting the fields of A, while placing the first and the third fields of instances of A contiguously, ensures coallocation (without intervening instances of objects for the second field of A).
- FIG. 7E demonstrates a layout 741 after splitting the fields of A.
- Fields in a hot data stream are “object merge coallocatable” if merging the types of the respective object allocation sites (either with or without field reordering) ensures coallocation of these fields.
- the fields b 1,4 and c 2,10 are object merge coallocatable, since merging the types allocated at B and C (and reordering the fields of the combined type to place the first field of B together with the second field of C) ensures coallocation.
- FIG. 7E demonstrates a layout 742 resulting after merging B and C and reordering the fields to place together the first field of B and the second field of C.
- FIG. 8 shows a generalized technique 800 for finding a solution for field coallocation.
- a tool such as the analysis tool cmanal performs the technique 800 .
- another tool performs the technique.
- FIGS. 9A and 9B shows a corresponding pseudocode listing for an algorithm for field coallocation in some embodiments.
- the tool first changes 810 the coallocation criteria for the first phase of the basic algorithm.
- the parameter coallocatable can reflect different coallocation criteria.
- the tools instead of checking that objects in a hot data stream H are coallocatable, the tools check that objects in H are field reorder coallocatable, object split coallocatable, and/or object merge coallocatable.
- the pseudocode shown in FIG. 9A all three options are evaluated in the “first phase” shown in FIG. 9A .
- the tool uses additional and/or different coallocation criteria.
- the tool explores 820 simple placement opportunities at the level of field allocation sites.
- the tool then computes 830 the benefits of different placement techniques when allocation sites are at object level. For example, the tool uses WSP approximation to compute the benefit of using a particular simple placement technique at the object allocation site level. In the pseudocode shown in FIG. 5A , this occurs in the “second phase” shown in FIG. 5A .
- the tool uses additional and/or different strategies or measures for evaluating placement opportunities.
- splitting and reordering are not applicable to more than one allocation site at a time.
- the tool does not combine several allocation sites and then split them.
- Object merging is applicable for two or more allocation sites that have not been split.
- the tool considers the benefits of simple reordering in the merge coallocation solution.
- FIG. 10A shows analysis of coallocatable object fields and coallocation sets for hot data streams on a field access trace.
- the two leftmost columns of FIG. 10A show the result of hot data stream analysis on a field access trace.
- the stream H 1 a 1,1 a 3,3 a 1,6 a 3,8 covers 80% of the reduced trace
- the stream H 2 a 2,2 b 1,4 c 2,10 covers 20% of the reduced trace.
- FIGS. 10A and 10B show the results of applying different field coallocatability definitions to the hot data streams, where (R), (S), (M) denote field reorder benefits, object split benefits, and object merge benefits, respectively.
- FIG. 10C shows the results of applying WSP for fields, when object-level allocation sites are used.
- R field reordering
- S object splitting
- S again only A has benefits (by splitting objects coming from A so the first and third fields of these objects are placed contiguously— FIG. 10E shows A's split layout).
- M object merging
- the tool finds 840 a field allocation solution.
- the tool applies the WSP approximation once again for the placement techniques to approximate the optimal disjoint subsets of object allocation sites that are candidates for layout changes.
- WSP is actually the identity function since all the coallocation sets are of size 1, thus they are pairwise disjoint.
- R 1 r , R 1 s , and R 1 m denote the reduction of cache misses over a given layout for field reorder coallocation, object split coallocation and object merge coallocation, respectively.
- the tool uses additional and/or different strategies or criteria for finding a field coallocation solution.
- FIG. 10D shows the WSP field solutions for field reordering coallocation, object splitting coallocation, and object merging coallocation.
- a potential cache miss reduction of 0.57 is achieved.
- the WSP solution for object splitting (S) shows that splitting the objects at allocation site A (placing the first and third field together as shown in FIG. 10C ) results in a potential cache miss reduction of 0.86.
- the WSP solution for object merging (M) for merging A, B, and C results in a potential cache miss reduction of 0.14.
- FIG. 10D also shows a hybrid solution—split or merge—where an object at an allocation site may be either split or participate in a merge with object(s) from other object allocation sites.
- the fields of these objects may also just be reordered.
- the tool may compute the maximal benefit obtained by split and reorder or by merging (collectively denoted by w ).
- Applying WSP to object coallocation sets with w gives the coallocation solution for the “split or merge” option.
- FIG. 10D shows the benefit of the WSP solution for the s m option.
- Splitting A yields 0.86 potential cache miss reduction
- merging B and C gives 0.07 more potential reduction, for a total potential miss reduction of 0.93.
- FIG. 10E shows the layout obtained by applying the split or merge coallocation solution.
- the area between a 1,6 a 3,8 and a 2,2 is reserved for placing the first/third field objects of further allocations requests from that split of site A.
- the next section provides details on how this is implemented.
- R 1 denotes the expected cache miss reduction of split or merge over a given other layout (e.g., either “original layout” or “allocation order layout”). It is expected that both R 1 r ⁇ R 1 s ⁇ R 1 ⁇ R 1 and R 1 r ⁇ R 1 m ⁇ R 1 ⁇ R 1 .
- an analysis tool uses a different technique for finding a solution for field-level coallocation, using additional and/or different criteria for identifying the allocation context, identifying coallocation opportunities, or measuring coallocation benefits.
- a programming language such as Java or C#
- a compiler might implement such splitting at compile time.
- common language runtime software might implement such splitting at runtime, for example, controlling field level allocation at the policy level, or making other decisions about field reordering, object splitting, and/or object merging.
- FIG. 11 shows a generalized tool 1100 for cache-conscious coallocation.
- the tool 1100 includes modules for profiling and analysis, as well as modules for enforcement of a coallocation solution. While FIG. 11 shows a particular configuration of modules, modules of the tool 1100 can be split, merged, replaced with similar modules, or otherwise modified while still performing cache-conscious coallocation.
- the profiling and analysis module cmanal produces a coallocation solution that is enforced by the instrumentation module cminstr and a coallocation library comalloc.
- the tool 1100 includes one or more profiling modules 1110 for profiling a computer program 1101 .
- the profiling modules produce a profile 1115 such as a data reference trace that shows temporal access pattern behavior at object, field, or other level.
- the reference abstraction granularity is determined when the trace is produced or at some other time.
- the tool cmanal uses Sequitur to compress a reference trace, which is obtained with low-overhead using the technique described in the article Hirzel et al., “Bursty Tracing: A Framework for Low-overhead Temporal Profiling,” 4 th Workshop on Feedback-Directed and Dynamic Optimization, pages 117-126, December 2001. This produces a context-free grammar representation for the profile 1115 .
- the tool 1100 uses other and/or additional modules for profiling.
- the tool 1100 includes one or more analysis modules 1120 for analyzing the profile 1115 and determining one or more cache-conscious coallocation solutions 1125 .
- the analysis module(s) 1120 receive a profile 1115 such as a data reference trace for subsequent analysis.
- the tool cmanal efficiently analyzes a context-free grammar representation for a profile to find hot data streams with their associated normalized heat value.
- the tool cmanal applies a coallocation algorithm (as described in the previous sections) to produce a collection of coallocation sets.
- the allocation context abstraction level i.e., just the allocation site, or a calling context of some length l
- the analysis module(s) include other and/or additional modules.
- the tool 1100 includes one or more enforcement modules 1130 for enforcing the cache-conscious coallocation solution(s) 1125 during execution of the program 1101 .
- the enforcement module(s) 1130 include modules for changing the behavior of the program 1101 at development time, compile time, after compile time but before run time, or at run time.
- the enforcement module(s) 1130 may include modules for presenting the solution(s) 1125 to a programmer and assisting the programmer in implementing the solution(s) 1125 .
- the enforcement module(s) 1130 may include modules for modifying the executable code (output from the compiler) for the program 1101 in view of the solution(s) 1125 .
- the enforcement module(s) 1130 may include modules for altering the executable code for the program 1101 , for example, by binary rewriting.
- the enforcement module(s) 1130 may include modules for enforcing the solution at run time by modifying the program 1101 or enforcing the behavior of the program 1101 as previously modified for the solution(s) 1125 .
- cminstr is a binary instrumentation tool based on Microsoft's Vulcan binary editing tool (similar to ATOM; see Srivastava et al., “ATOM: A System for Building Customized Program Analysis Tools,” SIGPLAN Conf. on Prog. Lang. Design and Impl., 1994).
- the tool cminstr enforces a coallocation solution produced by the tool cmanal by replacing a program's original heap allocation calls (e.g., malloc calls) with calls to routines of the coallocation library comalloc.
- the comalloc calls are implemented by an independent dynamic link library that handles coallocation of objects or fields.
- the instrumented program, running in conjunction with the comalloc library, then uses memory according to the coallocation solution layout.
- the enforcement module(s) include other and or additional modules.
- the enforcement module(s) inject code into a program for enforcing coallocation in some other manner.
- a coallocation tool enforces a coallocation solution by reserving a separate memory arena M i for each coallocation set C i . For consecutive allocation requests from sites in the same coallocation set C i , the tool assigns consecutive addresses in M i . Allocation sites not in any coallocation set are assigned to a default heap arena M 0 .
- the tool uses the heap layers infrastructure to implement this memory management policy. (See Berger et al., “Composing High-performance Memory Allocators,” SIGPLAN Conf. on Prog. Lang. Design and Impl., June 2001.) Heap layers permit managing several separate heaps, where each heap is independently managed, possibly with a different policy. Alternatively, a tool uses other software to enforce a coallocation solution.
- each heap M i is managed with comalloc functions (i.e, functions of the form comalloc i ) allocating memory from the respective heap.
- comalloc functions i.e, functions of the form comalloc i
- corealloc i functions of the form comalloc i
- cocalloc i cocalloc i
- Free operations are implemented by recording the owner heap of every object in its header. Every heap includes a “ownerHeap” layer as part of its type. When calling a “free” function, an actual free i is dispatched to the owner heap of the object being freed. This additional owner header information may reduce some of the coallocation benefit. As an alternative, maintaining ownership information separately from the object likely produces larger coallocation benefits, since objects accesses are typically much more frequent than free operations.
- Instance interleaving is one way to split an object. Instance interleaving splits object instances, such that frequently accessed instance fields are laid out contiguously in memory. (See Truong et al., “Improving Cache Behavior of Dynamically Allocated Data Structures,” IEEE Parallel Arch. And Compiler Techniques, October 1998.) In one implementation, this is done by adding special padding fields to the object type definition and allocating the objects using an instance interleaving library ialloc, which maintains the invariant of placing frequently accessed fields of object instances contiguously.
- the ialloc library may be adapted to work with the heap layers infrastructure, in which case a iallocHeap layer is used to enforce splitting an object. Alternatively, a tool uses other software to implement and/or enforce object splitting.
- One way to do object merging is to combine the object type definitions (i.e., so that the combined type contains all the fields of the types being merged) and use a new heap layer mergeHeap for allocations. For example, suppose the allocation sites A, B, . . . , K are merged. Upon allocation requests from any of these sites, mergeHeap determines if a new object (large enough to store the combined object type) should be allocated, or if the last allocated combined object can be re-used to satisfy the allocation request.
- a and B are merged allocation sites, and allocation requests come from A, then B and then again from B.
- the routine mergeHeap In response to the allocation request from site A, the routine mergeHeap returns a new object o 1 large enough to hold the combined type of A, B.
- the first allocation request from B is returned the address of o 1 so that the B portion of o 1 can be used.
- the next allocation request from B causes allocation of a new combined object o 2 , since the last allocated combined object o 1 was already used to satisfy the prior request from site B.
- a tool uses other software to implement and/or enforce object splitting.
- coallocation tool there are several potential limitations on performance of a coallocation tool. These limitations relate to tracking information, cache block size, and programming languages.
- object size and owner heap information are kept in an 8 byte object header. This header may reduce some of the potential benefits of coallocation, since the header limits opportunities for cache block coallocation. In other implementations, the 8 byte object header has less of an effect because cache block sizes are larger. In still other implementations, object size and owner heap information are tracked in a separate structure.
- a coallocation tool lacks access to type information and support for type manipulations.
- field reordering is done by manual source code modification.
- the instrumentation tool cminstr relies on receiving the PC's of malloc calls to be replaced by comalloc calls.
- Type manipulations affect the binary code, however, so allocation site PC's for a field coallocation solution are adjusted.
- an object coallocation solution which does not involve type manipulations, is enforced automatically.
- splitting array elements is hard to enforce in some languages, since it involves changes in the pointer arithmetic around array expressions. In principle, a compiler can employ such layout, assuming precise type information. An alternative is to enforce array splitting manually (according to the coallocation solution), or to ignore array splitting suggested by the coallocation solution.
- C programs may pose a problem with respect to type manipulation.
- field reordering can crash an application due to hidden assumptions regarding structure layout. This problem is eliminated in strongly-typed languages, such as Java or C#.
- the benchmarks were instrumented with Microsoft's Vulcan tool to produce a data reference trace along with heap allocation information.
- the heap allocation information was used for trace abstraction at an object or field granularity depending on the experiment.
- the traces were processed by cmanal to produce object/field coallocation solutions, which were enforced using cminstr and the comalloc allocation library.
- the coallocation solutions were evaluated on different inputs than the one used to generate the solution. Measurements of CPU time and hardware performance counters were carried out on a 900 Mhz Pentium-III with 1 GB of memory running Windows 2000, with a 256 KB L2 data cache and 16 KB L1 data cache, and 32 byte cache blocks. To minimize the effect of different allocator policies on runtime, the same allocator is used for allocating objects in the original benchmark and for allocating objects not in a coallocation set in the optimized benchmark. The results represent an average of three runs.
- FIG. 12A shows that, except for mst, hot data streams cover around 90% of the trace for both object and field address abstractions.
- the number of hot object addresses is small ( ⁇ 10%) in four of the benchmarks, leading to fast coallocation analysis.
- traces exhibit less regularity, as around 20% of the field addresses participate in hot data streams. This difference is primarily due to arrays, since an array is abstracted to a single object address, but at a field granularity distinct identities are assigned to fields accesses from distinct array elements.
- the problem is that coallocating array elements breaks the invariant that consecutive array elements are contiguous, and can require significant source modifications.
- FIG. 12B presents allocation site characteristics of the benchmarks and the coallocation solutions.
- FIG. 12B shows results for various coallocation solutions in terms of allocation sites.
- the “sites” column presents the number of program allocation sites.
- the “hot” column shows the number of sites that allocate objects that participate in hot data streams.
- the next 6 columns present the number of sites participating in various coallocation solutions, respectively.
- the results indicate significant coallocation opportunity. More sophisticated coallocation techniques create more coallocation opportunity up to a certain point. For example, field reordering finds far less coallocation opportunities than split or merge.
- FIG. 12C shows potential cache miss reduction numbers when coallocating at the granularity of object addresses, where the base (i.e., no improvement) is 0%.
- FIG. 12C shows potential reduction in cache misses with respect to two alternative layouts, original layout and allocation order layout. For the benchmark health, there is no improvement over original layout, indicating that the objects are already placed in a cache-conscious way.
- the potential cache miss reduction is attributed mainly (more than 90% of it in mcf) to coallocating large objects (ranging from 120 bytes to several KBytes) that do not fit in a cache block. Having the analysis take the object and cache block size into account would avoid such situations, though larger cache blocks combined with sequential hardware pre-fetching implemented in next-generation processors could take better advantage of such a layout.
- the actual cache miss reduction is negligible for vpr and mcf. In twolf, the numbers are negative.
- the WSP coallocation solution yields potential benefits over the worst case scenario, but those benefits are less than the benefits of the original layout over the worst case scenario. This is due to the approximation algorithm used, which is bounded to 2 ⁇ square root over (
- FIG. 12B also presents allocation sites characteristics of a field coallocation solution.
- FIG. 12B presents numbers in terms of field allocation sites, since this solution assumes fields are split and merged arbitrarily.
- the number in parentheses is the number of hot sites in terms of field allocation sites.
- FIG. 12D shows potential cache miss reduction numbers for coallocating fields according to various field placement techniques, compared to the original layout.
- FIG. 12D shows results 5 strategies for field coallocation.
- split and merge is preferable, although in boxsim, split or merge performs almost as well. Since split and merge requires sophisticated source transformations, however, the split or merge is employed for twolf as well.
- mcf object splitting suffices to obtain most of the benefits. The solution for mst only requires reordering of fields. In health, object merging suffices to obtain the coallocation benefits.
- the numbers are negative (again due to the approximation algorithm) and coallocation is avoided since the benefits under the coallocation solution are less than the benefits of the original layout achieved by the current heap allocator.
- coallocation tools lack type information. As a result, some type manipulation suggested by the coallocation solutions may not be feasible or may be hard to employ. Other problems arise from the fact that type changes are not always feasible in a non-typesafe environment (i.e., C programs). The following problems may be encountered: (i) If it is assumed that fields are up to 4-bytes long, reordering part of an 8-byte field is not feasible. This can happen for fields of type double, for example, and is the case for boxsim and twolf. (ii) Several allocation sites may be associated with the same type. Thus, a coallocation algorithm that assumes that distinct allocation sites allocate distinct types may suggest conflicting type changes.
- FIG. 12E shows the normalized L2 cache miss reduction for solutions over the original layout, measured using the machine's hardware performance counters.
- FIG. 12E shows reduction in L2 cache misses for a medium and a large input (shown as train and ref, respectively). The coallocation solutions were obtained using a trace for a small input (test input).
- FIG. 12F shows normalized execution time for the benchmarks. Also, the time spent on L1+L2 data cache misses is shown. In boxsim, there is no speedup due to problems of employing the layout. In twolf, there is around 20% speedup, and the time spent in cache misses is reduced by around 27%. For mcf, there is around 12% speedup, and the time spent in cache misses is reduced by around 20%. For mst, there is no speedup. For health, there is around 6% speedup and the time spent in cache misses is reduced by around 30%.
- GC garbage-collection
- Novel profile-based analysis and enforcement techniques aimed at coallocating contemporaneously accessed hot heap objects were presented in various embodiments.
- the analysis techniques obtain cache-conscious coallocation of a hot data stream H by isolating and directing allocation sites of H to a sequentially allocated heap arena such that intervening allocations coming from other sites are separated.
- extended analysis techniques aim for coallocation at object field granularity, generalizing common restructuring techniques, such as field reordering, object splitting, and object merging. Initial results indicate that programs possess significant coallocation opportunities and that field coallocation reduces L2 cache misses by up to 40%, producing execution time improvements of up to 22%.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Memory System Of A Hierarchy Structure (AREA)
Abstract
Description
-
- int * buffer;
- buffer=(int *) malloc (sizeof (int1));
where “buffer” is a pointer and malloc is a memory allocation function that returns a pointer to a block of memory that here is the size of the variable int1. The tool makes the following change to the second line: - buffer=(int *) comalloci (sizeof (int1));
where comalloci is a call to a particular routine that causes the allocation to be in the ith arena. The tool enforces the coallocation solution inFIG. 2C by replacing allocations at sites A, B, and C with calls to comallocl, and replacing allocations at other sites (including sites D and E) with calls to comalloc0. Each comalloci routine manages a separate heap arena where consecutively allocated objects are coallocated. Alternatively, an instrumentation tool uses other and/or additional techniques for instrumentation, for example, replacing a program's heap allocation request in some other way at the programming stage, compile time, after compile time but before run time, or at run time. Or, the run-time allocation libraries use a different technique. For example, instead of having a different comalloci routine per arena, some other combination of routines manages the arenas.
min(T alloc(H))≦t≦max(T alloc(H)) X∉S alloc(H)
where:
- (i) H={xil jl, . . . , xik jk} is a set of k unique objects;
- (ii) xim jm denotes an object allocated at allocation site Xjm by the im th allocation request of the program;
- (iii) Salloc(H)={Xj1, . . . , Xjk} is the set of allocation sites for objects in H; and
- (iv) Talloc(H)={i1, . . . , ik} is the set of allocation request times for the objects in H.
- (i) allocation sites in Salloc(H) allocate in a designated memory arena M;
- (ii) memory arena M allocates consecutive allocation requests contiguously; and
- (iii) intervening objects allocated at sites not in Salloc(H) are allocated in a different memory arena.
- (i) only field reordering is used (“field reorder coallocation”),
- (ii) object splitting is used, either with or without field reordering (“object split coallocation”),
- (iii) merging of objects is used, either with or without field reordering (“object merge coallocation”), and
- (iv) an object can either be split or it can be merged with other objects, but it cannot participate in both operations (“split or merge coallocation”).
Claims (52)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/737,205 US7343598B2 (en) | 2003-04-25 | 2003-12-15 | Cache-conscious coallocation of hot data streams |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US46591703P | 2003-04-25 | 2003-04-25 | |
US10/737,205 US7343598B2 (en) | 2003-04-25 | 2003-12-15 | Cache-conscious coallocation of hot data streams |
Publications (2)
Publication Number | Publication Date |
---|---|
US20040215880A1 US20040215880A1 (en) | 2004-10-28 |
US7343598B2 true US7343598B2 (en) | 2008-03-11 |
Family
ID=33303296
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/737,205 Expired - Fee Related US7343598B2 (en) | 2003-04-25 | 2003-12-15 | Cache-conscious coallocation of hot data streams |
Country Status (1)
Country | Link |
---|---|
US (1) | US7343598B2 (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060070040A1 (en) * | 2002-11-25 | 2006-03-30 | Microsoft Corporation | Dynamic prefetching of hot data streams |
US20060265694A1 (en) * | 2005-05-20 | 2006-11-23 | Microsoft Corporation | Heap-based bug identification using anomaly detection |
US20070250820A1 (en) * | 2006-04-20 | 2007-10-25 | Microsoft Corporation | Instruction level execution analysis for debugging software |
US20070288908A1 (en) * | 2006-06-09 | 2007-12-13 | International Business Machines Corporation | Computer implemented method and system for accurate, efficient and adaptive calling context profiling |
US20080005208A1 (en) * | 2006-06-20 | 2008-01-03 | Microsoft Corporation | Data structure path profiling |
US20080222224A1 (en) * | 2007-03-09 | 2008-09-11 | Kumar Rangarajan | System and Method of Identification of Dangling Pointers |
US20080243300A1 (en) * | 2007-03-26 | 2008-10-02 | Jack Liu | Object relocation guided by data cache miss profile |
US20100115494A1 (en) * | 2008-11-03 | 2010-05-06 | Gorton Jr Richard C | System for dynamic program profiling |
US20100138819A1 (en) * | 2008-11-29 | 2010-06-03 | International Business Machines Corporation | Type merging technique to reduce class loading during java verfication |
US20100229164A1 (en) * | 2009-03-03 | 2010-09-09 | Samsung Electronics Co., Ltd. | Method and system generating execution file system device |
US7912877B2 (en) | 2005-05-20 | 2011-03-22 | Microsoft Corporation | Leveraging garbage collection to dynamically infer heap invariants |
US7962901B2 (en) | 2006-04-17 | 2011-06-14 | Microsoft Corporation | Using dynamic analysis to improve model checking |
US8612952B2 (en) | 2010-04-07 | 2013-12-17 | International Business Machines Corporation | Performance optimization based on data accesses during critical sections |
US20150082285A1 (en) * | 2013-03-15 | 2015-03-19 | Concurix Corporation | Runtime settings derived from relationships identified in tracer data |
US10318415B2 (en) | 2017-05-31 | 2019-06-11 | International Business Machines Corporation | Garbage collection facility grouping infrequently accessed data units in designated transient memory area |
US20200142677A1 (en) * | 2018-11-05 | 2020-05-07 | International Business Machines Corporation | Fields Hotness Based Object Splitting |
Families Citing this family (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002035548A2 (en) * | 2000-10-26 | 2002-05-02 | Matsushita Electric Industrial Co., Ltd. | Storing device, storing control method and program |
US7140008B2 (en) * | 2002-11-25 | 2006-11-21 | Microsoft Corporation | Dynamic temporal optimization framework |
US7124276B2 (en) * | 2003-04-14 | 2006-10-17 | Broadcom Corporation | Optimizing cache efficiency within application software |
US7587709B2 (en) * | 2003-10-24 | 2009-09-08 | Microsoft Corporation | Adaptive instrumentation runtime monitoring and analysis |
US7607119B2 (en) * | 2005-04-26 | 2009-10-20 | Microsoft Corporation | Variational path profiling |
US20070156640A1 (en) * | 2005-12-29 | 2007-07-05 | International Business Machines Corporation | Hot tuple detection |
US8789028B2 (en) * | 2006-06-30 | 2014-07-22 | International Business Machines Corporation | Memory access monitoring |
US20080148241A1 (en) * | 2006-10-11 | 2008-06-19 | Scott Thomas Jones | Method and apparatus for profiling heap objects |
US20080104323A1 (en) * | 2006-10-26 | 2008-05-01 | Colglazier Daniel J | Method for identifying, tracking, and storing hot cache lines in an smp environment |
US7613753B2 (en) * | 2006-11-28 | 2009-11-03 | Alcatel Lucent | Platform and method for functional programming (FP) processing |
US7991967B2 (en) * | 2007-06-29 | 2011-08-02 | Microsoft Corporation | Using type stability to facilitate contention management |
US20090327377A1 (en) * | 2008-06-26 | 2009-12-31 | Tatu Ylonen Oy Ltd | Copying entire subgraphs of objects without traversing individual objects |
US7937419B2 (en) * | 2008-06-26 | 2011-05-03 | Tatu Ylonen Oy | Garbage collection via multiobjects |
US8489840B2 (en) * | 2009-02-26 | 2013-07-16 | Red Hat, Inc. | Discontiguous object addressing |
US8473680B1 (en) * | 2009-09-18 | 2013-06-25 | Marvell International Ltd. | Hotspot detection and caching for storage devices |
US8621150B2 (en) * | 2010-04-09 | 2013-12-31 | International Business Machines Corporation | Data placement optimization using data context collected during garbage collection |
CN104011680B (en) * | 2011-12-26 | 2017-03-01 | 英特尔公司 | The virtual center processing unit of scheduling virtual machine in physical processing unit |
US20140006537A1 (en) * | 2012-06-28 | 2014-01-02 | Wiliam H. TSO | High speed record and playback system |
US10482658B2 (en) * | 2014-03-31 | 2019-11-19 | Gary Stephen Shuster | Visualization and control of remote objects |
JP6555259B2 (en) * | 2014-06-19 | 2019-08-07 | 日本電気株式会社 | Information processing apparatus, data storage method, and program |
US10515430B2 (en) * | 2015-11-03 | 2019-12-24 | International Business Machines Corporation | Allocating device buffer on GPGPU for an object with metadata using access boundary alignment |
US10481817B2 (en) * | 2018-06-28 | 2019-11-19 | Intel Corporation | Methods and apparatus to optimize dynamic memory assignments in multi-tiered memory systems |
CN115328821B (en) * | 2022-10-18 | 2022-12-23 | 北京红山微电子技术有限公司 | Reconfigurable Cache system, memory access system and memory access method based on GPU |
Citations (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5220667A (en) | 1989-09-25 | 1993-06-15 | Mitsubishi Denki Kabushiki Kaisha | Computer system |
US5333311A (en) | 1990-12-10 | 1994-07-26 | Alsoft, Inc. | Optimizing a magnetic disk by allocating files by the frequency a file is accessed/updated or by designating a file to a fixed location on a disk |
US5713008A (en) | 1995-06-08 | 1998-01-27 | Sun Microsystems | Determination of working sets by logging and simulating filesystem operations |
US5740443A (en) | 1995-08-14 | 1998-04-14 | International Business Machines Corporation | Call-site specific selective automatic inlining |
US5774685A (en) | 1995-04-21 | 1998-06-30 | International Business Machines Corporation | Method and apparatus for biasing cache LRU for prefetched instructions/data based upon evaluation of speculative conditions |
US5815720A (en) | 1996-03-15 | 1998-09-29 | Institute For The Development Of Emerging Architectures, L.L.C. | Use of dynamic translation to collect and exploit run-time information in an optimizing compilation system |
US5909578A (en) | 1996-09-30 | 1999-06-01 | Hewlett-Packard Company | Use of dynamic translation to burst profile computer applications |
US5925100A (en) | 1996-03-21 | 1999-07-20 | Sybase, Inc. | Client/server system with methods for prefetching and managing semantic objects based on object-based prefetch primitive present in client's executing application |
US5940618A (en) | 1997-09-22 | 1999-08-17 | International Business Machines Corporation | Code instrumentation system with non intrusive means and cache memory optimization for dynamic monitoring of code segments |
US5950003A (en) | 1995-08-24 | 1999-09-07 | Fujitsu Limited | Profile instrumentation method and profile data collection method |
US5950007A (en) | 1995-07-06 | 1999-09-07 | Hitachi, Ltd. | Method for compiling loops containing prefetch instructions that replaces one or more actual prefetches with one virtual prefetch prior to loop scheduling and unrolling |
US5953524A (en) | 1996-11-22 | 1999-09-14 | Sybase, Inc. | Development system with methods for runtime binding of user-defined classes |
US5960198A (en) | 1997-03-19 | 1999-09-28 | International Business Machines Corporation | Software profiler with runtime control to enable and disable instrumented executable |
US6026234A (en) | 1997-03-19 | 2000-02-15 | International Business Machines Corporation | Method and apparatus for profiling indirect procedure calls in a computer program |
US6073232A (en) | 1997-02-25 | 2000-06-06 | International Business Machines Corporation | Method for minimizing a computer's initial program load time after a system reset or a power-on using non-volatile storage |
US6079032A (en) | 1998-05-19 | 2000-06-20 | Lucent Technologies, Inc. | Performance analysis of computer systems |
US6122638A (en) * | 1997-11-26 | 2000-09-19 | International Business Machines Corporation | Object-oriented processor and method for caching intermediate data in an object-oriented processor |
US6148437A (en) | 1998-05-04 | 2000-11-14 | Hewlett-Packard Company | System and method for jump-evaluated trace designation |
US6311260B1 (en) | 1999-02-25 | 2001-10-30 | Nec Research Institute, Inc. | Method for perfetching structured data |
US6321240B1 (en) | 1999-03-15 | 2001-11-20 | Trishul M. Chilimbi | Data structure partitioning with garbage collection to optimize cache utilization |
US6330556B1 (en) * | 1999-03-15 | 2001-12-11 | Trishul M. Chilimbi | Data structure partitioning to optimize cache utilization |
US6360361B1 (en) | 1999-03-15 | 2002-03-19 | Microsoft Corporation | Field reordering to optimize cache utilization |
US6405292B1 (en) * | 2000-01-04 | 2002-06-11 | International Business Machines Corp. | Split pending buffer with concurrent access of requests and responses to fully associative and indexed components |
US6404455B1 (en) * | 1997-05-14 | 2002-06-11 | Hitachi Denshi Kabushiki Kaisha | Method for tracking entering object and apparatus for tracking and monitoring entering object |
US20020133639A1 (en) * | 1995-12-27 | 2002-09-19 | International Business Machines Corporation | Method and system for migrating an object between a split status and a merged status |
US20020144245A1 (en) | 2001-03-30 | 2002-10-03 | Guei-Yuan Lueh | Static compilation of instrumentation code for debugging support |
US6560693B1 (en) | 1999-12-10 | 2003-05-06 | International Business Machines Corporation | Branch history guided instruction/data prefetching |
US6571318B1 (en) | 2001-03-02 | 2003-05-27 | Advanced Micro Devices, Inc. | Stride based prefetcher with confidence counter and dynamic prefetch-ahead mechanism |
US20030163658A1 (en) * | 2002-02-28 | 2003-08-28 | Lanzatella Thomas W. | System and method for controlling the creation of stable disk images |
US6628835B1 (en) * | 1998-08-31 | 2003-09-30 | Texas Instruments Incorporated | Method and system for defining and recognizing complex events in a video sequence |
US6651243B1 (en) | 1997-12-12 | 2003-11-18 | International Business Machines Corporation | Method and system for periodic trace sampling for real-time generation of segments of call stack trees |
US6658652B1 (en) | 2000-06-08 | 2003-12-02 | International Business Machines Corporation | Method and system for shadow heap memory leak detection and other heap analysis in an object-oriented environment during real-time trace processing |
US6675374B2 (en) | 1999-10-12 | 2004-01-06 | Hewlett-Packard Development Company, L.P. | Insertion of prefetch instructions into computer program code |
US20040015930A1 (en) | 2001-03-26 | 2004-01-22 | Youfeng Wu | Method and system for collaborative profiling for continuous detection of profile phase transitions |
US20040015897A1 (en) | 2001-05-15 | 2004-01-22 | Thompson Carlos L. | Method and apparatus for verifying invariant properties of data structures at run-time |
US20040025145A1 (en) | 1999-05-12 | 2004-02-05 | Dawson Peter S. | Dynamic software code instrumentation method and system |
US6704860B1 (en) | 2000-07-26 | 2004-03-09 | International Business Machines Corporation | Data processing system and method for fetching instruction blocks in response to a detected block sequence |
US20040103401A1 (en) | 2002-11-25 | 2004-05-27 | Microsoft Corporation | Dynamic temporal optimization framework |
US20040103408A1 (en) | 2002-11-25 | 2004-05-27 | Microsoft Corporation | Dynamic prefetching of hot data streams |
US20040111444A1 (en) | 2002-12-06 | 2004-06-10 | Garthwaite Alexander T. | Advancing cars in trains managed by a collector based on the train algorithm |
US20040133556A1 (en) * | 2003-01-02 | 2004-07-08 | Wolczko Mario I. | Method and apparatus for skewing a bi-directional object layout to improve cache performance |
US6848029B2 (en) | 2000-01-03 | 2005-01-25 | Dirk Coldewey | Method and apparatus for prefetching recursive data structures |
US20050091645A1 (en) | 2003-10-24 | 2005-04-28 | Microsoft Corporation | Adaptive instrumentation runtime monitoring and analysis |
US6951015B2 (en) | 2002-05-30 | 2005-09-27 | Hewlett-Packard Development Company, L.P. | Prefetch insertion by correlation of cache misses and previously executed instructions |
US20050246696A1 (en) | 2004-04-29 | 2005-11-03 | International Business Machines Corporation | Method and apparatus for hardware awareness of data types |
-
2003
- 2003-12-15 US US10/737,205 patent/US7343598B2/en not_active Expired - Fee Related
Patent Citations (47)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5220667A (en) | 1989-09-25 | 1993-06-15 | Mitsubishi Denki Kabushiki Kaisha | Computer system |
US5333311A (en) | 1990-12-10 | 1994-07-26 | Alsoft, Inc. | Optimizing a magnetic disk by allocating files by the frequency a file is accessed/updated or by designating a file to a fixed location on a disk |
US5774685A (en) | 1995-04-21 | 1998-06-30 | International Business Machines Corporation | Method and apparatus for biasing cache LRU for prefetched instructions/data based upon evaluation of speculative conditions |
US5713008A (en) | 1995-06-08 | 1998-01-27 | Sun Microsystems | Determination of working sets by logging and simulating filesystem operations |
US5950007A (en) | 1995-07-06 | 1999-09-07 | Hitachi, Ltd. | Method for compiling loops containing prefetch instructions that replaces one or more actual prefetches with one virtual prefetch prior to loop scheduling and unrolling |
US5740443A (en) | 1995-08-14 | 1998-04-14 | International Business Machines Corporation | Call-site specific selective automatic inlining |
US5950003A (en) | 1995-08-24 | 1999-09-07 | Fujitsu Limited | Profile instrumentation method and profile data collection method |
US6886167B1 (en) * | 1995-12-27 | 2005-04-26 | International Business Machines Corporation | Method and system for migrating an object between a split status and a merged status |
US20020133639A1 (en) * | 1995-12-27 | 2002-09-19 | International Business Machines Corporation | Method and system for migrating an object between a split status and a merged status |
US5815720A (en) | 1996-03-15 | 1998-09-29 | Institute For The Development Of Emerging Architectures, L.L.C. | Use of dynamic translation to collect and exploit run-time information in an optimizing compilation system |
US5925100A (en) | 1996-03-21 | 1999-07-20 | Sybase, Inc. | Client/server system with methods for prefetching and managing semantic objects based on object-based prefetch primitive present in client's executing application |
US5909578A (en) | 1996-09-30 | 1999-06-01 | Hewlett-Packard Company | Use of dynamic translation to burst profile computer applications |
US5953524A (en) | 1996-11-22 | 1999-09-14 | Sybase, Inc. | Development system with methods for runtime binding of user-defined classes |
US6073232A (en) | 1997-02-25 | 2000-06-06 | International Business Machines Corporation | Method for minimizing a computer's initial program load time after a system reset or a power-on using non-volatile storage |
US5960198A (en) | 1997-03-19 | 1999-09-28 | International Business Machines Corporation | Software profiler with runtime control to enable and disable instrumented executable |
US6026234A (en) | 1997-03-19 | 2000-02-15 | International Business Machines Corporation | Method and apparatus for profiling indirect procedure calls in a computer program |
US6404455B1 (en) * | 1997-05-14 | 2002-06-11 | Hitachi Denshi Kabushiki Kaisha | Method for tracking entering object and apparatus for tracking and monitoring entering object |
US5940618A (en) | 1997-09-22 | 1999-08-17 | International Business Machines Corporation | Code instrumentation system with non intrusive means and cache memory optimization for dynamic monitoring of code segments |
US6122638A (en) * | 1997-11-26 | 2000-09-19 | International Business Machines Corporation | Object-oriented processor and method for caching intermediate data in an object-oriented processor |
US6651243B1 (en) | 1997-12-12 | 2003-11-18 | International Business Machines Corporation | Method and system for periodic trace sampling for real-time generation of segments of call stack trees |
US6148437A (en) | 1998-05-04 | 2000-11-14 | Hewlett-Packard Company | System and method for jump-evaluated trace designation |
US6079032A (en) | 1998-05-19 | 2000-06-20 | Lucent Technologies, Inc. | Performance analysis of computer systems |
US6628835B1 (en) * | 1998-08-31 | 2003-09-30 | Texas Instruments Incorporated | Method and system for defining and recognizing complex events in a video sequence |
US6311260B1 (en) | 1999-02-25 | 2001-10-30 | Nec Research Institute, Inc. | Method for perfetching structured data |
US6360361B1 (en) | 1999-03-15 | 2002-03-19 | Microsoft Corporation | Field reordering to optimize cache utilization |
US6330556B1 (en) * | 1999-03-15 | 2001-12-11 | Trishul M. Chilimbi | Data structure partitioning to optimize cache utilization |
US6321240B1 (en) | 1999-03-15 | 2001-11-20 | Trishul M. Chilimbi | Data structure partitioning with garbage collection to optimize cache utilization |
US20040025145A1 (en) | 1999-05-12 | 2004-02-05 | Dawson Peter S. | Dynamic software code instrumentation method and system |
US6675374B2 (en) | 1999-10-12 | 2004-01-06 | Hewlett-Packard Development Company, L.P. | Insertion of prefetch instructions into computer program code |
US6560693B1 (en) | 1999-12-10 | 2003-05-06 | International Business Machines Corporation | Branch history guided instruction/data prefetching |
US6848029B2 (en) | 2000-01-03 | 2005-01-25 | Dirk Coldewey | Method and apparatus for prefetching recursive data structures |
US6405292B1 (en) * | 2000-01-04 | 2002-06-11 | International Business Machines Corp. | Split pending buffer with concurrent access of requests and responses to fully associative and indexed components |
US6658652B1 (en) | 2000-06-08 | 2003-12-02 | International Business Machines Corporation | Method and system for shadow heap memory leak detection and other heap analysis in an object-oriented environment during real-time trace processing |
US6704860B1 (en) | 2000-07-26 | 2004-03-09 | International Business Machines Corporation | Data processing system and method for fetching instruction blocks in response to a detected block sequence |
US6571318B1 (en) | 2001-03-02 | 2003-05-27 | Advanced Micro Devices, Inc. | Stride based prefetcher with confidence counter and dynamic prefetch-ahead mechanism |
US20040015930A1 (en) | 2001-03-26 | 2004-01-22 | Youfeng Wu | Method and system for collaborative profiling for continuous detection of profile phase transitions |
US7032217B2 (en) | 2001-03-26 | 2006-04-18 | Intel Corporation | Method and system for collaborative profiling for continuous detection of profile phase transitions |
US20020144245A1 (en) | 2001-03-30 | 2002-10-03 | Guei-Yuan Lueh | Static compilation of instrumentation code for debugging support |
US20040015897A1 (en) | 2001-05-15 | 2004-01-22 | Thompson Carlos L. | Method and apparatus for verifying invariant properties of data structures at run-time |
US20030163658A1 (en) * | 2002-02-28 | 2003-08-28 | Lanzatella Thomas W. | System and method for controlling the creation of stable disk images |
US6951015B2 (en) | 2002-05-30 | 2005-09-27 | Hewlett-Packard Development Company, L.P. | Prefetch insertion by correlation of cache misses and previously executed instructions |
US20040103408A1 (en) | 2002-11-25 | 2004-05-27 | Microsoft Corporation | Dynamic prefetching of hot data streams |
US20040103401A1 (en) | 2002-11-25 | 2004-05-27 | Microsoft Corporation | Dynamic temporal optimization framework |
US20040111444A1 (en) | 2002-12-06 | 2004-06-10 | Garthwaite Alexander T. | Advancing cars in trains managed by a collector based on the train algorithm |
US20040133556A1 (en) * | 2003-01-02 | 2004-07-08 | Wolczko Mario I. | Method and apparatus for skewing a bi-directional object layout to improve cache performance |
US20050091645A1 (en) | 2003-10-24 | 2005-04-28 | Microsoft Corporation | Adaptive instrumentation runtime monitoring and analysis |
US20050246696A1 (en) | 2004-04-29 | 2005-11-03 | International Business Machines Corporation | Method and apparatus for hardware awareness of data types |
Non-Patent Citations (99)
Title |
---|
"JProfiler Manual," ejtechnologies, GmbH (2004), pp. 1-141. |
"Technologies for Measuring Software Performance", Intel Corporation (2003), 4 pages. |
Ammons et al., "Exploiting Hardware Performance Counters with Flow and Context Sensitive Profiling", PLDI (Jun. 1997), 12 pages. |
Ammons et al., "Improving Data-Flow Analysis with Path Profiles," SIGPLAN '98 (1998), pp. 72-84. |
Anderson et al., "Continuous Profiling: Where Have All the Cycles Gone?", ACM (Nov. 1997), pp. 357-390. |
Annavaram et al., "Data Prefetching by Dependence Graph Precomputation", IEEE (2001), pp. 52-61. |
Arnold et al., "A Framework for Reducing the Cost of Instrumented Code", Rutgers University Technical Report DCS-TR-424 (Nov. 2000), pp. 1-10. |
Bala et al., "Dynamo: A Transparent Dynamic Optimization System", ACM (2000), pp. 1-12. |
Balakrishnan et al., "Analyzing Memory Accesses in chi86 Binary Executables", Proc. 13th Intl. Conference on Compiler Construction, LNCS 2985 (Apr. 2004), pp. 5-23. |
Ball et al., "Efficient Path Profiling", IEEE (1996), pp. 46-57. |
Berger et al., "Composing High-Performance Memory Allocators," ACM, 11 pp. (2001). |
Bush et al., "A Static Analyzer for Finding Dynamic Programming Errors", Software: Practice and Experience (2000), pp. 775-802. |
Cahoon et al., "Data Flow Analysis for Software Prefetching Linked Data Structures in Java", IEEE (2001), 12 pages. |
Calder et al., "Cache-Conscious Data Placement," ACM, 11 pp. (1998). |
Chen et al., "Reducing Memory Latency via Non-Blocking and Prefetching Caches", ACM (1992), pp. 51-61. |
Chilimbi et al., "Cache-Conscious Structure Definition," Proc. ACM SIGPLAN '99 Conf. on Programming Language Design and Impl., 12 pp. (May 1999). |
Chilimbi et al., "Cache-Conscious Structure Layout," Proc. ACM SIGPLAN '99 Conf. on Programming Language Design and Impl., 12 pp. (May 1999). |
Chilimbi et al., "Dynamic Hot Data Stream Prefetching for General-Purpose Programs", ACM (2002), 11 pages. |
Chilimbi et al., "Low-Overhead Memory Leak Detection Using Adaptive Statistical Profiling", Proc. 11th Intl. Conference on Architectural Support for Programming Language and Operating Systems (ASPLOS), (Oct. 2004), pp. 156-164. |
Chilimbi et al., "Using Generational Garbage Collection to Implement Cache-Conscious Data Placement", Proceedings of the First International Symposium on Memory Management (Oct. 1998), vol. 34(3), pp. 37-48. |
Chilimbi, "Efficient Representations and Abstractions for Quantifying and Exploiting Data Reference Locality," ACM SIGPLAN '01 Conf. on Programming Language Design and Implementation, 12 pp. (Jun. 2001). |
Chilimbi, "On the Stability of Temporal Data Reference Profiles," Int'l Conf. on Parallel Architectures and Compilation Techniques, 10 pp. (Sep. 2001). |
Cooksey et al., "A Stateless, Content-Directed Data Prefetching Mechanism", ACM (2002), pp. 279-290. |
Cowan et al., "Buffer Overflows: Attacks and Defenses for the Vulnerability of the Decade", DARPA information survivability conference and expo (DISCEX) (2000), pp. 1-11. |
Crescenzi et al., "A Compendium of NP Optimization Problems," 20 pp. [Downloaded from the World Wide Web on Dec. 12, 2003]. |
Dean et al., "ProfileMe: Hardware Support for Instruction-Level Profiling on Out-of-Order Processors", Proc. 30th Annual Intl. Symp. On Microarchitecture (Dec. 1997), 12 pages. |
Deaver et al., "Wiggins/Restone: An On-line Program Specializer", Proceedings of the IEEE Hot Chips XI Conference (Aug. 1999), 29 pages. |
Demsky et al., "Automatic Detection and Repair of Errors in Data Structures", Proceedings of 18th ACM SIGPLAN Conference on Object-Oriented Programming Systems, Language and Applications (OOPLSA) (Oct. 2003), pp. 78-95. |
Demsky et al., "Role-Based Exploration of Object-Oriented Programs", Proceedings of 24th International Conference on Software Engineering (ISCE) (May 2002), pp. 313-334. |
Dor et al., "Cleanness Checking of String Manipulations in C Programs via Integer Analysis", Proc. 8th Int'l Static Analysis Symposium (Jun. 2001), 19 pages. |
Dor et al., "CSSV: Towards a Realistic Tool for Statically Detecting All Buffer Overflows in C", PLDI'03 (Jun. 9-11, 2003), pp. 155-167. |
Duesterwald et al., "Software profiling for hot path prediction: Less is more", Ninth International Conference on Architectural Support for Programming Languages and Operating Systems (Nov. 2000), pp. 202-211. |
Edwards, "Black-Box Testing Using Flowgraphs: An Experimental Assessment of Effectiveness and Automation Potential", Software Testing, Verification and Reliability, vol. 10, No. 4 (Dec. 2000), pp. 249-262. |
Eeckhout et al., "Workload Design: Selecting Representative Program-Input Pairs", Proceedings of the 2002 International Conference on Parallel Architectures and Compilation Techniques (2002), 12 pages. |
English et al., "Loge: a self-organizing disk controller," Proc. USENIX Winter 1992 Tech. Conf., San Francisco, pp. 237-251 (Jan. 1992). |
Ernst, "Dynamically Discovering Likely Program Invariants", PhD Thesis, University of Washington (Aug. 2000), pp. 1-127. |
Evans et al., "Improving Security Using Extensible Lightweight Static Analysis", IEEE Software (Jan./Feb. 2002), pp. 42-51. |
Evans et al., "LCLint: A Tool for Using Specifications to Check Code", SIGSOFT Symposium on the Foundations of Software Engineering (Dec. 1994), 10 pages. |
Evans et al., "Splint Manual, Version 3.1.1-1", Secure Programming Group, University of Virginia Department of Computer Science (Jun. 5, 2003), 121 pages. |
Evans, "Static Detection of Dynamic Memory Errors", SIGPLAN Conf. on Programming Language and Design Implementation (May 1996), 10 pages. |
Evans, "Using Specifications to Check Source Code", TR-628, MIT Lab for Computer Science (Jun. 1994), 97 pages. |
Foster et al., "A Theory of Type Qualifiers", Proc. ACM SIGPLAN '99 Conf. on Programming Language and Design Implementation (PLDI) (May 1999), 12 pages. |
Gloy et al., "Procedure Placement Using Temporal-Ordering Information," ACM Transactions on Programming Languages and System, vol. 21, pp. 111-161 (1999). |
Griffioen et al., "Reducing File System Latency Using a Predictive Approach," 11 pp., no date. |
Guyer et al., "An Annotation Language for Optimizing Software Libraries", Proc. Second Conf. on Domain Specific Languages (Oct. 1999), 14 pages. |
Halldorsson, "Approximations of Weighted Independent Set and Hereditary Subset Problems," JGAA, vol. 4, No. 1, pp. 1-16 (Apr. 2000). |
Hangal et al., "Tracking Down Software Bugs Using Automatic Anomaly Detection", Proceedings of 22nd International Conference on Software Engineering (ICSE) (Jan. 1992), pp. 125-136. |
Harris, "Dynamic Adaptive Pre-tenuring", Proceedings of the International Symposium on Memory Management (Oct. 2000), 9 pages. |
Hastings et al., "Purify: Fast Detection of Memory Leaks and Access Errors", Proceedings of the Winter 1992 USENIX Conference (1992), 10 pages. |
Hatfield et al., "Program Restructuring for Virtual Memory," IBM Sys. J., No. 3, pp. 168-192 (1971). |
Heil et al., "Relational Profiling: Enabling Thread-Level Parallelism in Virtual Machines", Proc. 33rd International Symposium on Microarchitecture (Dec. 2000), pp. 1-10. |
Hirzel et al., "Bursty Tracing: A Framework for Low-Overhead Temporal Profiling," Microsoft Research, 10 pp. (2001). |
Hirzel et al., "Understanding the Connectivity of Heap Objects", Proceedings of International Symposium on Memory Management (ISMM) (Jun. 2002), pp. 143-156. |
Hollingsworth et al., "MDL: A Language and Compiler for Dynamic Program Instrumentation", Proc. Of the International Conference on Parallel Architectures and Compilations Techniques (Nov. 1997), 12 pages. |
Hölzle et al., "Reconciling Responsiveness with Performance in Purse Object-Oriented Languages", ACM Transactions on Programming Languages and Systems (Jul. 1996), pp. 1-40. |
Horning, "The Larch Shared Language: Some Open Problems", Compass/ADT Workshop (Sep. 1995), 16 pages. |
Joseph et al., "Prefetching Using Markov Predictors", ACM (1997), pp. 252-263. |
Jouppi, "Improving Direct-Mapped Cache Performance by the Addition of a Small Fully-Associative Cache and Prefetch Buffers", IEEE (1990), pp. 364-373. |
Karlsson et al., "A Prefetching Technique for Irregular Accesses to Linked Data Structures", Sixth International Symposium on High-Performance Computer Architecture (Jan. 2000), 12 pages. |
Khurshid et al., "An Analyzable Annotation Language," OOPSLA '02 (Nov. 2002), 15 pages. |
Kistler et al., "Automated Data-Member Layout of Heap Objects to Improve Memory-Hierarchy Performance," ACM Transactions on Programming Language and Systems, 16 pp. (2000). |
Klaiber et al., "An Architecture for Software-Controlled Data Prefetching", ACM (1991), pp. 43-53. |
Kramer, "Examples of Design by Contract in Java Using Contract, the Design by Contract(TM) Tool for Java(TM)", Object World Berlin '99, Design & Components (May 17-20, 1999), 26 pages. |
Kroeger et al., "Predicting Future File-system Actions from Prior Events," Proc. USENIX Annual Tech. Conf., San Diego, pp. 1-10 (Jan. 1996). |
Larochelle et al., "Statistically Detecting Likely Buffer Overflow Vulnerabilities", 2001 USENIX Security Symposium (Aug. 2001), 5 pages. |
Larus, "Whole Program Paths," SIGPLAN '99 Conference on Programming Languages and Design (1999), 11 pages. |
Leavens et al., "Enhancing the Pre-and Postcondition Technique for More Expressive Specifications", Proc. World Congress on Formal Methods in the Development of Computing Systems (Sep. 1999), 21 pages. |
Leavens et al., "Preliminary Design of JML", Technical Report 98-06v, Iowa State University Department of Computer Science (Jun. 1998-2003; revised May 2003), 94 pages. |
Leino, "Checking Correctness Properties of Object-Oriented Programs," Internet, http://research.microsoft.com/leino/paper/1 (Aug. 19, 2002), 49 pages. |
Luk et al., "Compiler-Based Prefetching for Recursive Data Structures", ACM (1996), pp. 222-233. |
Melski et al., "Interprocedural Path Profiling", University of Wisconsin (1998), pp. 1-50. |
Microsoft Corporation, "Scalable Program Analysis", Internet, http://research.microsoft.com/spa/ (downloaded on Sep. 5, 2003), 3 pages. |
Mowry et al., "Design and Evaluation of a Compiler Algorithm for Prefetching", ACM (1992), pp. 62-73. |
Mowry et al., "Predicting Data Cache Misses in Non-Numeric Applications Through Correlation Profiling", International Symposium on Microarchitecture (1997), 7 pages. |
Nevill-Manning et al., "Identifying Hierarchical Structure in Sequences: A Linear-time Algorithm", Journal of Artificial Intelligence Research (1997), 7:67-82. |
Palmer et al., "Fido: A Cache that Learns to Fetch," Proc. 17th Int'l Conf. on Very Large Databases, Barcelona, pp. 255-264 (Sep. 1991). |
Patterson et al., "Informed Prefetching and Caching," Proc. 15th ACM Symp. on Operating System Principles, Copper Mountain Resort, CO, pp. 79-95 (Dec. 1995). |
Petrank et al., "The Hardness of Cache Conscious Data Placement," 29<SUP>th </SUP>Annual ACM Symposium on Principles of Programming Languages, 33 pp. (2002). |
Roth et al., "Dependence Based Prefetching for Linked Data Structures", Proceedings of the 8th International Conference on Architectural Support (Oct. 1998), pp. 115-126. |
Roth et al., "Effective Jump-Pointer Prefetching for Linked Data Structures", IEEE (1999), pp. 111-121. |
Rubin et al., An Efficient Profile-Analysis Framework for Data-Layout Optimizations, POPL, 14 pp. (Jan. 2002). |
Saavedra et al., "Improving the Effectiveness of Software Prefetching With Adaptive Execution", IEEE (1996), pp. 68-78. |
Sastry et al., "Rapid Profiling Via Stratified Sampling", International Symposium on Computer Architecture (2001), pp. 1-12. |
Savage et al., "Eraser: a dynamic data race detector for multighreaded programs", ACM Transactions on Computer Systems (TOCS) (1997), 391-411. |
Seidl et al., "Segregating Heap Objects by Reference Behavior and Lifetime," Eighth Int'l Conf. on Architectural Support for Programming Languages and Operating Systems, 12 pp. (Oct. 1998). |
Shaham, "Automatic Removal of Array Memory Leaks in Java," 54 pp. (1999). |
Srivastava et al., "Vulcan Binary Transformation in a Distributed Environment", Microsoft Research, MSR-TR-99-76 (Apr. 20, 2001), pp. 1-12. |
Srivastava, "ATOM: A System for Building Customized Program Analysis Tools," SIGPLAN '94 Conf. on Programming Language Design and Impl., 25 pp. (1994). |
Staelin et al., "Smart Filesystems," Proc. USENIX-Winter '91, Dallas, TX, pp. 45-51 (1991). |
Stoutchinin et al., "Speculative Prefetching of Induction Pointers", Proceedings of 10th International Conference on Compiler Construction (2001), 15 pages. |
Traub et al., "Ephemeral instrumentation for lightweight program profiling", Technical report, Harvard University (2000), pp. 1-13. |
Truong et al., "Improving Cache Behavior of Dynamically Allocated Data Structures," PACT, 8 pp. (1998). |
U.S. Appl. No. 11/115,924, filed Apr. 26, 2005, Chilimbi et al. |
U.S. Appl. No. 11/134,796, filed May 20, 2005, Shankar et al. |
U.S. Appl. No. 11/134,812, filed May 20, 2005, Ganapathy et al. |
Vanderwiel et al., "Data Prefetch Mechanisms", ACM Computing Surveys, vol. 32, No. 2 (Jun. 2000), pp. 174-199. |
Wahls et al., "The Direct Execution of SPECS-C++: A Model-Based Specification Language for C++ Classes", Technical Report TR94-02b, Iowa State University Department of Computer Science (Nov. 18, 1994), 52 pages. |
Zhou et al., "AccMon: Automatically Detecting Memory-Related Bugs via Program Counter-Based Invariants," Proceedings of 37th International Symposium on Micro-Architecture (MICRO) (Dec. 2004). |
Zilles et al., "A Programmable Co-processor for Profiling", Proceedings of the 7<SUP>th </SUP>International Symposium on High Performance Architecture (Jan. 2001), 12 pages. |
Cited By (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060070040A1 (en) * | 2002-11-25 | 2006-03-30 | Microsoft Corporation | Dynamic prefetching of hot data streams |
US8046752B2 (en) | 2002-11-25 | 2011-10-25 | Microsoft Corporation | Dynamic prefetching of hot data streams |
US20060265694A1 (en) * | 2005-05-20 | 2006-11-23 | Microsoft Corporation | Heap-based bug identification using anomaly detection |
US7770153B2 (en) | 2005-05-20 | 2010-08-03 | Microsoft Corporation | Heap-based bug identification using anomaly detection |
US7912877B2 (en) | 2005-05-20 | 2011-03-22 | Microsoft Corporation | Leveraging garbage collection to dynamically infer heap invariants |
US7962901B2 (en) | 2006-04-17 | 2011-06-14 | Microsoft Corporation | Using dynamic analysis to improve model checking |
US20070250820A1 (en) * | 2006-04-20 | 2007-10-25 | Microsoft Corporation | Instruction level execution analysis for debugging software |
US20070288908A1 (en) * | 2006-06-09 | 2007-12-13 | International Business Machines Corporation | Computer implemented method and system for accurate, efficient and adaptive calling context profiling |
US20080288926A1 (en) * | 2006-06-09 | 2008-11-20 | International Business Machine Corporation | Computer Implemented Method and System for Accurate, Efficient and Adaptive Calling Context Profiling |
US8122438B2 (en) | 2006-06-09 | 2012-02-21 | International Business Machines Corporation | Computer implemented method and system for accurate, efficient and adaptive calling context profiling |
US7818722B2 (en) * | 2006-06-09 | 2010-10-19 | International Business Machines Corporation | Computer implemented method and system for accurate, efficient and adaptive calling context profiling |
US7926043B2 (en) * | 2006-06-20 | 2011-04-12 | Microsoft Corporation | Data structure path profiling |
US20080005208A1 (en) * | 2006-06-20 | 2008-01-03 | Microsoft Corporation | Data structure path profiling |
US7861231B2 (en) * | 2007-03-09 | 2010-12-28 | International Business Machines Corporation | System and method of identification of dangling pointers |
US20080222224A1 (en) * | 2007-03-09 | 2008-09-11 | Kumar Rangarajan | System and Method of Identification of Dangling Pointers |
US7650464B2 (en) * | 2007-03-26 | 2010-01-19 | Intel Corporation | Object relocation guided by data cache miss profile |
US20080243300A1 (en) * | 2007-03-26 | 2008-10-02 | Jack Liu | Object relocation guided by data cache miss profile |
US20100115494A1 (en) * | 2008-11-03 | 2010-05-06 | Gorton Jr Richard C | System for dynamic program profiling |
US8875115B2 (en) * | 2008-11-29 | 2014-10-28 | International Business Machines Corporation | Type merging technique to reduce class loading during Java verification |
US20100138819A1 (en) * | 2008-11-29 | 2010-06-03 | International Business Machines Corporation | Type merging technique to reduce class loading during java verfication |
US8566813B2 (en) * | 2009-03-03 | 2013-10-22 | Samsung Electronics Co., Ltd. | Method and system generating execution file system device |
US20100229164A1 (en) * | 2009-03-03 | 2010-09-09 | Samsung Electronics Co., Ltd. | Method and system generating execution file system device |
US8612952B2 (en) | 2010-04-07 | 2013-12-17 | International Business Machines Corporation | Performance optimization based on data accesses during critical sections |
US20150082285A1 (en) * | 2013-03-15 | 2015-03-19 | Concurix Corporation | Runtime settings derived from relationships identified in tracer data |
US9864676B2 (en) | 2013-03-15 | 2018-01-09 | Microsoft Technology Licensing, Llc | Bottleneck detector application programming interface |
US10318415B2 (en) | 2017-05-31 | 2019-06-11 | International Business Machines Corporation | Garbage collection facility grouping infrequently accessed data units in designated transient memory area |
US10339049B2 (en) | 2017-05-31 | 2019-07-02 | International Business Machines Corporation | Garbage collection facility grouping infrequently accessed data units in designated transient memory area |
US20200142677A1 (en) * | 2018-11-05 | 2020-05-07 | International Business Machines Corporation | Fields Hotness Based Object Splitting |
US10747515B2 (en) * | 2018-11-05 | 2020-08-18 | International Business Machines Corporation | Fields hotness based object splitting |
Also Published As
Publication number | Publication date |
---|---|
US20040215880A1 (en) | 2004-10-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7343598B2 (en) | Cache-conscious coallocation of hot data streams | |
Calder et al. | Cache-conscious data placement | |
Maas et al. | Learning-based memory allocation for C++ server workloads | |
Rubin et al. | An efficient profile-analysis framework for data-layout optimizations | |
US5815720A (en) | Use of dynamic translation to collect and exploit run-time information in an optimizing compilation system | |
US7367024B2 (en) | Compiler-driven dynamic memory allocation methodology for scratch-pad based embedded systems | |
Chilimbi et al. | Dynamic hot data stream prefetching for general-purpose programs | |
Kistler et al. | Automated data-member layout of heap objects to improve memory-hierarchy performance | |
Blackburn et al. | Barriers: Friend or foe? | |
Clifford et al. | Memento mori: Dynamic allocation-site-based optimizations | |
Chen et al. | Profile-guided proactive garbage collection for locality optimization | |
Lee et al. | Feedback directed optimization of tcmalloc | |
Chilimbi et al. | Cache-conscious coallocation of hot data streams | |
US8359435B2 (en) | Optimization of software instruction cache by line re-ordering | |
Inagaki et al. | Stride prefetching by dynamically inspecting objects | |
Curial et al. | Mpads: memory-pooling-assisted data splitting | |
Wang et al. | On improving heap memory layout by dynamic pool allocation | |
US9146719B2 (en) | Data layout using data type information | |
Effler et al. | Evaluating the effectiveness of program data features for guiding memory management | |
Inoue et al. | Identifying the sources of cache misses in Java programs without relying on hardware counters | |
Duck et al. | An extended low fat allocator API and applications | |
Ye et al. | A unifying abstraction for data structure splicing | |
Nguyen et al. | Scratch-pad memory allocation without compiler support for java applications | |
Wu et al. | Reducing trace selection footprint for large-scale java applications without performance loss | |
Chakraborty et al. | Integrating software caches with scratch pad memory |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHILIMBI, TRISHUL;SHAHAM, RAN;REEL/FRAME:014810/0383;SIGNING DATES FROM 20031212 TO 20031214 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034541/0477 Effective date: 20141014 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20200311 |