US9141525B2 - Adaptive prestaging in a storage controller - Google Patents

Adaptive prestaging in a storage controller Download PDF

Info

Publication number
US9141525B2
US9141525B2 US13/018,305 US201113018305A US9141525B2 US 9141525 B2 US9141525 B2 US 9141525B2 US 201113018305 A US201113018305 A US 201113018305A US 9141525 B2 US9141525 B2 US 9141525B2
Authority
US
United States
Prior art keywords
prestage
value
cache
trigger
read data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/018,305
Other versions
US20120198148A1 (en
Inventor
Michael T. Benhase
Nedlaya Y. Francisco
Binny S. Gill
Lokesh M. Gupta
Suguang Li
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US13/018,305 priority Critical patent/US9141525B2/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BENHASE, MICHAEL T., GILL, BINNY S., FRANCISCO, NEDLAYA Y., GUPTA, LOKESH M., LI, SUGUANG
Publication of US20120198148A1 publication Critical patent/US20120198148A1/en
Priority to US13/786,981 priority patent/US8874840B2/en
Application granted granted Critical
Publication of US9141525B2 publication Critical patent/US9141525B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/023Free address space management
    • G06F12/0238Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory
    • G06F12/0246Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory in block erasable memory, e.g. flash memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0862Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches with prefetch
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0866Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches for peripheral storage systems, e.g. disk cache
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0866Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches for peripheral storage systems, e.g. disk cache
    • G06F12/0868Data transfer between cache memory and other subsystems, e.g. storage devices or host systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/50Control mechanisms for virtual memory, cache or TLB
    • G06F2212/502Control mechanisms for virtual memory, cache or TLB using adaptive policy

Definitions

  • the present description relates to a method, system, and program product for prestaging data into cache from a storage system in preparation for data transfer operations.
  • a storage controller manages data transfer operations between a direct access storage device (DASD), which may be a string of hard disk drives or other non-volatile storage devices and host systems and their application programs.
  • DASD direct access storage device
  • DAR data access request
  • the storage controller physically accesses the data stored in tracks in the DASD.
  • a DAR is a set of contiguous data sets, such as tracks, records, fixed blocks, or any other grouping of data.
  • the process of physically rotating the disk in the DASD to the requested track then physically moving the reading unit to the disk section to read data is often a time consuming process. For this reason, current systems frequently stage data into a cache memory of the storage controller in advance of the host requests for such data.
  • storage systems may also include solid state storage devices such as flash memory.
  • flash memory tends to be substantially more expensive than DASD storage.
  • a storage controller typically includes a large buffer managed as a cache to buffer data accessed from the attached storage.
  • DARs data access requests
  • the storage controller Prior to receiving the actual DAR, the storage controller receives information on a sequence of tracks in the DASD involved in the upcoming read operation or that data is being accessed sequentially. The storage controller will then proceed to stage the sequential tracks into the cache. The storage controller would then process DARs by accessing the staged data in the cache. In this manner, the storage controller can return cached data to a read request at the data transfer speed in the storage controller channels as opposed to non-cached data which is transferred at the speed of the DASD device or other storage from which the data was transferred to the cache.
  • an application program such as a batch program, will process numerous data records stored at contiguous locations in the storage device. It is desirable during such sequential read operations to prestage the sequential data into cache in anticipation of the requests from the application program. As a result of prestaging, often one of the tracks being staged is already in cache.
  • Present techniques used to prestage sequential blocks of data include sequential caching algorithms systems, such as those described in the commonly assigned patent entitled “CACHE DASD Sequential Staging and Method,” having U.S. Pat. No. 5,426,761.
  • a sequential caching algorithm detects when a device is requesting data as part of a sequential access operation. Upon making such a detection, the storage controller will begin prestaging sequential data records following the last requested data record into cache in anticipation of future sequential accesses.
  • U.S. Pat. No. 5,426,761 discloses an algorithm for sequential read operations to prestage multiple tracks into cache when the tracks are in the extent range of the sequential read operation, the track is not already in the cache, and a maximum number of tracks have not been prestaged into cache.
  • the cached records may then be returned to the application performing the sequential data operations at speeds substantially faster than retrieving the records from a non-volatile storage device.
  • Another prestaging technique includes specifying a block of contiguous data records to prestage into cache in anticipation of a sequential data request.
  • the Small Computer System Interface provides a prestage or prefetch command, PRE-FETCH, that specifies a logical block address where the prestaging operation begins and a transfer length of contiguous logical blocks of data to transfer to cache.
  • PRE-FETCH prestage or prefetch command
  • the SCSI PRE-FETCH command is described in the publication “Information Technology-Small Computer System Interface-2,” published by ANSI on Apr. 19, 1996, reference no. X3.131-199x, Revision 10L, which publication is incorporated herein by reference in its entirety.
  • AMP Adaptive Multi-stream Prefetching
  • This technique seeks to provide an algorithm which can adapt both the amount of data being prefetched in a prefetch operation, and the value of a trigger for triggering the prefetch operation, on a per stream basis in response to evolving workloads.
  • a host when a host reads tracks sequentially, it benefits greatly to pre-stage tracks so that the tracks are already in cache before the host requests access to those tracks.
  • a storage controller has a limited Cache and limited bandwidth. If pre-stage stages too few tracks then the host will get more cache misses since fewer tracks will be in cache. Conversely, if the host pre-stages too many tracks then it can starve other kinds of I/O in the storage controller.
  • Tracks staged into cache may be demoted (dropped from cache) according to a Least Recently Used (LRU) algorithm to insure that the staged data in cache does not exceed a predetermined threshold.
  • LRU Least Recently Used
  • the LRU algorithm discards the “least recently used” items first, which are generally the items which have been in the cache the longest without being used.
  • the data when first placed into cache may be designated (“time stamped”) MRU (“most recently used”) data. As the data remains unused in cache, the timestamp for that data is updated. The longer the data stays in cache unused, the closer the data moves toward an LRU timestamp.
  • the LRU algorithm assumes that the oldest unused data is the data least likely to be used and therefore may be discarded from the cache to make room in the cache for additional data which is more likely to be accessed.
  • LRU data may be redesignated with an MRU timestamp to give that data another chance to be used before being demoted, that is, discarded from cache, should the timestamp for that data again arrive at LRU status.
  • At least one of the value of a prestage trigger and the value of a prestage amount may be modified as a function of the drive speed of the storage drive from which the units of read data are prestaged into a cache memory.
  • the operations include reading units of data of a sequential stream of read data stored in at least one of a first storage drive having a drive speed and a second storage drive having a drive speed.
  • a prestage trigger is initialized to initiate prestaging of units of read data.
  • units of data of the sequential stream of read data are prestaged from a storage drive into a cache memory in anticipation of a read request.
  • At least one of the value of the prestage trigger and the value of the prestage amount may be modified as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory.
  • the drive speed of the first storage drive may be faster than that of the second storage drive.
  • the drive speeds of various storage drives of the system may be the same, depending upon the particular application.
  • At least one of the value of the prestage trigger and the value of the prestage amount may be reduced as a function of the drive speed in association with the demoting of a unit of data from the cache.
  • the prestaging operations may further include comparing the modified value of the prestage trigger to the amount of data to be prestaged in response to the trigger; and prestaging additional units of data of the sequential stream of read data into the cache memory as a function of the comparison.
  • the prestaging operations may further include staging units of data of the sequential stream of read data into the cache memory, and determining if a cache miss occurs because a staging has not completed.
  • the value of the prestage trigger may be modified by increasing the value of the prestage trigger as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory.
  • the value of the prestage trigger may be modified by increasing the value of the prestage trigger by a first value if the read data being prestaged is from a hard disk drive, and may be modified by increasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from a solid state drive.
  • the prestaging operations may further include determining if a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache, wherein the value of the prestage amount may be modified by decreasing the value of the prestage amount as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, in response to a determination that a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache.
  • the value of the prestage amount may be decreased by decreasing the value of the prestage trigger by a first value if the read data being prestaged is from a hard disk drive and by decreasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from a solid state storage drive.
  • FIG. 1 illustrates a trigger distance in a stream of data for triggering the prestaging of data into a cache.
  • FIG. 2 shows an example of cache operations in accordance with a prior art Adaptive Multi-stream Prefetching (AMP) algorithm.
  • AMP Adaptive Multi-stream Prefetching
  • FIG. 3 illustrates an embodiment of a computing environment in which an aspect of the present description may be employed.
  • FIG. 4 shows an example of operations of a cache in connection with prestaging tracks into the cache in accordance with one embodiment of the present description.
  • FIG. 5 shows a more detailed example of operations for prestage trigger and amount modification and other features in accordance with one embodiment of the present description.
  • FIG. 6 illustrates an example of prior art cache operations in which tracks may become candidates for demotion.
  • FIG. 1 shows a set 100 of a stream of sequential read data which has been prefetched or prestaged into a cache.
  • the set 100 may be a set of sequential pages, tracks or other units 110 of sequential read data.
  • read data is prestaged in a prestageAmount (p) of tracks.
  • a first portion 120 of the set 100 is indicated as being accessed by a storage controller or host in connection with a read operation.
  • the remaining portion 130 of the set 100 is indicated as being unaccessed, that is, having not yet been accessed.
  • a trigger track 140 is identified at a trigger distance (g) from the end 150 of the unaccessed tracks 130 of the prestaged set 100 . Once the tracks being accessed reaches the trigger track 140 , another set 100 of the next p sequential tracks of the stream of sequential read data is prefetched or prestaged into the cache. Once the tracks being accessed reaches the trigger track 140 of that next set 100 , yet another set 100 of the next p sequential tracks of the stream of sequential read data is prefetched or prestaged into the cache, and so on.
  • a prior art algorithm called AMP can adapt both the amount (p) of data being prestaged in a prestage operation, and the value of the trigger (g) for triggering the prestage operation, on a per stream basis in response to changing workloads.
  • the prestageTrigger g+1 r*t(p), 0 ⁇ g ⁇ p where t(p) is the average time to fetch p consecutive tracks. It is believed that when the prestageTrigger g ⁇ r*t(p), a cache miss is likely to happen.
  • FIG. 2 shows an example of cache operations in accordance with a prior art AMP algorithm, for determining both the number p of tracks to prestage and the value of the trigger g which triggers a prestage operation.
  • the prestage amount (p) and the value of the prestage trigger (g) are initialized (block 202 ).
  • the prestage amount p may be initialized at 8 tracks and the value of the prestage trigger g may be initialized at distance of 3 tracks, for example.
  • the prestage amount p is increased (block 208 ) by 1. By increasing the prestage amount p, it is intended to decrease the chances of running out of prestaged data and causing a cache miss. However, if it is previously determined (block 210 ) that the prestage amount p has reached a maximum, such as a maximum of 84, for example, the prestage amount p is not increased.
  • the prestage trigger distance g is increased (block 214 ) by 1. By increasing the prestage trigger distance g, it is intended to trigger the prestaging of additional data sooner to reduce the chances of future cache misses. However, if it determined (block 216 ) that the prestage trigger g plus 2 (g+2) has exceeded the prestage amount p, the prestage trigger is reduced (block 218 ) to the prestage amount p minus 2 (p ⁇ 2).
  • FIG. 6 illustrates an example of cache operations in which tracks may become candidates for demotion.
  • a determination is made (block 600 ) as to whether additional space is needed in cache. If so, a determination is made (block 610 ) as to whether the LRU track is a candidate for demotion. If not, in this algorithm, LRU data is redesignated with an MRU timestamp (block 620 ) to give that data another chance to be used before being demoted, that is, discarded from cache, should the timestamp for that data again arrive at LRU status. LRU data redesignated with an MRU timestamp is referred to herein as a “candidate” for demotion.
  • a flag is also set (block 620 ) to mark the LRU data redesignated with an MRU timestamp as a “candidate” for demotion.
  • both the prestage amount (p) and the prestage trigger g are each decreased (block 222 ) by 1.
  • FIG. 3 illustrates an example of a hardware environment in which embodiments may be implemented.
  • a plurality of host systems 310 a, 310 b, 310 c are in data communication with a plurality of storage units such as a hard disk drive 312 such as a DASD, and a solid state storage 314 such as a Flash memory, via a storage controller 320 .
  • a hard disk drive 312 such as a DASD
  • solid state storage 314 such as a Flash memory
  • the storage controller further includes a cache 330 in which data is prestaged and staged.
  • the storage controller 320 can prestage data in such a manner that one or both of the prestage amount and the prestage trigger may be modified as a function of the speeds and bandwidths of the various storage devices coupled to the storage controller 320 .
  • the solid state storage 314 has a substantially faster access time as compared to that of the hard drive storage 312 .
  • the prestage amount or prestage trigger for a stream of data being read from the faster flash memory storage 314 may be modified a lower amount, for example, as compared to a modification of the prestage amount or prestage trigger for a stream of data being read from slower hard drive storage 312 .
  • Other features and advantages will also be explained in connection with the present description.
  • the host systems 310 a, 310 b, 310 c may be any host system known in the art, such as a mainframe computer, workstations, etc., including an operating system such as WINDOWS.®., AIX.®., UNIX.®., MVS.TM., etc.
  • AIX is a registered trademark of IBM
  • MVS is a trademark of IBM
  • WINDOWS is a registered trademark of Microsoft Corporation
  • UNIX is a registered trademark licensed by the X/Open Company LTD.
  • a plurality of channel paths 322 a, 322 b, 322 c in the host systems 310 a , 310 b, 310 c provide communication paths to the storage controller 320 .
  • the storage controller 320 and host systems 310 a, 310 b, 310 c may communicate via any network or communication system known in the art, such as LAN, TCP/IP, ESCON.®., SAN, SNA, Fibre Channel, SCSI, etc.
  • ESCON is a registered trademark of International Business Machines Corporation (“IBM”).
  • the host system 310 a, 310 b, 310 c executes commands and receives returned data along a selected channel 322 a, 322 b, 322 c.
  • the storage controller 320 issues commands via a path 324 a to physically position the electromechanical devices to read the storage 312 .
  • the storage controller 320 also issues commands via a path 324 b to read data from the solid state storage 314 .
  • the cache 330 is implemented in a high speed, volatile storage area within the storage controller 320 , such as a DRAM, RAM, etc.
  • the length of time since the last use of a record in cache 330 is maintained to determine the frequency of use of cache 330 .
  • the least recently used (LRU) data in cache 330 may be demoted according to LRU or other cache memory management techniques known in the art.
  • the cache 330 may be implemented in alternative storage areas accessible to the storage controller 320 .
  • Data can be transferred between the channels 322 a, 322 b, 322 c and the cache 330 , between the channels 322 a, 322 b, 322 c and the storage 312 , 314 and between the storage 312 , 314 and the cache 330 .
  • data retrieved from the storage 312 , 314 in response to a read miss can be concurrently transferred to both the channel 322 a, 322 b, 322 c and the cache 330 and a data write can be concurrently transferred from the channel 322 a, 322 b , 322 c to both a non-volatile storage unit and cache 330 .
  • the number of hosts, storage controllers, caches and storage drives may vary, depending upon the particular application.
  • FIG. 4 shows one embodiment of operations of the cache 330 in connection with prestaging tracks into the cache 330 .
  • a track is read (block 400 ).
  • a determination is made as to whether (block 404 ) the track being read is on a sequential read stream. If so, a determination is made as to whether (block 408 ) the track being read is a trigger track. If the track being read is a trigger track, the next group of tracks is prestaged (block 412 ) into cache.
  • FIG. 5 discussed below shows an example of such a determination and an example of such a prestage trigger or amount modification as a function of different drives speeds.
  • FIG. 5 discussed below shows an example of such a determination.
  • FIG. 5 shows a more detailed example of certain operations for prestage trigger and amount modification and other features.
  • g is the prestage trigger
  • r is the request rate of the stream (tracks/sec)
  • L is the life of the shared cache (MRU timestamp ⁇ LRU timestamp)
  • t(N) is the average time to fetch N consecutive tracks
  • N is the stride size of the device (4*rank width).
  • the prestage amount (p) and the value of the prestage trigger (g) are initialized (block 500 ).
  • the prestage amount p may be initialized at 8 tracks and the value of the prestage trigger g may be initialized at distance of 3 tracks, for example.
  • the prestage trigger of this track is the larger one of 3 or the prestage trigger of the previous track. It is appreciated that other values may be selected for initialization of the prestage amount p and the prestage trigger g, depending upon the particular application
  • a track is read (block 504 ) and a determination is made as to whether (block 508 ) the track being read is on a sequential read stream. If so, another determination is made as to whether (block 510 ) the track being read was in connection with a cache miss. If so, a determination is made as to whether (block 512 ) a stage issued in connection with the track being read. If so, the prestage trigger is increased (block 516 ) as a function of different speeds of different drives and tracks are staged (block 524 ).
  • the prestage trigger of this track may be increased based on the speed of the drives. For example, if the track was read from the hard drive storage 312 , the prestage trigger may be increased by 1, for example. If the track was read from a faster drive such as the solid state storage 314 , the prestage trigger may be increased by a greater amount such as by 2, for example. Such differentiated increments may facilitate aiding the completion of the issued stages before a cache miss occurs. It is appreciated that other values may be selected for incrementing the prestage trigger g as a function of different drive speeds, depending upon the particular application.
  • the cache miss may have occurred as a result of that track having been demoted before being used. As a result, it may be appropriate to delay prestaging additional tracks or to reduce (block 530 ) the amount being prestaged.
  • both the prestage trigger and the prestage amount may be decreased (block 530 ) as a function of the different speeds of the different drives. For example, if the track was read from the hard drive storage 312 , the prestage trigger may be decreased by 1, for example, and the prestage amount may be reduced by 1, for example. If the track was read from a faster drive such as solid state storage 314 , the prestage trigger may be decreased by a greater amount such as by 2, for example, and the prestage amount may be reduced by 2, for example. Such differentiated decrements may facilitate reducing the number of candidate tracks for demotion. It is appreciated that other values may be selected for decrementing the prestage trigger g and the prestage amount p as a function of different drive speeds, depending upon the particular application. In addition, tracks are staged (block 524 ).
  • the track being read was a cache hit rather than a cache miss
  • another determination is made as to whether (block 540 ) the track being read is a prestage track candidate for demotion. If so, it may be appropriate to reduce the amount of data being prestaged to reduce the number of candidates for demotion. In this embodiment, the prestage amount is decreased (block 544 ) as a function of different speeds of different drives.
  • the prestage amount may be decreased by 1, for example. If the track was read from a faster drive such as solid state storage 314 , the prestage amount may be decreased by a greater amount such as by 2, for example.
  • Such differentiated decrements may facilitate reducing the number of candidate tracks for demotion. It is appreciated that other values may be selected for decrementing the prestage amount p as a function of different drive speeds, depending upon the particular application.
  • the prestage amount p may be increased (block 558 ) as a function of different speeds of different drives. For example, if the track was read from the hard drive storage 312 which is slower than the solid state drive 314 , the prestage amount may be increased by 1, for example. If the track was read from a faster drive such as solid state storage 314 , the prestage amount may be increased by a larger amount such as by 2, for example. Such differentiated increments may facilitate prestaging sufficient data to reduce cache misses. It is appreciated that other values may be selected for incrementing the prestage amount p as a function of different drive speeds, depending upon the particular application.
  • the prestage amount p By increasing the prestage amount p, it is intended to decrease the chances of running out of prestaged data and causing a cache miss. However, if it is previously determined (block 560 ) that the prestage amount p has reached a maximum, such as a maximum of 84, for example, the prestage amount p is not increased. It is appreciated that other values may be selected for a maximum, depending upon the particular application.
  • cache prestaging operations in accordance with one aspect of the present description may take into account storage devices of varying speeds and bandwidths for purposes of modifying one or both of a prestage trigger and the prestage amount.
  • Such a capability may be particularly suitable for storage controllers having a mix of solid state storage and hard drive storage, for example.
  • a cache prestaging operation in accordance with another aspect of the present description may decrease one or both of the prestage trigger and the prestage amount in circumstances such as a cache miss which may have resulted from prestaged tracks being demoted before they are used. Conversely, a cache prestaging operation in accordance with another aspect of the present description may increase one or both of the prestage trigger and the prestage amount in circumstances such as a cache miss which may have resulted from waiting for a stage to complete.
  • the prestage trigger may not be limited by the prestage amount p. Instead, the pre-stage trigger may be permitted to expand as conditions warrant it by prestaging additional tracks and thereby effectively increasing the potential range for the prestage trigger.
  • aspects of the embodiments may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.”
  • aspects of the embodiments may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • the system of FIG. 2 may be implemented as a cloud component part in a cloud computing environment.
  • the systems architecture of the hardware and software components involved in the delivery of cloud computing may comprise a plurality of cloud components communicating with each other over a network, such as the Internet.
  • the system of FIG. 2 may provide clients, and other servers and software and/or hardware components in the networked cloud, with scheduling services.
  • an embodiment means “one or more (but not all) embodiments of the present invention(s)” unless expressly specified otherwise.
  • Devices that are in communication with each other need not be in continuous communication with each other, unless expressly specified otherwise.
  • devices that are in communication with each other may communicate directly or indirectly through one or more intermediaries.
  • FIGS. 4 and 5 shows certain events occurring in a certain order. In alternative embodiments, certain operations may be performed in a different order, modified or removed. Moreover, steps may be added to the above described logic and still conform to the described embodiments. Further, operations described herein may occur sequentially or certain operations may be processed in parallel. Yet further, operations may be performed by a single processing unit or by distributed processing units.
  • a solid state storage 314 was described as having a substantially faster drive speed or access time as compared to that of a hard drive storage 312 .
  • features of the present description are also applicable or beneficial to additional embodiments in which the drive speeds of various drives of the system may be the same, as well, depending upon the particular application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Memory System Of A Hierarchy Structure (AREA)

Abstract

In one aspect of the present description, at least one of the value of a prestage trigger and the value of the prestage amount, may be modified as a function of the drive speed of the storage drive from which the units of read data are prestaged into a cache memory. Thus, cache prestaging operations in accordance with another aspect of the present description may take into account storage devices of varying speeds and bandwidths for purposes of modifying a prestage trigger and the prestage amount. Still further, a cache prestaging operation in accordance with further aspects may decrease one or both of the prestage trigger and the prestage amount as a function of the drive speed in circumstances such as a cache miss which may have resulted from prestaged tracks being demoted before they are used. Conversely, a cache prestaging operation in accordance with another aspect may increase one or both of the prestage trigger and the prestage amount as a function of the drive speed in circumstances such as a cache miss which may have resulted from waiting for a stage to complete. In yet another aspect, the prestage trigger may not be limited by the prestage amount. Instead, the pre-stage trigger may be permitted to expand as conditions warrant it by prestaging additional tracks and thereby effectively increasing the potential range for the prestage trigger. Other features and aspects may be realized, depending upon the particular application.

Description

BACKGROUND
1. Field
The present description relates to a method, system, and program product for prestaging data into cache from a storage system in preparation for data transfer operations.
2. Description of Related Art
In current storage systems, a storage controller manages data transfer operations between a direct access storage device (DASD), which may be a string of hard disk drives or other non-volatile storage devices and host systems and their application programs. To execute a read operation presented from a host system, i.e., a data access request (DAR), the storage controller physically accesses the data stored in tracks in the DASD. A DAR is a set of contiguous data sets, such as tracks, records, fixed blocks, or any other grouping of data. The process of physically rotating the disk in the DASD to the requested track then physically moving the reading unit to the disk section to read data is often a time consuming process. For this reason, current systems frequently stage data into a cache memory of the storage controller in advance of the host requests for such data.
In addition, storage systems may also include solid state storage devices such as flash memory. However, flash memory tends to be substantially more expensive than DASD storage.
A storage controller typically includes a large buffer managed as a cache to buffer data accessed from the attached storage. In this way, data access requests (DARs) can be serviced at electronic speeds directly from the cache thereby avoiding the delays caused by reading from storage such as DASD storage. Prior to receiving the actual DAR, the storage controller receives information on a sequence of tracks in the DASD involved in the upcoming read operation or that data is being accessed sequentially. The storage controller will then proceed to stage the sequential tracks into the cache. The storage controller would then process DARs by accessing the staged data in the cache. In this manner, the storage controller can return cached data to a read request at the data transfer speed in the storage controller channels as opposed to non-cached data which is transferred at the speed of the DASD device or other storage from which the data was transferred to the cache.
During a sequential read operation, an application program, such as a batch program, will process numerous data records stored at contiguous locations in the storage device. It is desirable during such sequential read operations to prestage the sequential data into cache in anticipation of the requests from the application program. As a result of prestaging, often one of the tracks being staged is already in cache. Present techniques used to prestage sequential blocks of data include sequential caching algorithms systems, such as those described in the commonly assigned patent entitled “CACHE DASD Sequential Staging and Method,” having U.S. Pat. No. 5,426,761. A sequential caching algorithm detects when a device is requesting data as part of a sequential access operation. Upon making such a detection, the storage controller will begin prestaging sequential data records following the last requested data record into cache in anticipation of future sequential accesses.
U.S. Pat. No. 5,426,761 discloses an algorithm for sequential read operations to prestage multiple tracks into cache when the tracks are in the extent range of the sequential read operation, the track is not already in the cache, and a maximum number of tracks have not been prestaged into cache. The cached records may then be returned to the application performing the sequential data operations at speeds substantially faster than retrieving the records from a non-volatile storage device.
Another prestaging technique includes specifying a block of contiguous data records to prestage into cache in anticipation of a sequential data request. For instance, the Small Computer System Interface (SCSI) provides a prestage or prefetch command, PRE-FETCH, that specifies a logical block address where the prestaging operation begins and a transfer length of contiguous logical blocks of data to transfer to cache. The SCSI PRE-FETCH command is described in the publication “Information Technology-Small Computer System Interface-2,” published by ANSI on Apr. 19, 1996, reference no. X3.131-199x, Revision 10L, which publication is incorporated herein by reference in its entirety.
Yet another prestaging technique is referred to as AMP (Adaptive Multi-stream Prefetching) in a shared cache. This technique seeks to provide an algorithm which can adapt both the amount of data being prefetched in a prefetch operation, and the value of a trigger for triggering the prefetch operation, on a per stream basis in response to evolving workloads.
As set forth above, when a host reads tracks sequentially, it benefits greatly to pre-stage tracks so that the tracks are already in cache before the host requests access to those tracks. However, a storage controller has a limited Cache and limited bandwidth. If pre-stage stages too few tracks then the host will get more cache misses since fewer tracks will be in cache. Conversely, if the host pre-stages too many tracks then it can starve other kinds of I/O in the storage controller.
Tracks staged into cache may be demoted (dropped from cache) according to a Least Recently Used (LRU) algorithm to insure that the staged data in cache does not exceed a predetermined threshold. The LRU algorithm discards the “least recently used” items first, which are generally the items which have been in the cache the longest without being used.
The data when first placed into cache may be designated (“time stamped”) MRU (“most recently used”) data. As the data remains unused in cache, the timestamp for that data is updated. The longer the data stays in cache unused, the closer the data moves toward an LRU timestamp. The LRU algorithm assumes that the oldest unused data is the data least likely to be used and therefore may be discarded from the cache to make room in the cache for additional data which is more likely to be accessed. In some LRU algorithms, LRU data may be redesignated with an MRU timestamp to give that data another chance to be used before being demoted, that is, discarded from cache, should the timestamp for that data again arrive at LRU status.
SUMMARY
In one aspect of the present description, at least one of the value of a prestage trigger and the value of a prestage amount may be modified as a function of the drive speed of the storage drive from which the units of read data are prestaged into a cache memory. In one embodiment, the operations include reading units of data of a sequential stream of read data stored in at least one of a first storage drive having a drive speed and a second storage drive having a drive speed. A prestage trigger is initialized to initiate prestaging of units of read data. In response to reaching the trigger as the units of the sequential stream of read data are read, units of data of the sequential stream of read data are prestaged from a storage drive into a cache memory in anticipation of a read request. In addition, at least one of the value of the prestage trigger and the value of the prestage amount may be modified as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory. In some embodiments, the drive speed of the first storage drive may be faster than that of the second storage drive. However, it is appreciated that in other embodiments, the drive speeds of various storage drives of the system may be the same, depending upon the particular application.
In another aspect, upon demoting a unit of data from the cache before the unit being demoted is read from cache in response to a read request, at least one of the value of the prestage trigger and the value of the prestage amount may be reduced as a function of the drive speed in association with the demoting of a unit of data from the cache.
In yet another aspect, the prestaging operations may further include comparing the modified value of the prestage trigger to the amount of data to be prestaged in response to the trigger; and prestaging additional units of data of the sequential stream of read data into the cache memory as a function of the comparison.
In still another aspect, the prestaging operations may further include staging units of data of the sequential stream of read data into the cache memory, and determining if a cache miss occurs because a staging has not completed. In one embodiment, the value of the prestage trigger may be modified by increasing the value of the prestage trigger as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory.
In yet another embodiment, the value of the prestage trigger may be modified by increasing the value of the prestage trigger by a first value if the read data being prestaged is from a hard disk drive, and may be modified by increasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from a solid state drive.
In still another aspect, the prestaging operations may further include determining if a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache, wherein the value of the prestage amount may be modified by decreasing the value of the prestage amount as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, in response to a determination that a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache. In one embodiment, the value of the prestage amount may be decreased by decreasing the value of the prestage trigger by a first value if the read data being prestaged is from a hard disk drive and by decreasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from a solid state storage drive.
Other features and aspects may be realized, depending upon the particular application.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 illustrates a trigger distance in a stream of data for triggering the prestaging of data into a cache.
FIG. 2 shows an example of cache operations in accordance with a prior art Adaptive Multi-stream Prefetching (AMP) algorithm.
FIG. 3 illustrates an embodiment of a computing environment in which an aspect of the present description may be employed.
FIG. 4 shows an example of operations of a cache in connection with prestaging tracks into the cache in accordance with one embodiment of the present description.
FIG. 5 shows a more detailed example of operations for prestage trigger and amount modification and other features in accordance with one embodiment of the present description.
FIG. 6 illustrates an example of prior art cache operations in which tracks may become candidates for demotion.
DETAILED DESCRIPTION
In the following description, reference is made to the accompanying drawings which form a part hereof, and which is shown, by way of illustration, several embodiments of the present invention. It is understood that other embodiments may be utilized and structural changes may be made without departing from the scope of the present invention.
FIG. 1 shows a set 100 of a stream of sequential read data which has been prefetched or prestaged into a cache. The set 100 may be a set of sequential pages, tracks or other units 110 of sequential read data. In this example, read data is prestaged in a prestageAmount (p) of tracks. A first portion 120 of the set 100 is indicated as being accessed by a storage controller or host in connection with a read operation. The remaining portion 130 of the set 100 is indicated as being unaccessed, that is, having not yet been accessed.
A trigger track 140 is identified at a trigger distance (g) from the end 150 of the unaccessed tracks 130 of the prestaged set 100. Once the tracks being accessed reaches the trigger track 140, another set 100 of the next p sequential tracks of the stream of sequential read data is prefetched or prestaged into the cache. Once the tracks being accessed reaches the trigger track 140 of that next set 100, yet another set 100 of the next p sequential tracks of the stream of sequential read data is prefetched or prestaged into the cache, and so on.
A prior art algorithm called AMP (Adaptive Multi-stream Prefetching) can adapt both the amount (p) of data being prestaged in a prestage operation, and the value of the trigger (g) for triggering the prestage operation, on a per stream basis in response to changing workloads. The AMP algorithm is based upon the theorem that improved performance may be obtained where, for each individual stream of sequential read data, the prestageAmount (p) should be: p=r*L, where r is the request rate of the stream (measured in tracks/second) and L is the life of the shared cache (MRU (Most Recently Used) timestamp minus LRU (Least Recently Used) timestamp). Accordingly, it is believed that when the prestageAmount p>r*L, that data in cache will likely be demoted before being used.
Also, for each individual stream the prestageTrigger g+1=r*t(p), 0≦g<p where t(p) is the average time to fetch p consecutive tracks. It is believed that when the prestageTrigger g<r*t(p), a cache miss is likely to happen.
FIG. 2 shows an example of cache operations in accordance with a prior art AMP algorithm, for determining both the number p of tracks to prestage and the value of the trigger g which triggers a prestage operation. Upon reading a track (block 200), the prestage amount (p) and the value of the prestage trigger (g) are initialized (block 202). In one prior art AMP algorithm, the prestage amount p may be initialized at 8 tracks and the value of the prestage trigger g may be initialized at distance of 3 tracks, for example.
If it is determined (block 204) that the track being read is on a sequential read stream and it is determined (block 206) that the track being read is at the last track of a stage group, the prestage amount p is increased (block 208) by 1. By increasing the prestage amount p, it is intended to decrease the chances of running out of prestaged data and causing a cache miss. However, if it is previously determined (block 210) that the prestage amount p has reached a maximum, such as a maximum of 84, for example, the prestage amount p is not increased.
If it is determined (block 206) that the track being read is not at the last track of a stage group, and it is determined (block 212) that the track being read is a cache miss, the prestage trigger distance g is increased (block 214) by 1. By increasing the prestage trigger distance g, it is intended to trigger the prestaging of additional data sooner to reduce the chances of future cache misses. However, if it determined (block 216) that the prestage trigger g plus 2 (g+2) has exceeded the prestage amount p, the prestage trigger is reduced (block 218) to the prestage amount p minus 2 (p−2).
If it is determined (block 212) that the track that was read was not a cache miss (that is, it was a cache hit), a determination is made (block 220) as to whether that track was prestaged in cache and whether it was a “candidate” for demotion from cache.
FIG. 6 illustrates an example of cache operations in which tracks may become candidates for demotion. A determination is made (block 600) as to whether additional space is needed in cache. If so, a determination is made (block 610) as to whether the LRU track is a candidate for demotion. If not, in this algorithm, LRU data is redesignated with an MRU timestamp (block 620) to give that data another chance to be used before being demoted, that is, discarded from cache, should the timestamp for that data again arrive at LRU status. LRU data redesignated with an MRU timestamp is referred to herein as a “candidate” for demotion. A flag is also set (block 620) to mark the LRU data redesignated with an MRU timestamp as a “candidate” for demotion. Once candidate data again arrives (block 610) at LRU status because the track has remained unused in cache for too long, the candidate data is demoted (block 640) from cache.
Referring back to FIG. 2, if it is determined (block 220) that the track being read was prestaged in cache and was a “candidate” for demotion from cache, both the prestage amount (p) and the prestage trigger g are each decreased (block 222) by 1. By decreasing the prestage amount p and the trigger g, it is intended to reduce the amount of data being prestaged each time and to delay the prestaging of additional data into cache, to reduce the number of candidates for demotion and to reduce demotion of data from cache.
Hardware and Software Environment
FIG. 3 illustrates an example of a hardware environment in which embodiments may be implemented. A plurality of host systems 310 a, 310 b, 310 c are in data communication with a plurality of storage units such as a hard disk drive 312 such as a DASD, and a solid state storage 314 such as a Flash memory, via a storage controller 320.
The storage controller further includes a cache 330 in which data is prestaged and staged. In accordance with one aspect of the present description, and as explained in greater detail below, the storage controller 320 can prestage data in such a manner that one or both of the prestage amount and the prestage trigger may be modified as a function of the speeds and bandwidths of the various storage devices coupled to the storage controller 320. Thus, in this embodiment, the solid state storage 314 has a substantially faster access time as compared to that of the hard drive storage 312. Consequently, when cache conditions indicate that it is appropriate to modify the prestage amount or prestage trigger, the prestage amount or prestage trigger for a stream of data being read from the faster flash memory storage 314 may be modified a lower amount, for example, as compared to a modification of the prestage amount or prestage trigger for a stream of data being read from slower hard drive storage 312. Other features and advantages will also be explained in connection with the present description.
The host systems 310 a, 310 b, 310 c may be any host system known in the art, such as a mainframe computer, workstations, etc., including an operating system such as WINDOWS.®., AIX.®., UNIX.®., MVS.™., etc. AIX is a registered trademark of IBM; MVS is a trademark of IBM; WINDOWS is a registered trademark of Microsoft Corporation; and UNIX is a registered trademark licensed by the X/Open Company LTD. A plurality of channel paths 322 a, 322 b, 322 c in the host systems 310 a, 310 b, 310 c provide communication paths to the storage controller 320. The storage controller 320 and host systems 310 a, 310 b, 310 c may communicate via any network or communication system known in the art, such as LAN, TCP/IP, ESCON.®., SAN, SNA, Fibre Channel, SCSI, etc. ESCON is a registered trademark of International Business Machines Corporation (“IBM”). The host system 310 a, 310 b, 310 c executes commands and receives returned data along a selected channel 322 a, 322 b, 322 c. The storage controller 320 issues commands via a path 324 a to physically position the electromechanical devices to read the storage 312. The storage controller 320 also issues commands via a path 324 b to read data from the solid state storage 314.
In preferred embodiments, the cache 330 is implemented in a high speed, volatile storage area within the storage controller 320, such as a DRAM, RAM, etc. The length of time since the last use of a record in cache 330 is maintained to determine the frequency of use of cache 330. The least recently used (LRU) data in cache 330 may be demoted according to LRU or other cache memory management techniques known in the art. In alternative embodiments, the cache 330 may be implemented in alternative storage areas accessible to the storage controller 320. Data can be transferred between the channels 322 a, 322 b, 322 c and the cache 330, between the channels 322 a, 322 b, 322 c and the storage 312, 314 and between the storage 312, 314 and the cache 330. In alternative embodiments with branching, data retrieved from the storage 312, 314 in response to a read miss can be concurrently transferred to both the channel 322 a, 322 b, 322 c and the cache 330 and a data write can be concurrently transferred from the channel 322 a, 322 b, 322 c to both a non-volatile storage unit and cache 330. It is appreciated that the number of hosts, storage controllers, caches and storage drives may vary, depending upon the particular application.
FIG. 4 shows one embodiment of operations of the cache 330 in connection with prestaging tracks into the cache 330. In a first operation, a track is read (block 400). A determination is made as to whether (block 404) the track being read is on a sequential read stream. If so, a determination is made as to whether (block 408) the track being read is a trigger track. If the track being read is a trigger track, the next group of tracks is prestaged (block 412) into cache.
A determination is also made as to whether the conditions in the cache are such that the prestage trigger or prestage amount should be modified (block 414). If so, the prestage trigger may be modified (block 418) as a function of the different speeds of the different drives coupled to the storage controller. FIG. 5 discussed below shows an example of such a determination and an example of such a prestage trigger or amount modification as a function of different drives speeds.
In another operation, a determination is made as to whether to stage tracks (block 430). If so, additional tracks are staged (block 434). FIG. 5 discussed below shows an example of such a determination.
FIG. 5 shows a more detailed example of certain operations for prestage trigger and amount modification and other features. In this example, modifications to the prestage trigger are based upon a theorem that:
r*t(N)<=g+1<=r*L
where g is the prestage trigger, r is the request rate of the stream (tracks/sec), L is the life of the shared cache (MRU timestamp−LRU timestamp), t(N) is the average time to fetch N consecutive tracks, and N is the stride size of the device (4*rank width).
In one operation, the prestage amount (p) and the value of the prestage trigger (g) are initialized (block 500). In this example, the prestage amount p may be initialized at 8 tracks and the value of the prestage trigger g may be initialized at distance of 3 tracks, for example. Thus, it will be seen that when a track is read, the prestage trigger of this track is the larger one of 3 or the prestage trigger of the previous track. It is appreciated that other values may be selected for initialization of the prestage amount p and the prestage trigger g, depending upon the particular application
A track is read (block 504) and a determination is made as to whether (block 508) the track being read is on a sequential read stream. If so, another determination is made as to whether (block 510) the track being read was in connection with a cache miss. If so, a determination is made as to whether (block 512) a stage issued in connection with the track being read. If so, the prestage trigger is increased (block 516) as a function of different speeds of different drives and tracks are staged (block 524).
If a stage had issued in connection with the read which resulted in a cache miss, the cache miss may have occurred as a result of waiting for that issued stage to complete. In this example, the prestage trigger of this track may be increased based on the speed of the drives. For example, if the track was read from the hard drive storage 312, the prestage trigger may be increased by 1, for example. If the track was read from a faster drive such as the solid state storage 314, the prestage trigger may be increased by a greater amount such as by 2, for example. Such differentiated increments may facilitate aiding the completion of the issued stages before a cache miss occurs. It is appreciated that other values may be selected for incrementing the prestage trigger g as a function of different drive speeds, depending upon the particular application.
Alternatively, if a stage had not issued (block 512) in connection with the read which resulted in a cache miss, the cache miss may have occurred as a result of that track having been demoted before being used. As a result, it may be appropriate to delay prestaging additional tracks or to reduce (block 530) the amount being prestaged.
In this example, both the prestage trigger and the prestage amount may be decreased (block 530) as a function of the different speeds of the different drives. For example, if the track was read from the hard drive storage 312, the prestage trigger may be decreased by 1, for example, and the prestage amount may be reduced by 1, for example. If the track was read from a faster drive such as solid state storage 314, the prestage trigger may be decreased by a greater amount such as by 2, for example, and the prestage amount may be reduced by 2, for example. Such differentiated decrements may facilitate reducing the number of candidate tracks for demotion. It is appreciated that other values may be selected for decrementing the prestage trigger g and the prestage amount p as a function of different drive speeds, depending upon the particular application. In addition, tracks are staged (block 524).
If it is determined (block 510) the track being read was a cache hit rather than a cache miss, another determination is made as to whether (block 540) the track being read is a prestage track candidate for demotion. If so, it may be appropriate to reduce the amount of data being prestaged to reduce the number of candidates for demotion. In this embodiment, the prestage amount is decreased (block 544) as a function of different speeds of different drives.
For example, if the track was read from the hard drive storage 312 which is slower than the solid state drive 314, the prestage amount may be decreased by 1, for example. If the track was read from a faster drive such as solid state storage 314, the prestage amount may be decreased by a greater amount such as by 2, for example. Such differentiated decrements may facilitate reducing the number of candidate tracks for demotion. It is appreciated that other values may be selected for decrementing the prestage amount p as a function of different drive speeds, depending upon the particular application.
If it is determined (block 556) that the track being read is at the last track of a stage group, the prestage amount p may be increased (block 558) as a function of different speeds of different drives. For example, if the track was read from the hard drive storage 312 which is slower than the solid state drive 314, the prestage amount may be increased by 1, for example. If the track was read from a faster drive such as solid state storage 314, the prestage amount may be increased by a larger amount such as by 2, for example. Such differentiated increments may facilitate prestaging sufficient data to reduce cache misses. It is appreciated that other values may be selected for incrementing the prestage amount p as a function of different drive speeds, depending upon the particular application.
By increasing the prestage amount p, it is intended to decrease the chances of running out of prestaged data and causing a cache miss. However, if it is previously determined (block 560) that the prestage amount p has reached a maximum, such as a maximum of 84, for example, the prestage amount p is not increased. It is appreciated that other values may be selected for a maximum, depending upon the particular application.
In the illustrated embodiment, another determination is made as to whether the size of the prestage trigger has reached or exceeded (block 590) the prestage amount p. If so, additional tracks may prestaged (block 595). It is believed that issuing additional prestages will permit the prestage trigger g to continue to grow to facilitate avoiding a cache miss due to prestage triggers of insufficient size. It is appreciated that other values may be selected for determining when to prestage additional tracks as a function of prestage trigger size, depending upon the particular application.
It is seen that cache prestaging operations in accordance with one aspect of the present description may take into account storage devices of varying speeds and bandwidths for purposes of modifying one or both of a prestage trigger and the prestage amount. Such a capability may be particularly suitable for storage controllers having a mix of solid state storage and hard drive storage, for example.
Still further, a cache prestaging operation in accordance with another aspect of the present description may decrease one or both of the prestage trigger and the prestage amount in circumstances such as a cache miss which may have resulted from prestaged tracks being demoted before they are used. Conversely, a cache prestaging operation in accordance with another aspect of the present description may increase one or both of the prestage trigger and the prestage amount in circumstances such as a cache miss which may have resulted from waiting for a stage to complete.
In yet another aspect, the prestage trigger may not be limited by the prestage amount p. Instead, the pre-stage trigger may be permitted to expand as conditions warrant it by prestaging additional tracks and thereby effectively increasing the potential range for the prestage trigger.
Additional Embodiment Details
The described operations may be implemented as a method, apparatus or computer program product using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof. Accordingly, aspects of the embodiments may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the embodiments may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, interconnected storage devices, an array of storage devices, multiple memory or storage devices or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
In certain embodiments, the system of FIG. 2 may be implemented as a cloud component part in a cloud computing environment. In the cloud computing environment, the systems architecture of the hardware and software components involved in the delivery of cloud computing may comprise a plurality of cloud components communicating with each other over a network, such as the Internet. For example, in certain embodiments, the system of FIG. 2 may provide clients, and other servers and software and/or hardware components in the networked cloud, with scheduling services.
The terms “an embodiment”, “embodiment”, “embodiments”, “the embodiment”, “the embodiments”, “one or more embodiments”, “some embodiments”, and “one embodiment” mean “one or more (but not all) embodiments of the present invention(s)” unless expressly specified otherwise.
The terms “including”, “comprising”, “having” and variations thereof mean “including but not limited to”, unless expressly specified otherwise.
The enumerated listing of items does not imply that any or all of the items are mutually exclusive, unless expressly specified otherwise.
The terms “a”, “an” and “the” mean “one or more”, unless expressly specified otherwise.
Devices that are in communication with each other need not be in continuous communication with each other, unless expressly specified otherwise. In addition, devices that are in communication with each other may communicate directly or indirectly through one or more intermediaries.
A description of an embodiment with several components in communication with each other does not imply that all such components are required. On the contrary a variety of optional components are described to illustrate the wide variety of possible embodiments of the present invention.
Further, although process steps, method steps, algorithms or the like may be described in a sequential order, such processes, methods and algorithms may be configured to work in alternate orders. In other words, any sequence or order of steps that may be described does not necessarily indicate a requirement that the steps be performed in that order. The steps of processes described herein may be performed in any order practical. Further, some steps may be performed simultaneously.
When a single device or article is described herein, it will be readily apparent that more than one device/article (whether or not they cooperate) may be used in place of a single device/article. Similarly, where more than one device or article is described herein (whether or not they cooperate), it will be readily apparent that a single device/article may be used in place of the more than one device or article or a different number of devices/articles may be used instead of the shown number of devices or programs. The functionality and/or the features of a device may be alternatively embodied by one or more other devices which are not explicitly described as having such functionality/features. Thus, other embodiments of the present invention need not include the device itself.
The illustrated operations of FIGS. 4 and 5 shows certain events occurring in a certain order. In alternative embodiments, certain operations may be performed in a different order, modified or removed. Moreover, steps may be added to the above described logic and still conform to the described embodiments. Further, operations described herein may occur sequentially or certain operations may be processed in parallel. Yet further, operations may be performed by a single processing unit or by distributed processing units.
In some embodiments, a solid state storage 314 was described as having a substantially faster drive speed or access time as compared to that of a hard drive storage 312. However, it is appreciated that features of the present description are also applicable or beneficial to additional embodiments in which the drive speeds of various drives of the system may be the same, as well, depending upon the particular application.
The foregoing description of various embodiments of the invention has been presented for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed. Many modifications and variations are possible in light of the above teaching. It is intended that the scope of the invention be limited not by this detailed description, but rather by the claims appended hereto. The above specification, examples and data provide a complete description of the manufacture and use of the composition of the invention. Since many embodiments of the invention can be made without departing from the spirit and scope of the invention, the invention resides in the claims hereinafter appended.

Claims (12)

What is claimed is:
1. A system for use with a host, comprising:
a first storage drive having a first drive speed;
a second storage drive having a second drive speed different from that of the first drive speed;
a storage controller adapted to control the first and second storage drives, said storage controller including a processor and a cache memory which are adapted to perform operations including:
reading units of data of a sequential stream of read data stored in at least one of the first storage drive and the second storage drive;
setting a prestage trigger to initiate prestaging of units of read data;
in response to reaching the trigger as the units of the sequential stream of read data are read, prestaging an amount of units of data of the sequential stream of read data from a storage drive into the cache memory in anticipation of a read request from the host; and
modifying at least one of the value of the prestage trigger and the value of the prestage amount, as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, being different from the other storage drive.
2. The system of claim 1, wherein the operations further comprise:
staging units of data of the sequential stream of read data into the cache memory;
demoting a unit of data from the cache before the unit being demoted is read from cache in response to a read request from the host; and
determining if a cache miss occurs in connection with a unit of data that had been demoted from the cache;
wherein modifying at least one of the value of the prestage trigger and the value of the prestage amount includes reducing both the value of the prestage trigger and the value of the prestage amount, in response to a determination that a cache miss occurs in connection with a unit of read data that had been demoted from the cache, and as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, wherein the reducing at least one of the value of the prestage trigger and the value of the prestage amount, in association with the demoting of a unit of data from the cache includes determining if a cache miss occurred and determining if a stage was in progress at the time of the cache miss.
3. The system of claim 1, wherein the operations further comprise:
comparing the modified value of the prestage trigger to the amount of data to be prestaged in response to the trigger; and
prestaging units of data of the sequential stream of read data into the cache memory as a function of the comparison,
in response to reaching the modified value trigger as the units of the sequential stream of read data are read, prestaging units of data after staged units of read data of the sequential stream of read data from a storage drive into the cache memory in anticipation of a read request.
4. The system of claim 1, wherein the operations further comprise:
staging units of data of the sequential stream of read data into the cache memory; and
determining if a cache miss occurred because a staging has not completed;
wherein said modifying at least one of the value of the prestage trigger and the value of the prestage amount, includes increasing the value of the prestage trigger as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, if it is determined that a cache miss occurred because a staging has not completed; and
wherein the first storage drive is a hard disk drive and the second storage drive is a flash memory drive having a drive speed faster than that of the hard disk drive, and wherein said increasing the value of the prestage trigger includes increasing the value of the prestage trigger by a first value if the read data being prestaged is from the hard disk drive, and increasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from the flash drive.
5. The system of claim 1, wherein the operations further comprise:
monitoring unused units of read data in the cache memory;
determining whether unused units of read data in the cache are candidates for demotion from the cache as a function of time spent in the cache relative to other unused units of read data in the cache;
determining if a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache;
wherein said modifying at least one of the value of the prestage trigger and the value of the prestage amount includes decreasing the value of the prestage amount as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, in response to a determination that a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache; and
wherein the first storage drive is a hard disk drive and the second storage drive is a flash memory drive having a drive speed faster than that of the hard disk drive, and wherein said decreasing the value of the prestage trigger includes decreasing the value of the prestage trigger by a first value if the read data being prestaged is from the hard disk drive and decreasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from the flash drive.
6. The system of claim 1, wherein the operations further comprise:
staging units of data of the sequential stream of read data into the cache memory; and
determining if a read track is the last track of a stage group;
wherein modifying at least one of the value of the prestage trigger and the value of the prestage amount includes increasing the value of the prestage amount as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory.
7. A computer program product, for use with a host, a storage controller and storage drives, the computer program product comprising a nontransitory computer readable storage medium having computer readable program code embodied therein that executes to perform operations, the operations being in the storage controller and storage devices, the operations comprising:
reading units of data of a sequential stream of read data stored in at least one of a first storage drive having a first drive speed and a second storage drive having a second drive speed different from that of the first drive speed;
setting a prestage trigger to initiate prestaging of units of read data;
in response to reaching the trigger as the units of the sequential stream of read data are read, prestaging an amount of units of data of the sequential stream of read data from a storage drive into the cache memory in anticipation of a read request from the host; and
modifying at least one of the value of the prestage trigger and the value of the prestage amount, as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, being different from the other storage drive.
8. The computer program product of claim 7, wherein the operations further comprise:
staging units of data of the sequential stream of read data into the cache memory;
demoting a unit of data from the cache before the unit being demoted is read from cache in response to a read request from the host; and
determining if a cache miss occurs in connection with a unit of data that had been demoted from the cache;
wherein modifying at least one of the value of the prestage trigger and the value of the prestage amount includes reducing both the value of the prestage trigger and the value of the prestage amount, in response to a determination that a cache miss occurs in connection with a unit of read data that had been demoted from the cache, and as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, wherein the reducing at least one of the value of the prestage trigger and the value of the prestage amount, in association with the demoting of a unit of data from the cache includes determining if a cache miss occurred and determining if a stage was in progress at the time of the cache miss.
9. The computer program product of claim 7, wherein the operations further comprise:
comparing the modified value of the prestage trigger to the amount of data to be prestaged in response to the trigger; and
prestaging units of data of the sequential stream of read data into the cache memory as a function of the comparison,
in response to reaching the modified value trigger as the units of the sequential stream of read data are read, prestaging units of data after staged units of read data of the sequential stream of read data from a storage drive into the cache memory in anticipation of a read request.
10. The computer program product of claim 7, wherein the operations further comprise:
staging units of data of the sequential stream of read data into the cache memory; and
determining if a cache miss occurred because a staging has not completed;
wherein said modifying at least one of the value of the prestage trigger and the value of the prestage amount, includes increasing the value of the prestage trigger as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, if it is determined that a cache miss occurred because a staging has not completed; and
wherein the first storage drive is a hard disk drive and the second storage drive is a flash memory drive having a drive speed faster than that of the hard disk drive, and wherein said increasing the value of the prestage trigger includes increasing the value of the prestage trigger by a first value if the read data being prestaged is from the hard disk drive, and increasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from the flash drive.
11. The computer program product of claim 7, wherein the operations further comprise:
monitoring unused units of read data in the cache memory;
determining whether unused units of read data in the cache are candidates for demotion from the cache as a function of time spent in the cache relative to other unused units of read data in the cache;
determining if a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache;
wherein said modifying at least one of the value of the prestage trigger and the value of the prestage amount includes decreasing the value of the prestage amount as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory, in response to a determination that a cache hit occurs in connection with a unit of read data which was a candidate for demotion from the cache; and
wherein the first storage drive is a hard disk drive and the second storage drive is a flash memory drive having a drive speed faster than that of the hard disk drive, and wherein said decreasing the value of the prestage trigger includes decreasing the value of the prestage trigger by a first value if the read data being prestaged is from the hard disk drive and decreasing the value of the prestage trigger by a second value higher than the first value, if the read data being prestaged is from the flash drive.
12. The computer program product of claim 7, wherein the operations further comprise:
staging units of data of the sequential stream of read data into the cache memory; and
determining if a read track is the last track of a stage group;
wherein modifying at least one of the value of the prestage trigger and the value of the prestage amount includes increasing the value of the prestage amount as a function of the drive speed of the storage drive from which the units of read data are prestaged into the cache memory.
US13/018,305 2011-01-31 2011-01-31 Adaptive prestaging in a storage controller Expired - Fee Related US9141525B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/018,305 US9141525B2 (en) 2011-01-31 2011-01-31 Adaptive prestaging in a storage controller
US13/786,981 US8874840B2 (en) 2011-01-31 2013-03-06 Adaptive prestaging in a storage controller

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/018,305 US9141525B2 (en) 2011-01-31 2011-01-31 Adaptive prestaging in a storage controller

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/786,981 Continuation US8874840B2 (en) 2011-01-31 2013-03-06 Adaptive prestaging in a storage controller

Publications (2)

Publication Number Publication Date
US20120198148A1 US20120198148A1 (en) 2012-08-02
US9141525B2 true US9141525B2 (en) 2015-09-22

Family

ID=46578364

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/018,305 Expired - Fee Related US9141525B2 (en) 2011-01-31 2011-01-31 Adaptive prestaging in a storage controller
US13/786,981 Expired - Fee Related US8874840B2 (en) 2011-01-31 2013-03-06 Adaptive prestaging in a storage controller

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/786,981 Expired - Fee Related US8874840B2 (en) 2011-01-31 2013-03-06 Adaptive prestaging in a storage controller

Country Status (1)

Country Link
US (2) US9141525B2 (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9141525B2 (en) 2011-01-31 2015-09-22 International Business Machines Corporation Adaptive prestaging in a storage controller
US9201794B2 (en) * 2011-05-20 2015-12-01 International Business Machines Corporation Dynamic hierarchical memory cache awareness within a storage system
US9128854B1 (en) * 2012-09-28 2015-09-08 Emc Corporation System and method for data prediction
US9323682B1 (en) * 2013-04-05 2016-04-26 Emc Corporation Non-intrusive automated storage tiering using information of front end storage activities
US9292454B2 (en) * 2013-09-27 2016-03-22 Microsoft Technology Licensing, Llc Data caching policy in multiple tenant enterprise resource planning system
US11940920B2 (en) * 2018-06-20 2024-03-26 International Business Machines Corporation Dynamically determining tracks to prestage from storage to cache using a machine learning module
US11068417B2 (en) 2018-06-26 2021-07-20 International Business Machines Corporation Allocation of cache storage among applications that indicate minimum retention time for tracks in least recently used demoting schemes
US11074197B2 (en) 2018-06-26 2021-07-27 International Business Machines Corporation Integration of application indicated minimum time to cache and maximum time to cache to least recently used track demoting schemes in a cache management system of a storage controller
US11144474B2 (en) 2018-06-26 2021-10-12 International Business Machines Corporation Integration of application indicated maximum time to cache to least recently used track demoting schemes in a cache management system of a storage controller
US11068413B2 (en) 2018-06-26 2021-07-20 International Business Machines Corporation Allocation of cache storage among applications based on application priority and minimum retention time for tracks in least recently used demoting schemes
US11061826B2 (en) 2018-06-26 2021-07-13 International Business Machines Corporation Integration of application indicated minimum time to cache to least recently used track demoting schemes in a cache management system of a storage controller
US11262946B2 (en) * 2019-11-25 2022-03-01 Micron Technology, Inc. Cache-based memory read commands

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5426761A (en) * 1992-12-22 1995-06-20 International Business Machines Corporation Cache DASD sequential staging and method
US5694581A (en) * 1993-09-07 1997-12-02 Industrial Technology Research Institute Concurrent disk array management system implemented with CPU executable extension
US5764430A (en) * 1996-04-01 1998-06-09 International Business Machines Corporation Disk drive having optimized spindle speed for environment
US5898891A (en) * 1996-06-04 1999-04-27 Micron Electronics, Inc. Method for transferring data directly between the first and second data storage devices without transferring data to the memory array or over the input-output bus
US6260115B1 (en) 1999-05-13 2001-07-10 Storage Technology Corporation Sequential detection and prestaging methods for a disk storage subsystem
US6345338B1 (en) 1999-03-23 2002-02-05 Storage Technology Corporation Adapting resource use to improve performance in a caching memory system
US6381677B1 (en) 1998-08-19 2002-04-30 International Business Machines Corporation Method and system for staging data into cache
US6449697B1 (en) * 1999-04-23 2002-09-10 International Business Machines Corporation Prestaging data into cache in preparation for data transfer operations
US20050240724A1 (en) 2004-04-21 2005-10-27 Hiroshi Koizumi Disk array device and cache control method for disk array device
US7080212B1 (en) 2003-12-29 2006-07-18 Storage Technology Corporation Closed loop adaptive prestage method, system, and product for prestaging cache blocks
US7079341B2 (en) * 2004-03-11 2006-07-18 International Business Machines Corporation Multiple disk, variable RPM data storage system for reducing power consumption
US7120736B2 (en) * 2003-05-22 2006-10-10 Hitachi, Ltd. Storage unit and circuit for shaping communication signal
US20090198738A1 (en) 2008-02-05 2009-08-06 Berger Jeffrey A System and Method for an Adaptive List Prefetch
US20090248964A1 (en) 2008-03-01 2009-10-01 Kabushiki Kaisha Toshiba Memory system and method for controlling a nonvolatile semiconductor memory
US20100138620A1 (en) 2008-12-03 2010-06-03 Martin Jess Capturing and naming dynamic storage tiering configurations to support data pre-staging
US20100195243A1 (en) * 2009-02-03 2010-08-05 Seagate Technology Llc Variable Spindle Speed Control for Data Storage Devices
US7822920B2 (en) * 2007-07-13 2010-10-26 Korea Advanced Institute Of Science And Technology Mass prefetching method for disk array

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9141525B2 (en) 2011-01-31 2015-09-22 International Business Machines Corporation Adaptive prestaging in a storage controller

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5426761A (en) * 1992-12-22 1995-06-20 International Business Machines Corporation Cache DASD sequential staging and method
US5694581A (en) * 1993-09-07 1997-12-02 Industrial Technology Research Institute Concurrent disk array management system implemented with CPU executable extension
US5764430A (en) * 1996-04-01 1998-06-09 International Business Machines Corporation Disk drive having optimized spindle speed for environment
US5898891A (en) * 1996-06-04 1999-04-27 Micron Electronics, Inc. Method for transferring data directly between the first and second data storage devices without transferring data to the memory array or over the input-output bus
US6381677B1 (en) 1998-08-19 2002-04-30 International Business Machines Corporation Method and system for staging data into cache
US6345338B1 (en) 1999-03-23 2002-02-05 Storage Technology Corporation Adapting resource use to improve performance in a caching memory system
US6993629B2 (en) 1999-04-23 2006-01-31 International Business Machines Corporation Prestaging data into cache in preparation for data transfer operations
US6449697B1 (en) * 1999-04-23 2002-09-10 International Business Machines Corporation Prestaging data into cache in preparation for data transfer operations
US6260115B1 (en) 1999-05-13 2001-07-10 Storage Technology Corporation Sequential detection and prestaging methods for a disk storage subsystem
US7120736B2 (en) * 2003-05-22 2006-10-10 Hitachi, Ltd. Storage unit and circuit for shaping communication signal
US7080212B1 (en) 2003-12-29 2006-07-18 Storage Technology Corporation Closed loop adaptive prestage method, system, and product for prestaging cache blocks
US7079341B2 (en) * 2004-03-11 2006-07-18 International Business Machines Corporation Multiple disk, variable RPM data storage system for reducing power consumption
US20050240724A1 (en) 2004-04-21 2005-10-27 Hiroshi Koizumi Disk array device and cache control method for disk array device
US7822920B2 (en) * 2007-07-13 2010-10-26 Korea Advanced Institute Of Science And Technology Mass prefetching method for disk array
US20090198738A1 (en) 2008-02-05 2009-08-06 Berger Jeffrey A System and Method for an Adaptive List Prefetch
US20090248964A1 (en) 2008-03-01 2009-10-01 Kabushiki Kaisha Toshiba Memory system and method for controlling a nonvolatile semiconductor memory
US20100138620A1 (en) 2008-12-03 2010-06-03 Martin Jess Capturing and naming dynamic storage tiering configurations to support data pre-staging
US20100195243A1 (en) * 2009-02-03 2010-08-05 Seagate Technology Llc Variable Spindle Speed Control for Data Storage Devices

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
B.S. Gill, et al., "AMP: Adaptive Multi-stream Prefetching in a Shared Cache", USENIX Association, FAST '07: 5th USENIX Conference on File and Storage Technologies, pp. 185-198.
Final Office Action dated Mar. 11, 2014, pp. 25, for U.S. Appl. No. 13/786,981, filed Mar. 6, 2013.
Notice of Allowance dated Jun. 23, 2014, pp. 9, for U.S. Appl. No. 13/786,981, filed Mar. 6, 2013.
Office Action dated Sep. 12, 2013, pp. 17, for U.S. Appl. No. 13/786,981, filed Mar. 6, 2013.
Preliminary Remarks filed Mar. 6, 2013, pp. 1-2, for U.S. Appl. No. 13/786,981, filed Mar. 6, 2013 by inventors M.T. Benhase, et al.
Response dated Dec. 12, 2013, pp. 8, to Office Action dated Dec. 12, 2013, pp. 17, for U.S. Appl. No. 13/786,981, filed Mar. 6, 2013.
Response dated May 12, 2014, pp. 9, to Final Office Action dated Mar. 11, 2014, pp. 25, for U.S. Appl. No. 13/786,981, filed Mar. 6, 2013.
U.S. Appl. No. 13/786,981, filed Mar. 6, 2013 by inventors M.T. Benhase, et al.

Also Published As

Publication number Publication date
US8874840B2 (en) 2014-10-28
US20130246691A1 (en) 2013-09-19
US20120198148A1 (en) 2012-08-02

Similar Documents

Publication Publication Date Title
US8874840B2 (en) Adaptive prestaging in a storage controller
US10380035B2 (en) Using an access increment number to control a duration during which tracks remain in cache
US7702857B2 (en) Adjusting parameters used to prefetch data from storage into cache
JP7216741B2 (en) Select one of several cache deletion algorithms to use to remove tracks from the cache
US9811474B2 (en) Determining cache performance using a ghost cache list indicating tracks demoted from a cache list of tracks in a cache
US8762650B2 (en) Prefetching tracks using multiple caches
US11030116B2 (en) Processing cache miss rates to determine memory space to add to an active cache to reduce a cache miss rate for the active cache
US6381677B1 (en) Method and system for staging data into cache
US8935479B2 (en) Adaptive cache promotions in a two level caching system
US8677062B2 (en) Caching data in a storage system having multiple caches including non-volatile storage cache in a sequential access storage device
US8788742B2 (en) Using an attribute of a write request to determine where to cache data in a storage system having multiple caches including non-volatile storage cache in a sequential access storage device
US9477607B2 (en) Adaptive record caching for solid state disks
US20030088739A1 (en) Adaptive data insertion for caching
US20080183969A1 (en) System and method for dynamic sizing of cache sequential list
US9996476B2 (en) Management of cache lists via dynamic sizing of the cache lists
US20170052897A1 (en) Using cache lists for multiple processors to cache and demote tracks in a storage system
US10628331B2 (en) Demote scan processing to demote tracks from cache
US20140359063A1 (en) Computer system, cache control method, and server
US8819343B2 (en) Periodic destages from inside and outside diameters of disks to improve read response times
US8825952B2 (en) Handling high priority requests in a sequential access storage device having a non-volatile storage cache
US8996789B2 (en) Handling high priority requests in a sequential access storage device having a non-volatile storage cache
US7277991B2 (en) Method, system, and program for prefetching data into cache

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BENHASE, MICHAEL T.;FRANCISCO, NEDLAYA Y.;GILL, BINNY S.;AND OTHERS;SIGNING DATES FROM 20110124 TO 20110127;REEL/FRAME:025918/0160

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20190922