US20130173849A1 - Write bandwidth management for flashdevices - Google Patents
Write bandwidth management for flashdevices Download PDFInfo
- Publication number
- US20130173849A1 US20130173849A1 US13/525,017 US201213525017A US2013173849A1 US 20130173849 A1 US20130173849 A1 US 20130173849A1 US 201213525017 A US201213525017 A US 201213525017A US 2013173849 A1 US2013173849 A1 US 2013173849A1
- Authority
- US
- United States
- Prior art keywords
- flash memory
- vms
- memory device
- write
- tokens
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/061—Improving I/O performance
- G06F3/0613—Improving I/O performance in relation to throughput
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0655—Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
- G06F3/0659—Command handling arrangements, e.g. command buffers, queues, command scheduling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/0671—In-line storage system
- G06F3/0673—Single storage device
- G06F3/0679—Non-volatile semiconductor memory device, e.g. flash memory, one time programmable memory [OTP]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5011—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
- G06F9/5016—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals the resource being the memory
Definitions
- the present invention relates to flash memory device reliability and more particularly to flash memory write cycle bandwidth management.
- Flash memory refers to non-volatile computer readable storage medium that can be electrically erased and reprogrammed. Flash memory initially had been developed from electrically erasable programmable read-only memory and stores information in an array of memory cells made from floating-gate transistors. In traditional single-level cell devices, each cell of flash memory stores only one bit of information. Newer flash memory, known as multi-level cell devices, can store more than one bit per cell by choosing between multiple levels of electrical charge to apply to the floating gates of its cells.
- Flash cells of a flash memory device can provide only a finite number of erase cycles. When the number of erases in the flash memory device exceeds this limit, the cell fails and eventually, the flash memory device fails. Flash memory devices are also capable of very high performance, which in effect, accelerates the end of life of the flash memory device. Flash memory devices also provide a warranty which guarantee reliable operation for a finite duration of time. Of note, it is well within the performance envelope to exceed the number of erase cycles of a flash memory device within the warranty period. Thus, in order to meet warranty requirements, flash memory devices may throttle write transactions. By slowing down the write transactions, the number of erase transactions is reduced, thus increasing the lifetime of the flash memory device.
- a single flash memory device may be shared by multiple VMs. So much is not a problem so long as the available write bandwidth of the flash memory device exceeds the bandwidth required by all of the VMs. However, as the available bandwidth of the flash memory decreases during write transaction throttling, a resource sharing issue can arise.
- the limited write bandwidth of a flash memory device is not intelligently partitioned among the VMs. Instead, in a conventional VM-agnostic flash memory device controller, the write bandwidth remains a resource that is freely available to any VM demanding the write access to the flash memory device. This leads to a situation where the most aggressive VM is provided all or most of the available write bandwidth of the flash memory device, potentially leaving other VMs starving for access to the flash memory device.
- Embodiments of the present invention address deficiencies of the art in respect to flash memory device access management in a virtualized environment and provide a novel and non-obvious method, system and computer program product for flash memory device write-access management amongst different VMs in a virtualized computing environment.
- a virtualized computing data processing system can include a host computer with at least one processor and memory and different VMs executing in the host computer.
- the system also can include a flash memory device coupled to the host computer and accessible by the VMs.
- a flash memory controller can manage access to the flash memory device.
- the controller can include program code enabled to compute a contemporaneous bandwidth of requests for write operations for the flash memory device, to allocate a corresponding number of tokens to the VMs, to accept write requests to the flash memory device from the VMs only when accompanied by a token and to repeat the computing, allocating and accepting after a lapse of a pre-determined time period.
- the program code of the controller further detects the lapse of the predetermined time period, and in response to the detecting of the lapse reallocates unused tokens in over-allocated ones of the VMs to under-allocated ones of the VMs for use during a subsequent pre-determined time period.
- the program code of the controller further identifies ones of the VMs experiencing a burst of requests to write to the flash memory device and allocates in response to the identification of VMs experiencing a burst of requests additional tokens to the VMs experiencing a burst of requests.
- FIG. 1 is a pictorial illustration of a process for flash memory device write-access management amongst different VMs in a virtualized computing environment
- FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment;
- FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
- FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
- Embodiments of the invention provide for flash memory device write-access management amongst different VMs in a virtualized computing environment.
- a quantity of tokens can be computed corresponding to a measured available write-cycle bandwidth of a flash memory device provided during throttling of the flash memory device.
- different VMs seeking access to the flash memory device can be allocated an equal number of the quantity of the tokens with which access is granted to institute a write operation on the flash memory device.
- unused tokens allocated to one or more of the VMs can be redistributed to VMs identified as lacking enough tokens to institute all desired write operations and the process can repeat.
- VMs demonstrating a burst of required write access to the flash memory device can be allocated additional tokens on a temporary basis. In this way, write access to the flash memory device can be intelligently and fairly managed in a virtualized environment despite throttling performed by the flash memory device.
- FIG. 1 shows a process for flash memory device write-access management amongst different VMs in a virtualized computing environment.
- different VMs 110 A, 110 B, 110 N of a virtualized computing environment 100 can request write operations in a flash memory device 130 of the virtualized computing environment 100 .
- Each write request 120 issued by a corresponding one of the VMs 110 A, 110 B, 110 N can be tracked within a corresponding request queue 150 A, 150 B, 150 N and can be accompanied by a token 160 A, 160 B, 160 N.
- a device controller 140 can manage access to the flash memory device 130 and can grant a request 120 issued by a VM 110 A, 110 B, 110 N only if the request 120 is accompanied by a token 160 A, 160 B, 160 N.
- the device controller 140 can maintain a steady state pool of tokens 180 A, a free pool of tokens 180 B and a burst pool of tokens 180 C.
- the steady state pool of tokens 180 A can include a number of tokens equivalent to a determined bandwidth 170 of the flash memory device 130 in respect to a number of write requests able to be accommodated by the flash memory device 130 irrespective of throttling established by the flash memory device 130 .
- An additional number of tokens can be provided for the burst pool of tokens 180 C.
- the device controller 140 initially can allocate the tokens in the steady state pool of tokens 180 A evenly amongst the VMs 110 A, 110 B, 110 N.
- unused tokens 160 A, 160 B, 160 N provided to the VMs 110 A, 110 B, 110 N can be placed in the free pool of tokens 180 B for use by ones of the VMs 110 A, 110 B, 110 N that suffer a deficiency of tokens 160 A, 160 B, 160 N to satisfy desired write operations in the flash memory device 130 .
- the queues 150 A, 150 B, 150 N can be monitored to identify VMs 110 A, 110 B, 110 N demonstrating an enhanced number of requests 120 for write operations in the flash memory device 130 .
- tokens from the burst pool of tokens 180 C can be provided to those of the VMs 110 A, 110 B, 110 N demonstrating an enhanced number of requests 120 for write operations in the flash memory device 130 .
- FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment.
- the system can include a host computer 210 with at least one processor and memory.
- the host computer 210 can support the execution of an operating system 220 which in turn can manage the operation of different VMs 260 .
- a flash device 230 can be coupled to the host computer 210 and can provide a fixed storage medium for use by the VMs 260 .
- a flash device controller 250 can be provided to manage access requests by the VMs 260 to the flash device 230 and in particular, write operation requests by the VMs 260 .
- the flash device controller 250 can include program code enabled upon execution by the host computer 210 to compute a number of tokens for placement in a token data store 240 , to distribute the tokens to the VMs 260 and to permit write operation access to the flash memory device 230 only when such a request by a corresponding one of the VMs 260 is accompanied by a token.
- FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
- the available bandwidth of the flash memory device can be retrieved, for example by contemporaneously monitoring a number of I/O operations or write operations accepted by the flash memory device over a set period of time.
- a pool of steady state tokens can be populated in accordance with the bandwidth and in block 320 , a count of a number of VMs active in the virtualized environment can be determined. Thereafter, in block 330 the tokens in the steady state pool can be allocated evenly across the VMs for use by the VMs whenever a memory write operation is required in respect to the flash memory device.
- a delay can be incurred and in decision block 350 , it can be determined whether a predetermined period of time—namely an epoch—has transpired. If so, in block 360 , those of the VMs with tokens remaining and unused can be identified as VMs which received an unnecessary over-allocation of tokens and those tokens can be placed into a free pool of tokens in block 370 . Additionally, in block 380 , request queues for write operation requests of the VMs can be inspected to identify VMs which are likely not to have enough tokens to satisfy desired requests for write operations on the flash memory device.
- tokens from the free pool of tokens can be allocated to those of the VMs determined to have been under-allocated.
- the process can repeat in block 300 with a new determination of the contemporaneous bandwidth of the flash memory device.
- FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
- a burst pool of tokens can be initialized to include a selection of tokens and in block 420 , a sub-interval of the epoch can be determined.
- a delay can be incurred and in decision block 440 , if the sub-interval has elapsed, in block 450 a first VM can be selected for analysis.
- the request queue of the selected VM can be inspected to determine the rate of additions of requests to the queue.
- decision block 470 if the rate of request growth in the request queue has exceeded a threshold value, in block 480 one or more tokens in the burst pool of tokens can be allocated to the selected VM for use only once before being returned to the burst pool of tokens. Thereafter, in decision block 490 , if more VMs remain to be analyzed, the process can repeat in block 450 for a next selected VM. Otherwise, again it can be determined if a next sub-interval has lapsed and the process can begin anew.
- aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
- the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
- a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
- a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
- a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
- a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
- Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radiofrequency, and the like, or any suitable combination of the foregoing.
- Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language and conventional procedural programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- LAN local area network
- WAN wide area network
- Internet Service Provider an Internet Service Provider
- each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures.
- each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams can be implemented by computer program instructions.
- These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Memory System (AREA)
Abstract
Description
- 1. Field of the Invention
- The present invention relates to flash memory device reliability and more particularly to flash memory write cycle bandwidth management.
- 2. Description of the Related Art
- Flash memory refers to non-volatile computer readable storage medium that can be electrically erased and reprogrammed. Flash memory initially had been developed from electrically erasable programmable read-only memory and stores information in an array of memory cells made from floating-gate transistors. In traditional single-level cell devices, each cell of flash memory stores only one bit of information. Newer flash memory, known as multi-level cell devices, can store more than one bit per cell by choosing between multiple levels of electrical charge to apply to the floating gates of its cells.
- Flash cells of a flash memory device can provide only a finite number of erase cycles. When the number of erases in the flash memory device exceeds this limit, the cell fails and eventually, the flash memory device fails. Flash memory devices are also capable of very high performance, which in effect, accelerates the end of life of the flash memory device. Flash memory devices also provide a warranty which guarantee reliable operation for a finite duration of time. Of note, it is well within the performance envelope to exceed the number of erase cycles of a flash memory device within the warranty period. Thus, in order to meet warranty requirements, flash memory devices may throttle write transactions. By slowing down the write transactions, the number of erase transactions is reduced, thus increasing the lifetime of the flash memory device.
- In a virtual environment in which multiple virtual machines (VMs) execute in the physical memory of one or more computing systems, a single flash memory device may be shared by multiple VMs. So much is not a problem so long as the available write bandwidth of the flash memory device exceeds the bandwidth required by all of the VMs. However, as the available bandwidth of the flash memory decreases during write transaction throttling, a resource sharing issue can arise. Today, the limited write bandwidth of a flash memory device is not intelligently partitioned among the VMs. Instead, in a conventional VM-agnostic flash memory device controller, the write bandwidth remains a resource that is freely available to any VM demanding the write access to the flash memory device. This leads to a situation where the most aggressive VM is provided all or most of the available write bandwidth of the flash memory device, potentially leaving other VMs starving for access to the flash memory device.
- Embodiments of the present invention address deficiencies of the art in respect to flash memory device access management in a virtualized environment and provide a novel and non-obvious method, system and computer program product for flash memory device write-access management amongst different VMs in a virtualized computing environment. In an embodiment of the invention, a virtualized computing data processing system can include a host computer with at least one processor and memory and different VMs executing in the host computer. The system also can include a flash memory device coupled to the host computer and accessible by the VMs. Finally, a flash memory controller can manage access to the flash memory device. The controller can include program code enabled to compute a contemporaneous bandwidth of requests for write operations for the flash memory device, to allocate a corresponding number of tokens to the VMs, to accept write requests to the flash memory device from the VMs only when accompanied by a token and to repeat the computing, allocating and accepting after a lapse of a pre-determined time period.
- In one aspect of the embodiment, the program code of the controller further detects the lapse of the predetermined time period, and in response to the detecting of the lapse reallocates unused tokens in over-allocated ones of the VMs to under-allocated ones of the VMs for use during a subsequent pre-determined time period. In another aspect of the embodiment the program code of the controller further identifies ones of the VMs experiencing a burst of requests to write to the flash memory device and allocates in response to the identification of VMs experiencing a burst of requests additional tokens to the VMs experiencing a burst of requests.
- Additional aspects of the invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The aspects of the invention will be realized and attained by means of the elements and combinations particularly pointed out in the appended claims. It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
- The accompanying drawings, which are incorporated in and constitute part of this specification, illustrate embodiments of the invention and together with the description, serve to explain the principles of the invention. The embodiments illustrated herein are presently preferred, it being understood, however, that the invention is not limited to the precise arrangements and instrumentalities shown, wherein:
-
FIG. 1 is a pictorial illustration of a process for flash memory device write-access management amongst different VMs in a virtualized computing environment; -
FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment; -
FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment; and, -
FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment. - Embodiments of the invention provide for flash memory device write-access management amongst different VMs in a virtualized computing environment. In accordance with an embodiment of the invention, a quantity of tokens can be computed corresponding to a measured available write-cycle bandwidth of a flash memory device provided during throttling of the flash memory device. Subsequently, different VMs seeking access to the flash memory device can be allocated an equal number of the quantity of the tokens with which access is granted to institute a write operation on the flash memory device. After an end of a pre-determined period of time, unused tokens allocated to one or more of the VMs can be redistributed to VMs identified as lacking enough tokens to institute all desired write operations and the process can repeat. During the pre-determined period of time, VMs demonstrating a burst of required write access to the flash memory device can be allocated additional tokens on a temporary basis. In this way, write access to the flash memory device can be intelligently and fairly managed in a virtualized environment despite throttling performed by the flash memory device.
- In more particular illustration,
FIG. 1 shows a process for flash memory device write-access management amongst different VMs in a virtualized computing environment. As shown inFIG. 1 ,different VMs virtualized computing environment 100 can request write operations in aflash memory device 130 of thevirtualized computing environment 100. Eachwrite request 120 issued by a corresponding one of theVMs corresponding request queue 150A, 150B, 150N and can be accompanied by atoken device controller 140 can manage access to theflash memory device 130 and can grant arequest 120 issued by aVM request 120 is accompanied by atoken - Of note, the
device controller 140 can maintain a steady state pool oftokens 180A, a free pool oftokens 180B and a burst pool oftokens 180C. The steady state pool oftokens 180A can include a number of tokens equivalent to adetermined bandwidth 170 of theflash memory device 130 in respect to a number of write requests able to be accommodated by theflash memory device 130 irrespective of throttling established by theflash memory device 130. An additional number of tokens can be provided for the burst pool oftokens 180C. Thedevice controller 140 initially can allocate the tokens in the steady state pool oftokens 180A evenly amongst the VMs 110A, 110B, 110N. - Thereafter, when a pre-determined period of time, namely an
epoch 190, has passed,unused tokens VMs tokens 180B for use by ones of theVMs tokens flash memory device 130. During a sub-period of theepoch 190, thequeues 150A, 150B, 150N can be monitored to identify VMs 110A, 110B, 110N demonstrating an enhanced number ofrequests 120 for write operations in theflash memory device 130. In response, tokens from the burst pool oftokens 180C can be provided to those of the VMs 110A, 110B, 110N demonstrating an enhanced number ofrequests 120 for write operations in theflash memory device 130. - The process described in connection with
FIG. 1 can be implemented in a flash memory data processing system. In yet further illustration,FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment. The system can include ahost computer 210 with at least one processor and memory. Thehost computer 210 can support the execution of anoperating system 220 which in turn can manage the operation ofdifferent VMs 260. Aflash device 230 can be coupled to thehost computer 210 and can provide a fixed storage medium for use by the VMs 260. Aflash device controller 250 can be provided to manage access requests by the VMs 260 to theflash device 230 and in particular, write operation requests by the VMs 260. - Notably, the
flash device controller 250 can include program code enabled upon execution by thehost computer 210 to compute a number of tokens for placement in atoken data store 240, to distribute the tokens to theVMs 260 and to permit write operation access to theflash memory device 230 only when such a request by a corresponding one of theVMs 260 is accompanied by a token. In yet further illustration of the operation of theflash device controller 250,FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment. - Beginning in
block 300, the available bandwidth of the flash memory device can be retrieved, for example by contemporaneously monitoring a number of I/O operations or write operations accepted by the flash memory device over a set period of time. Inblock 310, a pool of steady state tokens can be populated in accordance with the bandwidth and inblock 320, a count of a number of VMs active in the virtualized environment can be determined. Thereafter, inblock 330 the tokens in the steady state pool can be allocated evenly across the VMs for use by the VMs whenever a memory write operation is required in respect to the flash memory device. - In
block 340, a delay can be incurred and indecision block 350, it can be determined whether a predetermined period of time—namely an epoch—has transpired. If so, inblock 360, those of the VMs with tokens remaining and unused can be identified as VMs which received an unnecessary over-allocation of tokens and those tokens can be placed into a free pool of tokens inblock 370. Additionally, inblock 380, request queues for write operation requests of the VMs can be inspected to identify VMs which are likely not to have enough tokens to satisfy desired requests for write operations on the flash memory device. Subsequently, inblock 390 tokens from the free pool of tokens can be allocated to those of the VMs determined to have been under-allocated. Finally, the process can repeat inblock 300 with a new determination of the contemporaneous bandwidth of the flash memory device. - At sub-intervals of the epoch, a determination can be made of any VMs requiring tokens from the burst pool of tokens. In more particular illustration,
FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment. Beginning inblock 410, a burst pool of tokens can be initialized to include a selection of tokens and inblock 420, a sub-interval of the epoch can be determined. Inblock 430, a delay can be incurred and indecision block 440, if the sub-interval has elapsed, in block 450 a first VM can be selected for analysis. - In
block 460, the request queue of the selected VM can be inspected to determine the rate of additions of requests to the queue. Indecision block 470, if the rate of request growth in the request queue has exceeded a threshold value, inblock 480 one or more tokens in the burst pool of tokens can be allocated to the selected VM for use only once before being returned to the burst pool of tokens. Thereafter, indecision block 490, if more VMs remain to be analyzed, the process can repeat inblock 450 for a next selected VM. Otherwise, again it can be determined if a next sub-interval has lapsed and the process can begin anew. - As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
- Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
- A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
- Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radiofrequency, and the like, or any suitable combination of the foregoing. Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language and conventional procedural programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- Aspects of the present invention have been described above with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. In this regard, the flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. For instance, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
- It also will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- Finally, the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
- The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
- Having thus described the invention of the present application in detail and by reference to embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of the invention defined in the appended claims as follows:
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/525,017 US20130173849A1 (en) | 2011-12-29 | 2012-06-15 | Write bandwidth management for flashdevices |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/339,685 US9081504B2 (en) | 2011-12-29 | 2011-12-29 | Write bandwidth management for flash devices |
US13/525,017 US20130173849A1 (en) | 2011-12-29 | 2012-06-15 | Write bandwidth management for flashdevices |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/339,685 Continuation US9081504B2 (en) | 2011-12-29 | 2011-12-29 | Write bandwidth management for flash devices |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130173849A1 true US20130173849A1 (en) | 2013-07-04 |
Family
ID=48695897
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/339,685 Active 2033-01-24 US9081504B2 (en) | 2011-12-29 | 2011-12-29 | Write bandwidth management for flash devices |
US13/525,017 Abandoned US20130173849A1 (en) | 2011-12-29 | 2012-06-15 | Write bandwidth management for flashdevices |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/339,685 Active 2033-01-24 US9081504B2 (en) | 2011-12-29 | 2011-12-29 | Write bandwidth management for flash devices |
Country Status (1)
Country | Link |
---|---|
US (2) | US9081504B2 (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140189700A1 (en) * | 2012-12-27 | 2014-07-03 | Advanced Micro Devices, Inc. | Resource management for northbridge using tokens |
US20150033045A1 (en) * | 2013-07-23 | 2015-01-29 | Apple Inc. | Power Supply Droop Reduction Using Feed Forward Current Control |
US20160132358A1 (en) * | 2014-11-06 | 2016-05-12 | Vmware, Inc. | Peripheral device sharing across virtual machines running on different host computing systems |
US20180115626A1 (en) * | 2016-10-21 | 2018-04-26 | Affirmed Networks, Inc. | Adaptive content optimization |
US10228856B2 (en) | 2016-01-28 | 2019-03-12 | Hewlett Packard Enterprise Development Lp | Storage space management in a thin provisioned virtual environment |
US10235297B2 (en) | 2015-11-04 | 2019-03-19 | International Business Machines Corporation | Mechanism for creating friendly transactions with credentials |
US10270775B2 (en) * | 2015-11-04 | 2019-04-23 | International Business Machines Corporation | Mechanism for creating friendly transactions with credentials |
CN110214437A (en) * | 2016-12-07 | 2019-09-06 | 马维尔国际贸易有限公司 | The system and method redistributed for memory access token |
CN110663019A (en) * | 2017-05-26 | 2020-01-07 | 微软技术许可有限责任公司 | File system for Shingled Magnetic Recording (SMR) |
US10965613B2 (en) * | 2019-01-10 | 2021-03-30 | Vmware, Inc. | Multi-pipe bandwidth control in hosted systems |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9111039B2 (en) * | 2012-08-29 | 2015-08-18 | Apple Ii 'c. | Limiting bandwidth for write transactions across networks of components in computer systems |
KR102164617B1 (en) | 2014-03-24 | 2020-10-12 | 삼성전자주식회사 | Method for operating a data storage device and method for operating a system including the same |
CN105988874B (en) | 2015-02-10 | 2020-08-28 | 阿里巴巴集团控股有限公司 | Resource processing method and device |
US10156999B2 (en) | 2016-03-28 | 2018-12-18 | Seagate Technology Llc | Dynamic bandwidth reporting for solid-state drives |
CN113467720B (en) * | 2021-06-29 | 2023-05-23 | 成都佰维存储科技有限公司 | Load balancing method and device, readable storage medium and electronic equipment |
WO2024000443A1 (en) * | 2022-06-30 | 2024-01-04 | Intel Corporation | Enforcement of maximum memory access latency for virtual machine instances |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070022129A1 (en) * | 2005-07-25 | 2007-01-25 | Parascale, Inc. | Rule driven automation of file placement, replication, and migration |
US7711789B1 (en) * | 2007-12-07 | 2010-05-04 | 3 Leaf Systems, Inc. | Quality of service in virtual computing environments |
US20120036239A1 (en) * | 2004-09-10 | 2012-02-09 | Freestyle Technology Pty Ltd | Client processor device for building application files from file fragments for different versions of an application |
US8291170B1 (en) * | 2010-08-19 | 2012-10-16 | Symantec Corporation | System and method for event driven backup data storage |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20060000193A (en) | 2004-06-28 | 2006-01-06 | 정민수 | The method of improving the speed of java card using memory management |
EP1679585A3 (en) | 2005-01-10 | 2009-08-05 | Microsoft Corporation | System and methods for an overlay disk and cache using portable flash memory |
US7917710B2 (en) * | 2006-06-05 | 2011-03-29 | Oracle America, Inc. | Memory protection in a computer system employing memory virtualization |
US20080209076A1 (en) | 2007-02-22 | 2008-08-28 | Wei Wang | Method and system for fast ethernet controller operation using a virtual cpu |
US7917689B2 (en) | 2007-12-28 | 2011-03-29 | Intel Corporation | Methods and apparatuses for nonvolatile memory wear leveling |
US20100083247A1 (en) | 2008-09-26 | 2010-04-01 | Netapp, Inc. | System And Method Of Providing Multiple Virtual Machines With Shared Access To Non-Volatile Solid-State Memory Using RDMA |
US8751738B2 (en) | 2010-02-08 | 2014-06-10 | Microsoft Corporation | Background migration of virtual storage |
-
2011
- 2011-12-29 US US13/339,685 patent/US9081504B2/en active Active
-
2012
- 2012-06-15 US US13/525,017 patent/US20130173849A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120036239A1 (en) * | 2004-09-10 | 2012-02-09 | Freestyle Technology Pty Ltd | Client processor device for building application files from file fragments for different versions of an application |
US20070022129A1 (en) * | 2005-07-25 | 2007-01-25 | Parascale, Inc. | Rule driven automation of file placement, replication, and migration |
US7711789B1 (en) * | 2007-12-07 | 2010-05-04 | 3 Leaf Systems, Inc. | Quality of service in virtual computing environments |
US8291170B1 (en) * | 2010-08-19 | 2012-10-16 | Symantec Corporation | System and method for event driven backup data storage |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9697146B2 (en) * | 2012-12-27 | 2017-07-04 | Advanced Micro Devices, Inc. | Resource management for northbridge using tokens |
US20140189700A1 (en) * | 2012-12-27 | 2014-07-03 | Advanced Micro Devices, Inc. | Resource management for northbridge using tokens |
US20150033045A1 (en) * | 2013-07-23 | 2015-01-29 | Apple Inc. | Power Supply Droop Reduction Using Feed Forward Current Control |
US10067800B2 (en) * | 2014-11-06 | 2018-09-04 | Vmware, Inc. | Peripheral device sharing across virtual machines running on different host computing systems |
US20160132358A1 (en) * | 2014-11-06 | 2016-05-12 | Vmware, Inc. | Peripheral device sharing across virtual machines running on different host computing systems |
US10255189B2 (en) | 2015-11-04 | 2019-04-09 | International Business Machines Corporation | Mechanism for creating friendly transactions with credentials |
US10235297B2 (en) | 2015-11-04 | 2019-03-19 | International Business Machines Corporation | Mechanism for creating friendly transactions with credentials |
US10270775B2 (en) * | 2015-11-04 | 2019-04-23 | International Business Machines Corporation | Mechanism for creating friendly transactions with credentials |
US10270773B2 (en) * | 2015-11-04 | 2019-04-23 | International Business Machines Corporation | Mechanism for creating friendly transactions with credentials |
US10228856B2 (en) | 2016-01-28 | 2019-03-12 | Hewlett Packard Enterprise Development Lp | Storage space management in a thin provisioned virtual environment |
US10129355B2 (en) * | 2016-10-21 | 2018-11-13 | Affirmed Networks, Inc. | Adaptive content optimization |
US20180115626A1 (en) * | 2016-10-21 | 2018-04-26 | Affirmed Networks, Inc. | Adaptive content optimization |
CN110214437A (en) * | 2016-12-07 | 2019-09-06 | 马维尔国际贸易有限公司 | The system and method redistributed for memory access token |
CN110663019A (en) * | 2017-05-26 | 2020-01-07 | 微软技术许可有限责任公司 | File system for Shingled Magnetic Recording (SMR) |
US10965613B2 (en) * | 2019-01-10 | 2021-03-30 | Vmware, Inc. | Multi-pipe bandwidth control in hosted systems |
Also Published As
Publication number | Publication date |
---|---|
US20130173843A1 (en) | 2013-07-04 |
US9081504B2 (en) | 2015-07-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9081504B2 (en) | Write bandwidth management for flash devices | |
US20180150242A1 (en) | Controller and storage device for efficient buffer allocation, and operating method of the storage device | |
US9928169B2 (en) | Method and system for improving swap performance | |
US8402470B2 (en) | Processor thread load balancing manager | |
US8479205B2 (en) | Schedule control program and schedule control method | |
US11340945B2 (en) | Memory congestion aware NUMA management | |
JP2013516716A (en) | Control and discrepancy to limit current spikes | |
KR20190132266A (en) | Memory access operation suspend/resume | |
US11126375B2 (en) | Arbiter circuit for commands from multiple physical functions in a memory sub-system | |
US20170031698A1 (en) | System and method for dynamically allocating resources of virtual machines based on service-level agreements (sla) and privilege levels of users | |
US9286199B2 (en) | Modifying memory space allocation for inactive tasks | |
TWI460659B (en) | Lock windows for reducing contention | |
US20130325543A1 (en) | Ledger-based resource tracking | |
US8244995B2 (en) | System and method for hierarchical wear leveling in storage devices | |
US9069621B2 (en) | Submitting operations to a shared resource based on busy-to-success ratios | |
Min et al. | {eZNS}: An elastic zoned namespace for commodity {ZNS}{SSDs} | |
US10235202B2 (en) | Thread interrupt offload re-prioritization | |
US11808806B2 (en) | Allocation of test resources to perform a test of memory components | |
US20160004473A1 (en) | Migration decision window selection based on hotspot characteristics | |
CN108139983A (en) | For the method and apparatus of the fixed memory page in multilevel system memory | |
Nguyen | Improving smartphone responsiveness through I/O optimizations | |
US11455261B2 (en) | First boot with one memory channel | |
US20220092725A1 (en) | Register compaction with early release | |
US8966133B2 (en) | Determining a mapping mode for a DMA data transfer | |
US9507702B2 (en) | Method of performing write access by distributing control rights to threads, memory controller and flash memory storage device using the same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE. LTD., SINGAPORE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:034194/0111 Effective date: 20140926 Owner name: LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE. LTD., Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:034194/0111 Effective date: 20140926 |
|
STCV | Information on status: appeal procedure |
Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE |