US20130173849A1 - Write bandwidth management for flashdevices - Google Patents

Write bandwidth management for flashdevices Download PDF

Info

Publication number
US20130173849A1
US20130173849A1 US13/525,017 US201213525017A US2013173849A1 US 20130173849 A1 US20130173849 A1 US 20130173849A1 US 201213525017 A US201213525017 A US 201213525017A US 2013173849 A1 US2013173849 A1 US 2013173849A1
Authority
US
United States
Prior art keywords
flash memory
vms
memory device
write
tokens
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/525,017
Inventor
Ganesh Balakrishnan
Anil Krishna
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Enterprise Solutions Singapore Pte Ltd
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US13/525,017 priority Critical patent/US20130173849A1/en
Publication of US20130173849A1 publication Critical patent/US20130173849A1/en
Assigned to LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE. LTD. reassignment LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE. LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERNATIONAL BUSINESS MACHINES CORPORATION
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/061Improving I/O performance
    • G06F3/0613Improving I/O performance in relation to throughput
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0655Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
    • G06F3/0659Command handling arrangements, e.g. command buffers, queues, command scheduling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device
    • G06F3/0679Non-volatile semiconductor memory device, e.g. flash memory, one time programmable memory [OTP]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5011Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals
    • G06F9/5016Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resources being hardware resources other than CPUs, Servers and Terminals the resource being the memory

Definitions

  • the present invention relates to flash memory device reliability and more particularly to flash memory write cycle bandwidth management.
  • Flash memory refers to non-volatile computer readable storage medium that can be electrically erased and reprogrammed. Flash memory initially had been developed from electrically erasable programmable read-only memory and stores information in an array of memory cells made from floating-gate transistors. In traditional single-level cell devices, each cell of flash memory stores only one bit of information. Newer flash memory, known as multi-level cell devices, can store more than one bit per cell by choosing between multiple levels of electrical charge to apply to the floating gates of its cells.
  • Flash cells of a flash memory device can provide only a finite number of erase cycles. When the number of erases in the flash memory device exceeds this limit, the cell fails and eventually, the flash memory device fails. Flash memory devices are also capable of very high performance, which in effect, accelerates the end of life of the flash memory device. Flash memory devices also provide a warranty which guarantee reliable operation for a finite duration of time. Of note, it is well within the performance envelope to exceed the number of erase cycles of a flash memory device within the warranty period. Thus, in order to meet warranty requirements, flash memory devices may throttle write transactions. By slowing down the write transactions, the number of erase transactions is reduced, thus increasing the lifetime of the flash memory device.
  • a single flash memory device may be shared by multiple VMs. So much is not a problem so long as the available write bandwidth of the flash memory device exceeds the bandwidth required by all of the VMs. However, as the available bandwidth of the flash memory decreases during write transaction throttling, a resource sharing issue can arise.
  • the limited write bandwidth of a flash memory device is not intelligently partitioned among the VMs. Instead, in a conventional VM-agnostic flash memory device controller, the write bandwidth remains a resource that is freely available to any VM demanding the write access to the flash memory device. This leads to a situation where the most aggressive VM is provided all or most of the available write bandwidth of the flash memory device, potentially leaving other VMs starving for access to the flash memory device.
  • Embodiments of the present invention address deficiencies of the art in respect to flash memory device access management in a virtualized environment and provide a novel and non-obvious method, system and computer program product for flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • a virtualized computing data processing system can include a host computer with at least one processor and memory and different VMs executing in the host computer.
  • the system also can include a flash memory device coupled to the host computer and accessible by the VMs.
  • a flash memory controller can manage access to the flash memory device.
  • the controller can include program code enabled to compute a contemporaneous bandwidth of requests for write operations for the flash memory device, to allocate a corresponding number of tokens to the VMs, to accept write requests to the flash memory device from the VMs only when accompanied by a token and to repeat the computing, allocating and accepting after a lapse of a pre-determined time period.
  • the program code of the controller further detects the lapse of the predetermined time period, and in response to the detecting of the lapse reallocates unused tokens in over-allocated ones of the VMs to under-allocated ones of the VMs for use during a subsequent pre-determined time period.
  • the program code of the controller further identifies ones of the VMs experiencing a burst of requests to write to the flash memory device and allocates in response to the identification of VMs experiencing a burst of requests additional tokens to the VMs experiencing a burst of requests.
  • FIG. 1 is a pictorial illustration of a process for flash memory device write-access management amongst different VMs in a virtualized computing environment
  • FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment;
  • FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • Embodiments of the invention provide for flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • a quantity of tokens can be computed corresponding to a measured available write-cycle bandwidth of a flash memory device provided during throttling of the flash memory device.
  • different VMs seeking access to the flash memory device can be allocated an equal number of the quantity of the tokens with which access is granted to institute a write operation on the flash memory device.
  • unused tokens allocated to one or more of the VMs can be redistributed to VMs identified as lacking enough tokens to institute all desired write operations and the process can repeat.
  • VMs demonstrating a burst of required write access to the flash memory device can be allocated additional tokens on a temporary basis. In this way, write access to the flash memory device can be intelligently and fairly managed in a virtualized environment despite throttling performed by the flash memory device.
  • FIG. 1 shows a process for flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • different VMs 110 A, 110 B, 110 N of a virtualized computing environment 100 can request write operations in a flash memory device 130 of the virtualized computing environment 100 .
  • Each write request 120 issued by a corresponding one of the VMs 110 A, 110 B, 110 N can be tracked within a corresponding request queue 150 A, 150 B, 150 N and can be accompanied by a token 160 A, 160 B, 160 N.
  • a device controller 140 can manage access to the flash memory device 130 and can grant a request 120 issued by a VM 110 A, 110 B, 110 N only if the request 120 is accompanied by a token 160 A, 160 B, 160 N.
  • the device controller 140 can maintain a steady state pool of tokens 180 A, a free pool of tokens 180 B and a burst pool of tokens 180 C.
  • the steady state pool of tokens 180 A can include a number of tokens equivalent to a determined bandwidth 170 of the flash memory device 130 in respect to a number of write requests able to be accommodated by the flash memory device 130 irrespective of throttling established by the flash memory device 130 .
  • An additional number of tokens can be provided for the burst pool of tokens 180 C.
  • the device controller 140 initially can allocate the tokens in the steady state pool of tokens 180 A evenly amongst the VMs 110 A, 110 B, 110 N.
  • unused tokens 160 A, 160 B, 160 N provided to the VMs 110 A, 110 B, 110 N can be placed in the free pool of tokens 180 B for use by ones of the VMs 110 A, 110 B, 110 N that suffer a deficiency of tokens 160 A, 160 B, 160 N to satisfy desired write operations in the flash memory device 130 .
  • the queues 150 A, 150 B, 150 N can be monitored to identify VMs 110 A, 110 B, 110 N demonstrating an enhanced number of requests 120 for write operations in the flash memory device 130 .
  • tokens from the burst pool of tokens 180 C can be provided to those of the VMs 110 A, 110 B, 110 N demonstrating an enhanced number of requests 120 for write operations in the flash memory device 130 .
  • FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment.
  • the system can include a host computer 210 with at least one processor and memory.
  • the host computer 210 can support the execution of an operating system 220 which in turn can manage the operation of different VMs 260 .
  • a flash device 230 can be coupled to the host computer 210 and can provide a fixed storage medium for use by the VMs 260 .
  • a flash device controller 250 can be provided to manage access requests by the VMs 260 to the flash device 230 and in particular, write operation requests by the VMs 260 .
  • the flash device controller 250 can include program code enabled upon execution by the host computer 210 to compute a number of tokens for placement in a token data store 240 , to distribute the tokens to the VMs 260 and to permit write operation access to the flash memory device 230 only when such a request by a corresponding one of the VMs 260 is accompanied by a token.
  • FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • the available bandwidth of the flash memory device can be retrieved, for example by contemporaneously monitoring a number of I/O operations or write operations accepted by the flash memory device over a set period of time.
  • a pool of steady state tokens can be populated in accordance with the bandwidth and in block 320 , a count of a number of VMs active in the virtualized environment can be determined. Thereafter, in block 330 the tokens in the steady state pool can be allocated evenly across the VMs for use by the VMs whenever a memory write operation is required in respect to the flash memory device.
  • a delay can be incurred and in decision block 350 , it can be determined whether a predetermined period of time—namely an epoch—has transpired. If so, in block 360 , those of the VMs with tokens remaining and unused can be identified as VMs which received an unnecessary over-allocation of tokens and those tokens can be placed into a free pool of tokens in block 370 . Additionally, in block 380 , request queues for write operation requests of the VMs can be inspected to identify VMs which are likely not to have enough tokens to satisfy desired requests for write operations on the flash memory device.
  • tokens from the free pool of tokens can be allocated to those of the VMs determined to have been under-allocated.
  • the process can repeat in block 300 with a new determination of the contemporaneous bandwidth of the flash memory device.
  • FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • a burst pool of tokens can be initialized to include a selection of tokens and in block 420 , a sub-interval of the epoch can be determined.
  • a delay can be incurred and in decision block 440 , if the sub-interval has elapsed, in block 450 a first VM can be selected for analysis.
  • the request queue of the selected VM can be inspected to determine the rate of additions of requests to the queue.
  • decision block 470 if the rate of request growth in the request queue has exceeded a threshold value, in block 480 one or more tokens in the burst pool of tokens can be allocated to the selected VM for use only once before being returned to the burst pool of tokens. Thereafter, in decision block 490 , if more VMs remain to be analyzed, the process can repeat in block 450 for a next selected VM. Otherwise, again it can be determined if a next sub-interval has lapsed and the process can begin anew.
  • aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radiofrequency, and the like, or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language and conventional procedural programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider an Internet Service Provider
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures.
  • each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams can be implemented by computer program instructions.
  • These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Software Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Memory System (AREA)

Abstract

Embodiments of the present invention provide a flash memory device write-access management amongst different virtual machines (VMs) in a virtualized computing environment. In one embodiment, a virtualized computing data processing system can include a host computer with at least one processor and memory and different VMs executing in the host computer. The system also can include a flash memory device coupled to the host computer and accessible by the VMs. Finally, a flash memory controller can manage access to the flash memory device. The controller can include program code enabled to compute a contemporaneous bandwidth of requests for write operations for the flash memory device, to allocate a corresponding number of tokens to the VMs, to accept write requests to the flash memory device from the VMs only when accompanied by a token and to repeat the computing, allocating and accepting after a lapse of a pre-determined time period.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to flash memory device reliability and more particularly to flash memory write cycle bandwidth management.
  • 2. Description of the Related Art
  • Flash memory refers to non-volatile computer readable storage medium that can be electrically erased and reprogrammed. Flash memory initially had been developed from electrically erasable programmable read-only memory and stores information in an array of memory cells made from floating-gate transistors. In traditional single-level cell devices, each cell of flash memory stores only one bit of information. Newer flash memory, known as multi-level cell devices, can store more than one bit per cell by choosing between multiple levels of electrical charge to apply to the floating gates of its cells.
  • Flash cells of a flash memory device can provide only a finite number of erase cycles. When the number of erases in the flash memory device exceeds this limit, the cell fails and eventually, the flash memory device fails. Flash memory devices are also capable of very high performance, which in effect, accelerates the end of life of the flash memory device. Flash memory devices also provide a warranty which guarantee reliable operation for a finite duration of time. Of note, it is well within the performance envelope to exceed the number of erase cycles of a flash memory device within the warranty period. Thus, in order to meet warranty requirements, flash memory devices may throttle write transactions. By slowing down the write transactions, the number of erase transactions is reduced, thus increasing the lifetime of the flash memory device.
  • In a virtual environment in which multiple virtual machines (VMs) execute in the physical memory of one or more computing systems, a single flash memory device may be shared by multiple VMs. So much is not a problem so long as the available write bandwidth of the flash memory device exceeds the bandwidth required by all of the VMs. However, as the available bandwidth of the flash memory decreases during write transaction throttling, a resource sharing issue can arise. Today, the limited write bandwidth of a flash memory device is not intelligently partitioned among the VMs. Instead, in a conventional VM-agnostic flash memory device controller, the write bandwidth remains a resource that is freely available to any VM demanding the write access to the flash memory device. This leads to a situation where the most aggressive VM is provided all or most of the available write bandwidth of the flash memory device, potentially leaving other VMs starving for access to the flash memory device.
  • BRIEF SUMMARY OF THE INVENTION
  • Embodiments of the present invention address deficiencies of the art in respect to flash memory device access management in a virtualized environment and provide a novel and non-obvious method, system and computer program product for flash memory device write-access management amongst different VMs in a virtualized computing environment. In an embodiment of the invention, a virtualized computing data processing system can include a host computer with at least one processor and memory and different VMs executing in the host computer. The system also can include a flash memory device coupled to the host computer and accessible by the VMs. Finally, a flash memory controller can manage access to the flash memory device. The controller can include program code enabled to compute a contemporaneous bandwidth of requests for write operations for the flash memory device, to allocate a corresponding number of tokens to the VMs, to accept write requests to the flash memory device from the VMs only when accompanied by a token and to repeat the computing, allocating and accepting after a lapse of a pre-determined time period.
  • In one aspect of the embodiment, the program code of the controller further detects the lapse of the predetermined time period, and in response to the detecting of the lapse reallocates unused tokens in over-allocated ones of the VMs to under-allocated ones of the VMs for use during a subsequent pre-determined time period. In another aspect of the embodiment the program code of the controller further identifies ones of the VMs experiencing a burst of requests to write to the flash memory device and allocates in response to the identification of VMs experiencing a burst of requests additional tokens to the VMs experiencing a burst of requests.
  • Additional aspects of the invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention. The aspects of the invention will be realized and attained by means of the elements and combinations particularly pointed out in the appended claims. It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
  • The accompanying drawings, which are incorporated in and constitute part of this specification, illustrate embodiments of the invention and together with the description, serve to explain the principles of the invention. The embodiments illustrated herein are presently preferred, it being understood, however, that the invention is not limited to the precise arrangements and instrumentalities shown, wherein:
  • FIG. 1 is a pictorial illustration of a process for flash memory device write-access management amongst different VMs in a virtualized computing environment;
  • FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment;
  • FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment; and,
  • FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • DETAILED DESCRIPTION OF THE INVENTION
  • Embodiments of the invention provide for flash memory device write-access management amongst different VMs in a virtualized computing environment. In accordance with an embodiment of the invention, a quantity of tokens can be computed corresponding to a measured available write-cycle bandwidth of a flash memory device provided during throttling of the flash memory device. Subsequently, different VMs seeking access to the flash memory device can be allocated an equal number of the quantity of the tokens with which access is granted to institute a write operation on the flash memory device. After an end of a pre-determined period of time, unused tokens allocated to one or more of the VMs can be redistributed to VMs identified as lacking enough tokens to institute all desired write operations and the process can repeat. During the pre-determined period of time, VMs demonstrating a burst of required write access to the flash memory device can be allocated additional tokens on a temporary basis. In this way, write access to the flash memory device can be intelligently and fairly managed in a virtualized environment despite throttling performed by the flash memory device.
  • In more particular illustration, FIG. 1 shows a process for flash memory device write-access management amongst different VMs in a virtualized computing environment. As shown in FIG. 1, different VMs 110A, 110B, 110N of a virtualized computing environment 100 can request write operations in a flash memory device 130 of the virtualized computing environment 100. Each write request 120 issued by a corresponding one of the VMs 110A, 110B, 110N can be tracked within a corresponding request queue 150A, 150B, 150N and can be accompanied by a token 160A, 160B, 160N. A device controller 140 can manage access to the flash memory device 130 and can grant a request 120 issued by a VM 110A, 110B, 110N only if the request 120 is accompanied by a token 160A, 160B, 160N.
  • Of note, the device controller 140 can maintain a steady state pool of tokens 180A, a free pool of tokens 180B and a burst pool of tokens 180C. The steady state pool of tokens 180A can include a number of tokens equivalent to a determined bandwidth 170 of the flash memory device 130 in respect to a number of write requests able to be accommodated by the flash memory device 130 irrespective of throttling established by the flash memory device 130. An additional number of tokens can be provided for the burst pool of tokens 180C. The device controller 140 initially can allocate the tokens in the steady state pool of tokens 180A evenly amongst the VMs 110A, 110B, 110N.
  • Thereafter, when a pre-determined period of time, namely an epoch 190, has passed, unused tokens 160A, 160B, 160N provided to the VMs 110A, 110B, 110N can be placed in the free pool of tokens 180B for use by ones of the VMs 110A, 110B, 110N that suffer a deficiency of tokens 160A, 160B, 160N to satisfy desired write operations in the flash memory device 130. During a sub-period of the epoch 190, the queues 150A, 150B, 150N can be monitored to identify VMs 110A, 110B, 110N demonstrating an enhanced number of requests 120 for write operations in the flash memory device 130. In response, tokens from the burst pool of tokens 180C can be provided to those of the VMs 110A, 110B, 110N demonstrating an enhanced number of requests 120 for write operations in the flash memory device 130.
  • The process described in connection with FIG. 1 can be implemented in a flash memory data processing system. In yet further illustration, FIG. 2 is a schematic illustration of a flash memory data processing system configured for write-access management amongst different VMs in a virtualized computing environment. The system can include a host computer 210 with at least one processor and memory. The host computer 210 can support the execution of an operating system 220 which in turn can manage the operation of different VMs 260. A flash device 230 can be coupled to the host computer 210 and can provide a fixed storage medium for use by the VMs 260. A flash device controller 250 can be provided to manage access requests by the VMs 260 to the flash device 230 and in particular, write operation requests by the VMs 260.
  • Notably, the flash device controller 250 can include program code enabled upon execution by the host computer 210 to compute a number of tokens for placement in a token data store 240, to distribute the tokens to the VMs 260 and to permit write operation access to the flash memory device 230 only when such a request by a corresponding one of the VMs 260 is accompanied by a token. In yet further illustration of the operation of the flash device controller 250, FIG. 3 is a flow chart illustrating a process for steady-state flash memory device write-access management amongst different VMs in a virtualized computing environment.
  • Beginning in block 300, the available bandwidth of the flash memory device can be retrieved, for example by contemporaneously monitoring a number of I/O operations or write operations accepted by the flash memory device over a set period of time. In block 310, a pool of steady state tokens can be populated in accordance with the bandwidth and in block 320, a count of a number of VMs active in the virtualized environment can be determined. Thereafter, in block 330 the tokens in the steady state pool can be allocated evenly across the VMs for use by the VMs whenever a memory write operation is required in respect to the flash memory device.
  • In block 340, a delay can be incurred and in decision block 350, it can be determined whether a predetermined period of time—namely an epoch—has transpired. If so, in block 360, those of the VMs with tokens remaining and unused can be identified as VMs which received an unnecessary over-allocation of tokens and those tokens can be placed into a free pool of tokens in block 370. Additionally, in block 380, request queues for write operation requests of the VMs can be inspected to identify VMs which are likely not to have enough tokens to satisfy desired requests for write operations on the flash memory device. Subsequently, in block 390 tokens from the free pool of tokens can be allocated to those of the VMs determined to have been under-allocated. Finally, the process can repeat in block 300 with a new determination of the contemporaneous bandwidth of the flash memory device.
  • At sub-intervals of the epoch, a determination can be made of any VMs requiring tokens from the burst pool of tokens. In more particular illustration, FIG. 4 is a flow chart illustrating a process for burst-state flash memory device write-access management amongst different VMs in a virtualized computing environment. Beginning in block 410, a burst pool of tokens can be initialized to include a selection of tokens and in block 420, a sub-interval of the epoch can be determined. In block 430, a delay can be incurred and in decision block 440, if the sub-interval has elapsed, in block 450 a first VM can be selected for analysis.
  • In block 460, the request queue of the selected VM can be inspected to determine the rate of additions of requests to the queue. In decision block 470, if the rate of request growth in the request queue has exceeded a threshold value, in block 480 one or more tokens in the burst pool of tokens can be allocated to the selected VM for use only once before being returned to the burst pool of tokens. Thereafter, in decision block 490, if more VMs remain to be analyzed, the process can repeat in block 450 for a next selected VM. Otherwise, again it can be determined if a next sub-interval has lapsed and the process can begin anew.
  • As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radiofrequency, and the like, or any suitable combination of the foregoing. Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language and conventional procedural programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • Aspects of the present invention have been described above with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. In this regard, the flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. For instance, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
  • It also will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • Finally, the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
  • The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
  • Having thus described the invention of the present application in detail and by reference to embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of the invention defined in the appended claims as follows:

Claims (4)

1. A flash memory device write-access management method for different virtual machines (VMs) in a virtualized computing environment, the method comprising:
computing a contemporaneous bandwidth of requests for write operations for a flash memory device;
allocating a corresponding number of tokens to virtual machines (VMs) in a virtualized computing environment seeking write access to the flash memory device;
accepting write requests to the flash memory device from VMs only when accompanied by a token; and,
repeating the computing, allocating and accepting after a lapse of a pre-determined time period.
2. The method of claim 1, further comprising:
detecting the lapse of the predetermined time period; and,
re-allocating in response to the detecting of the lapse unused tokens in over-allocated ones of the VMs to under-allocated ones of the VMs for use during a subsequent pre-determined time period.
3. The method of claim 1, further comprising:
identifying ones of the VMs experiencing a burst of requests to write to the flash memory device; and,
allocating in response to the identification of VMs experiencing a burst of requests, additional tokens to the VMs experiencing a burst of requests.
4-9. (canceled)
US13/525,017 2011-12-29 2012-06-15 Write bandwidth management for flashdevices Abandoned US20130173849A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/525,017 US20130173849A1 (en) 2011-12-29 2012-06-15 Write bandwidth management for flashdevices

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/339,685 US9081504B2 (en) 2011-12-29 2011-12-29 Write bandwidth management for flash devices
US13/525,017 US20130173849A1 (en) 2011-12-29 2012-06-15 Write bandwidth management for flashdevices

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/339,685 Continuation US9081504B2 (en) 2011-12-29 2011-12-29 Write bandwidth management for flash devices

Publications (1)

Publication Number Publication Date
US20130173849A1 true US20130173849A1 (en) 2013-07-04

Family

ID=48695897

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/339,685 Active 2033-01-24 US9081504B2 (en) 2011-12-29 2011-12-29 Write bandwidth management for flash devices
US13/525,017 Abandoned US20130173849A1 (en) 2011-12-29 2012-06-15 Write bandwidth management for flashdevices

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/339,685 Active 2033-01-24 US9081504B2 (en) 2011-12-29 2011-12-29 Write bandwidth management for flash devices

Country Status (1)

Country Link
US (2) US9081504B2 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140189700A1 (en) * 2012-12-27 2014-07-03 Advanced Micro Devices, Inc. Resource management for northbridge using tokens
US20150033045A1 (en) * 2013-07-23 2015-01-29 Apple Inc. Power Supply Droop Reduction Using Feed Forward Current Control
US20160132358A1 (en) * 2014-11-06 2016-05-12 Vmware, Inc. Peripheral device sharing across virtual machines running on different host computing systems
US20180115626A1 (en) * 2016-10-21 2018-04-26 Affirmed Networks, Inc. Adaptive content optimization
US10228856B2 (en) 2016-01-28 2019-03-12 Hewlett Packard Enterprise Development Lp Storage space management in a thin provisioned virtual environment
US10235297B2 (en) 2015-11-04 2019-03-19 International Business Machines Corporation Mechanism for creating friendly transactions with credentials
US10270775B2 (en) * 2015-11-04 2019-04-23 International Business Machines Corporation Mechanism for creating friendly transactions with credentials
CN110214437A (en) * 2016-12-07 2019-09-06 马维尔国际贸易有限公司 The system and method redistributed for memory access token
CN110663019A (en) * 2017-05-26 2020-01-07 微软技术许可有限责任公司 File system for Shingled Magnetic Recording (SMR)
US10965613B2 (en) * 2019-01-10 2021-03-30 Vmware, Inc. Multi-pipe bandwidth control in hosted systems

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9111039B2 (en) * 2012-08-29 2015-08-18 Apple Ii 'c. Limiting bandwidth for write transactions across networks of components in computer systems
KR102164617B1 (en) 2014-03-24 2020-10-12 삼성전자주식회사 Method for operating a data storage device and method for operating a system including the same
CN105988874B (en) 2015-02-10 2020-08-28 阿里巴巴集团控股有限公司 Resource processing method and device
US10156999B2 (en) 2016-03-28 2018-12-18 Seagate Technology Llc Dynamic bandwidth reporting for solid-state drives
CN113467720B (en) * 2021-06-29 2023-05-23 成都佰维存储科技有限公司 Load balancing method and device, readable storage medium and electronic equipment
WO2024000443A1 (en) * 2022-06-30 2024-01-04 Intel Corporation Enforcement of maximum memory access latency for virtual machine instances

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070022129A1 (en) * 2005-07-25 2007-01-25 Parascale, Inc. Rule driven automation of file placement, replication, and migration
US7711789B1 (en) * 2007-12-07 2010-05-04 3 Leaf Systems, Inc. Quality of service in virtual computing environments
US20120036239A1 (en) * 2004-09-10 2012-02-09 Freestyle Technology Pty Ltd Client processor device for building application files from file fragments for different versions of an application
US8291170B1 (en) * 2010-08-19 2012-10-16 Symantec Corporation System and method for event driven backup data storage

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20060000193A (en) 2004-06-28 2006-01-06 정민수 The method of improving the speed of java card using memory management
EP1679585A3 (en) 2005-01-10 2009-08-05 Microsoft Corporation System and methods for an overlay disk and cache using portable flash memory
US7917710B2 (en) * 2006-06-05 2011-03-29 Oracle America, Inc. Memory protection in a computer system employing memory virtualization
US20080209076A1 (en) 2007-02-22 2008-08-28 Wei Wang Method and system for fast ethernet controller operation using a virtual cpu
US7917689B2 (en) 2007-12-28 2011-03-29 Intel Corporation Methods and apparatuses for nonvolatile memory wear leveling
US20100083247A1 (en) 2008-09-26 2010-04-01 Netapp, Inc. System And Method Of Providing Multiple Virtual Machines With Shared Access To Non-Volatile Solid-State Memory Using RDMA
US8751738B2 (en) 2010-02-08 2014-06-10 Microsoft Corporation Background migration of virtual storage

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120036239A1 (en) * 2004-09-10 2012-02-09 Freestyle Technology Pty Ltd Client processor device for building application files from file fragments for different versions of an application
US20070022129A1 (en) * 2005-07-25 2007-01-25 Parascale, Inc. Rule driven automation of file placement, replication, and migration
US7711789B1 (en) * 2007-12-07 2010-05-04 3 Leaf Systems, Inc. Quality of service in virtual computing environments
US8291170B1 (en) * 2010-08-19 2012-10-16 Symantec Corporation System and method for event driven backup data storage

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9697146B2 (en) * 2012-12-27 2017-07-04 Advanced Micro Devices, Inc. Resource management for northbridge using tokens
US20140189700A1 (en) * 2012-12-27 2014-07-03 Advanced Micro Devices, Inc. Resource management for northbridge using tokens
US20150033045A1 (en) * 2013-07-23 2015-01-29 Apple Inc. Power Supply Droop Reduction Using Feed Forward Current Control
US10067800B2 (en) * 2014-11-06 2018-09-04 Vmware, Inc. Peripheral device sharing across virtual machines running on different host computing systems
US20160132358A1 (en) * 2014-11-06 2016-05-12 Vmware, Inc. Peripheral device sharing across virtual machines running on different host computing systems
US10255189B2 (en) 2015-11-04 2019-04-09 International Business Machines Corporation Mechanism for creating friendly transactions with credentials
US10235297B2 (en) 2015-11-04 2019-03-19 International Business Machines Corporation Mechanism for creating friendly transactions with credentials
US10270775B2 (en) * 2015-11-04 2019-04-23 International Business Machines Corporation Mechanism for creating friendly transactions with credentials
US10270773B2 (en) * 2015-11-04 2019-04-23 International Business Machines Corporation Mechanism for creating friendly transactions with credentials
US10228856B2 (en) 2016-01-28 2019-03-12 Hewlett Packard Enterprise Development Lp Storage space management in a thin provisioned virtual environment
US10129355B2 (en) * 2016-10-21 2018-11-13 Affirmed Networks, Inc. Adaptive content optimization
US20180115626A1 (en) * 2016-10-21 2018-04-26 Affirmed Networks, Inc. Adaptive content optimization
CN110214437A (en) * 2016-12-07 2019-09-06 马维尔国际贸易有限公司 The system and method redistributed for memory access token
CN110663019A (en) * 2017-05-26 2020-01-07 微软技术许可有限责任公司 File system for Shingled Magnetic Recording (SMR)
US10965613B2 (en) * 2019-01-10 2021-03-30 Vmware, Inc. Multi-pipe bandwidth control in hosted systems

Also Published As

Publication number Publication date
US20130173843A1 (en) 2013-07-04
US9081504B2 (en) 2015-07-14

Similar Documents

Publication Publication Date Title
US9081504B2 (en) Write bandwidth management for flash devices
US20180150242A1 (en) Controller and storage device for efficient buffer allocation, and operating method of the storage device
US9928169B2 (en) Method and system for improving swap performance
US8402470B2 (en) Processor thread load balancing manager
US8479205B2 (en) Schedule control program and schedule control method
US11340945B2 (en) Memory congestion aware NUMA management
JP2013516716A (en) Control and discrepancy to limit current spikes
KR20190132266A (en) Memory access operation suspend/resume
US11126375B2 (en) Arbiter circuit for commands from multiple physical functions in a memory sub-system
US20170031698A1 (en) System and method for dynamically allocating resources of virtual machines based on service-level agreements (sla) and privilege levels of users
US9286199B2 (en) Modifying memory space allocation for inactive tasks
TWI460659B (en) Lock windows for reducing contention
US20130325543A1 (en) Ledger-based resource tracking
US8244995B2 (en) System and method for hierarchical wear leveling in storage devices
US9069621B2 (en) Submitting operations to a shared resource based on busy-to-success ratios
Min et al. {eZNS}: An elastic zoned namespace for commodity {ZNS}{SSDs}
US10235202B2 (en) Thread interrupt offload re-prioritization
US11808806B2 (en) Allocation of test resources to perform a test of memory components
US20160004473A1 (en) Migration decision window selection based on hotspot characteristics
CN108139983A (en) For the method and apparatus of the fixed memory page in multilevel system memory
Nguyen Improving smartphone responsiveness through I/O optimizations
US11455261B2 (en) First boot with one memory channel
US20220092725A1 (en) Register compaction with early release
US8966133B2 (en) Determining a mapping mode for a DMA data transfer
US9507702B2 (en) Method of performing write access by distributing control rights to threads, memory controller and flash memory storage device using the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE. LTD., SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:034194/0111

Effective date: 20140926

Owner name: LENOVO ENTERPRISE SOLUTIONS (SINGAPORE) PTE. LTD.,

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:034194/0111

Effective date: 20140926

STCV Information on status: appeal procedure

Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE