US11592986B2 - Methods for minimizing fragmentation in SSD within a storage system and devices thereof - Google Patents

Methods for minimizing fragmentation in SSD within a storage system and devices thereof Download PDF

Info

Publication number
US11592986B2
US11592986B2 US17/486,160 US202117486160A US11592986B2 US 11592986 B2 US11592986 B2 US 11592986B2 US 202117486160 A US202117486160 A US 202117486160A US 11592986 B2 US11592986 B2 US 11592986B2
Authority
US
United States
Prior art keywords
allocation area
data
cache
computing device
address range
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US17/486,160
Other versions
US20220083232A1 (en
Inventor
Ravikanth Dronamraju
Shivali Gupta
Kyle Sterling
Atul Goel
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NetApp Inc
Original Assignee
NetApp Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NetApp Inc filed Critical NetApp Inc
Priority to US17/486,160 priority Critical patent/US11592986B2/en
Publication of US20220083232A1 publication Critical patent/US20220083232A1/en
Application granted granted Critical
Publication of US11592986B2 publication Critical patent/US11592986B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0608Saving storage space on storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/023Free address space management
    • G06F12/0238Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory
    • G06F12/0246Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory in block erasable memory, e.g. flash memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0866Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches for peripheral storage systems, e.g. disk cache
    • G06F12/0868Data transfer between cache memory and other subsystems, e.g. storage devices or host systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0629Configuration or reconfiguration of storage systems
    • G06F3/0631Configuration or reconfiguration of storage systems by allocating resources to storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0638Organizing or formatting or addressing of data
    • G06F3/064Management of blocks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device
    • G06F3/0679Non-volatile semiconductor memory device, e.g. flash memory, one time programmable memory [OTP]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1041Resource optimization
    • G06F2212/1044Space efficiency improvement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/21Employing a record carrier using a specific recording technology
    • G06F2212/214Solid state disk
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/28Using a specific disk cache architecture
    • G06F2212/281Single cache

Definitions

  • This technology generally relates to data storage management and, more particularly, methods for minimizing fragmentation in solid state devices (SSD) within a storage system and devices thereof.
  • SSD solid state devices
  • Fragmentation within the data storage mediums phenomenon reduces storage capacity, and increases access and read times. Fragmentation occurs when a computing device does not allocate enough contiguous storage space to encode a complete data set as a single sequence, and instead divides the data set into portions which are encoded in gaps between other data sets. Typically, these data storage medium gaps arise because an earlier data set was deleted, or excess space was allocated to another stored data set. The degree to which a data storage medium is fragmented heavily dependent upon the size and number of data sets stored on it.
  • prior technologies uses the technique of defragmentation where the datasets written to a data storage medium is rearranged such that segments of each data set are written contiguously.
  • the technique used in the prior technologies fails to significantly reduce fragmentation and are time consuming, memory intensive operations that can significantly reduce the performance of a computer system while being executed.
  • FIG. 1 is a block diagram of an environment with a storage management computing device that minimizes fragmentation in SSD within a storage system
  • FIG. 2 is a block diagram of the exemplary storage management computing device shown in FIG. 1 ;
  • FIG. 3 is an exemplary flow chart of an example of a method for minimizing fragmentation in SSD within a storage system.
  • FIG. 1 An environment 10 with a plurality of client computing devices 12 ( 1 )- 12 ( n ), an exemplary storage management computing device 14 , a plurality of storage devices 16 ( 1 )- 16 ( n ) is illustrated in FIG. 1 .
  • the environment 10 in FIG. 1 includes the plurality of client computing devices 12 ( 1 )- 12 ( n ), the storage management computing device 14 and a plurality of storage devices 16 ( 1 )- 16 ( n ) coupled via one or more communication networks 30 , although the environment could include other types and numbers of systems, devices, components, and/or other elements.
  • the example of a method for minimizing fragmentation in SSD within a storage system is executed by the storage management computing device 14 , although the approaches illustrated and described herein could be executed by other types and/or numbers of other computing systems and devices.
  • the environment 10 may include other types and numbers of other network elements and devices, as is generally known in the art and will not be illustrated or described herein. This technology provides a number of advantages including providing methods, non-transitory computer readable media and devices for minimizing fragmentation in SSD within a storage system.
  • the storage management computing device 14 includes a processor 18 , a memory 20 , and a communication interface 24 which are coupled together by a bus 26 , although the storage management computing device 14 may include other types and numbers of elements in other configurations.
  • the processor 18 of the storage management computing device 14 may execute one or more programmed instructions stored in the memory 20 for dynamic resource reservation based on classified input/output requests as illustrated and described in the examples herein, although other types and numbers of functions and/or other operation can be performed.
  • the processor 18 of the storage management computing device 14 may include one or more central processing units (“CPUs”) or general purpose processors with one or more processing cores, such as AMD® processor(s), although other types of processor(s) could be used (e.g., Intel®).
  • the memory 20 of the storage management computing device 14 stores the programmed instructions and other data for one or more aspects of the present technology as described and illustrated herein, although some or all of the programmed instructions could be stored and executed elsewhere.
  • a variety of different types of memory storage devices such as a non-volatile memory, random access memory (RAM) or a read only memory (ROM) in the system or a floppy disk, hard disk, CD ROM, DVD ROM, or other computer readable medium which is read from and written to by a magnetic, optical, or other reading and writing system that is coupled to the processor 18 , can be used for the memory 20 .
  • the communication interface 24 of the storage management computing device 14 operatively couples and communicates with the plurality of client computing devices 12 ( 1 )- 12 ( n ) and the plurality of storage devices 16 ( 1 )- 16 ( n ), which are all coupled together by the communication network 30 , although other types and numbers of communication networks or systems with other types and numbers of connections and configurations to other devices and elements.
  • the communication network 30 can use TCP/IP over Ethernet and industry-standard protocols, including NFS, CIFS, SOAP, XML, LDAP, and SNMP, although other types and numbers of communication networks, can be used.
  • the communication networks 30 in this example may employ any suitable interface mechanisms and network communication technologies, including, for example, any local area network, any wide area network (e.g., Internet), teletraffic in any suitable form (e.g., voice, modem, and the like), Public Switched Telephone Network (PSTNs), Ethernet-based Packet Data Networks (PDNs), and any combinations thereof and the like.
  • the bus 26 is a universal serial bus, although other bus types and links may be used, such as PCI-Express or hyper-transport bus.
  • Each of the plurality of client computing devices 12 ( 1 )- 12 ( n ) includes a central processing unit (CPU) or processor, a memory, and an I/O system, which are coupled together by a bus or other link, although other numbers and types of network devices could be used.
  • the plurality of client computing devices 12 ( 1 )- 12 ( n ) communicates with the storage management computing device 14 for storage management, although the client computing devices 12 ( 1 )- 12 ( n ) can interact with the storage management computing device 14 for other purposes.
  • the plurality of client computing devices 12 ( 1 )- 12 ( n ) may run application(s) that may provide an interface to make requests to access, modify, delete, edit, read or write data within storage management computing device 14 or the plurality of storage devices 16 ( 1 )- 16 ( n ) via the communication network 30 .
  • Each of the plurality of storage devices 16 ( 1 )- 16 ( n ) includes a central processing unit (CPU) or processor, and an I/O system, which are coupled together by a bus or other link, although other numbers and types of network devices could be used.
  • Each plurality of storage devices 16 ( 1 )- 16 ( n ) assists with storing data, although the plurality of storage devices 16 ( 1 )- 16 ( n ) can assist with other types of operations such as storing of files or data.
  • Various network processing applications such as CIFS applications, NFS applications, HTTP Web Data storage device applications, and/or FTP applications, may be operating on the plurality of storage devices 16 ( 1 )- 16 ( n ) and transmitting data (e.g., files or web pages) in response to requests from the storage management computing device 14 and the plurality of client computing devices 12 ( 1 )- 12 ( n ).
  • data e.g., files or web pages
  • the plurality of storage devices 16 ( 1 )- 16 ( n ) may be hardware such as hard disk drives, solid state devices (SSD), or magnetic tapes, or software or may represent a system with multiple external resource servers, which may include internal or external networks.
  • the plurality of storage devices 16 ( 1 )- 16 ( n ) are outside the storage management computing device 14
  • the plurality of storage devices 16 ( 1 )- 16 ( n ) can reside within the storage management computing device 14 in other examples.
  • the exemplary network environment 10 includes the plurality of client computing devices 12 ( 1 )- 12 ( n ), the storage management computing device 14 , and the plurality of storage devices 16 ( 1 )- 16 ( n ) described and illustrated herein, other types and numbers of systems, devices, components, and/or other elements in other topologies can be used. It is to be understood that the systems of the examples described herein are for exemplary purposes, as many variations of the specific hardware and software used to implement the examples are possible, as will be appreciated by those of ordinary skill in the art.
  • two or more computing systems or devices can be substituted for any one of the systems or devices in any example. Accordingly, principles and advantages of distributed processing, such as redundancy and replication also can be implemented, as desired, to increase the robustness and performance of the devices and systems of the examples.
  • the examples may also be implemented on computer system(s) that extend across any suitable network using any suitable interface mechanisms and traffic technologies, including by way of example only teletraffic in any suitable form (e.g., voice and modem), wireless traffic media, wireless traffic networks, cellular traffic networks, G3 traffic networks, Public Switched Telephone Network (PSTNs), Packet Data Networks (PDNs), the Internet, intranets, and combinations thereof.
  • PSTNs Public Switched Telephone Network
  • PDNs Packet Data Networks
  • the Internet intranets, and combinations thereof.
  • the examples also may be embodied as a non-transitory computer readable medium having instructions stored thereon for one or more aspects of the present technology as described and illustrated by way of the examples herein, as described herein, which when executed by the processor, cause the processor to carry out the steps necessary to implement the methods of this technology as described and illustrated with the examples herein.
  • the exemplary method begins at step 305 where the storage management computing device 14 receives a data operation request, such as a read request or a write request by way of example, from one of the plurality of client computing devices 12 ( 1 )- 12 ( n ), although the storage management computing device 14 can receive other types and/or numbers of requests from other devices.
  • a data operation request such as a read request or a write request by way of example, from one of the plurality of client computing devices 12 ( 1 )- 12 ( n ), although the storage management computing device 14 can receive other types and/or numbers of requests from other devices.
  • the received data operation request includes a write request on an object, although the storage management computing device 14 can receive other types and/or numbers of data operation requests from other devices.
  • step 310 the storage management computing device 14 caches the received write data operation into the cache within memory 20 of the storage management computing device 14 , although the storage management computing device 14 can cache or otherwise store the operations at other memory locations.
  • the storage management computing device 14 may determine when the cache within memory 20 has reached a threshold size.
  • the cache within the memory 20 has a stored or otherwise set threshold size after which the cache cannot receive further data. Accordingly, when the storage management computing device 14 determines that the cache has not yet reached the threshold size, then the No branch is taken back to step 305 where the storage management computing device 14 continues to receive the data operation request from the plurality of client computing devices 12 ( 1 )- 12 ( n ).
  • step 320 the storage management computing device 14 identifies an allocation area in one of the plurality of storage devices 16 ( 1 )- 16 ( n ) which is a solid state drive (SSD), although the storage management computing device 14 can identify the allocation area at other memory locations, such as in memory 20 .
  • allocation area relates to the memory space in the SSD where the contents from the cache could be transferred.
  • the storage management computing device 14 includes metadata stored in the memory 20 that includes the address range in the memory strip within the SSD that can stage the data being flushed out of the cache within the memory 20 .
  • the storage management computing device 14 may use the metadata to identify the allocation area, although the storage management computing device 14 can use other types and/or amounts of information to identify the allocation area.
  • the size allocation area identified by the storage management computing device 14 is at least equal to the size of the data in the cache that is going to be transferred, although the size of the allocation area can be more than the size of the data being transferred.
  • the address range of the identified allocation area can include a combination of free space and previously stored data (due to fragmented data that was previously stored) in this example.
  • step 325 the storage management computing device 14 determines or identifies the amount of data in the cache that can be transferred to the identified allocation area sequentially from the starting of the address range of the identified allocation area to the end of the address range of the identified allocation area.
  • step 330 during the traversing the storage management computing device 14 determines when there is data stored in the address range of the identified allocation area, although the storage management computing device 14 can use other techniques to make the determination.
  • the allocation area identified by the storage management computing device 14 can include data that was previously transferred from the cache within memory 20 , although the identified allocation area can include other types and/or amounts of data. Accordingly, when the storage management computing device 14 determines that the data is present in the address range, then the Yes branch is taken in to step 335 .
  • step 335 the storage management computing device 14 reads the previously stored data in the address range of the allocation area.
  • the storage management computing device 14 is required to read the previously stored data within the address range in the allocation area as the data has to be first read before writing the data back into the SSD.
  • step 340 the storage management computing device 14 combines the read data from step 335 with the cache data that is required to be transferred to the allocation area, although the storage management computing device 14 can add other types and/or amounts of information.
  • step 345 the storage management computing device 14 writes or transfers the data present in the cache within the memory 20 and the previously stored data that was read back into the identified allocation area within the SSD. Additionally back in step 330 , when the storage management computing device 14 determines that there was no data found during the traversing, then the No branch is taken to this step 345 .
  • the storage management computing device 14 only writes or transfers the data from the cache within memory 20 to the identified allocation area.
  • the storage management computing device 14 begins the write or transfers the data from the start of the address range of the identified allocation and proceeds to write or transfer the data sequentially in the address range.
  • the technology is able to reduce the fragmentation in the SSD within the plurality of storage devices 16 ( 1 )- 16 ( n ) and the exemplary method ends at step 350 .
  • this technology provides a number of advantages including providing methods, non-transitory computer readable media and devices for minimizing fragmentation in SSD within an enterprise storage system.
  • the technology By writing the data sequentially into the allocation area in the SSD, the technology able to significantly reduce the fragmentation in SSD thereby increases the memory utilization of the SSD.

Abstract

A method, non-transitory computer readable medium, and device that assists with reducing memory fragmentation in solid state devices includes identifying an allocation area within an address range to write data from a cache. Next, the identified allocation area is determined for including previously stored data. The previously stored data is read from the identified allocation area when it is determined that the identified allocation area comprises previously stored data. Next, both the write data from the cache and the read previously stored data are written back into the identified allocation area sequentially through the address range.

Description

FIELD
This application claims priority to and is a continuation of U.S. application Ser. No. 16/584,025, filed Sep. 26, 2019, now allowed, and titled “METHODS FOR MINIMIZING FRAGMENTATION IN SSD WITHIN A STORAGE SYSTEM AND DEVICES THEREOF,” which claims priority to and is a continuation of U.S. Pat. No. 10,430,081, filed Jun. 28, 2016, and titled “METHODS FOR MINIMIZING FRAGMENTATION IN SSD WITHIN A STORAGE SYSTEM AND DEVICES THEREOF,” which is incorporated herein by reference.
This technology generally relates to data storage management and, more particularly, methods for minimizing fragmentation in solid state devices (SSD) within a storage system and devices thereof.
BACKGROUND
Repeated writing data sets to, and erasing data sets from, a data storage medium causes storage space to be used inefficiently and this phenomenon is called fragmentation within the data storage medium. Fragmentation within the data storage mediums phenomenon reduces storage capacity, and increases access and read times. Fragmentation occurs when a computing device does not allocate enough contiguous storage space to encode a complete data set as a single sequence, and instead divides the data set into portions which are encoded in gaps between other data sets. Typically, these data storage medium gaps arise because an earlier data set was deleted, or excess space was allocated to another stored data set. The degree to which a data storage medium is fragmented heavily dependent upon the size and number of data sets stored on it.
To address the fragmentation issue, prior technologies uses the technique of defragmentation where the datasets written to a data storage medium is rearranged such that segments of each data set are written contiguously. Unfortunately, the technique used in the prior technologies fails to significantly reduce fragmentation and are time consuming, memory intensive operations that can significantly reduce the performance of a computer system while being executed.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of an environment with a storage management computing device that minimizes fragmentation in SSD within a storage system;
FIG. 2 is a block diagram of the exemplary storage management computing device shown in FIG. 1 ; and
FIG. 3 is an exemplary flow chart of an example of a method for minimizing fragmentation in SSD within a storage system.
DETAILED DESCRIPTION
An environment 10 with a plurality of client computing devices 12(1)-12(n), an exemplary storage management computing device 14, a plurality of storage devices 16(1)-16(n) is illustrated in FIG. 1 . In this particular example, the environment 10 in FIG. 1 includes the plurality of client computing devices 12(1)-12(n), the storage management computing device 14 and a plurality of storage devices 16(1)-16(n) coupled via one or more communication networks 30, although the environment could include other types and numbers of systems, devices, components, and/or other elements. The example of a method for minimizing fragmentation in SSD within a storage system is executed by the storage management computing device 14, although the approaches illustrated and described herein could be executed by other types and/or numbers of other computing systems and devices. The environment 10 may include other types and numbers of other network elements and devices, as is generally known in the art and will not be illustrated or described herein. This technology provides a number of advantages including providing methods, non-transitory computer readable media and devices for minimizing fragmentation in SSD within a storage system.
Referring to FIG. 2 , in this example the storage management computing device 14 includes a processor 18, a memory 20, and a communication interface 24 which are coupled together by a bus 26, although the storage management computing device 14 may include other types and numbers of elements in other configurations.
The processor 18 of the storage management computing device 14 may execute one or more programmed instructions stored in the memory 20 for dynamic resource reservation based on classified input/output requests as illustrated and described in the examples herein, although other types and numbers of functions and/or other operation can be performed. The processor 18 of the storage management computing device 14 may include one or more central processing units (“CPUs”) or general purpose processors with one or more processing cores, such as AMD® processor(s), although other types of processor(s) could be used (e.g., Intel®).
The memory 20 of the storage management computing device 14 stores the programmed instructions and other data for one or more aspects of the present technology as described and illustrated herein, although some or all of the programmed instructions could be stored and executed elsewhere. A variety of different types of memory storage devices, such as a non-volatile memory, random access memory (RAM) or a read only memory (ROM) in the system or a floppy disk, hard disk, CD ROM, DVD ROM, or other computer readable medium which is read from and written to by a magnetic, optical, or other reading and writing system that is coupled to the processor 18, can be used for the memory 20.
The communication interface 24 of the storage management computing device 14 operatively couples and communicates with the plurality of client computing devices 12(1)-12(n) and the plurality of storage devices 16(1)-16(n), which are all coupled together by the communication network 30, although other types and numbers of communication networks or systems with other types and numbers of connections and configurations to other devices and elements. By way of example only, the communication network 30 can use TCP/IP over Ethernet and industry-standard protocols, including NFS, CIFS, SOAP, XML, LDAP, and SNMP, although other types and numbers of communication networks, can be used. The communication networks 30 in this example may employ any suitable interface mechanisms and network communication technologies, including, for example, any local area network, any wide area network (e.g., Internet), teletraffic in any suitable form (e.g., voice, modem, and the like), Public Switched Telephone Network (PSTNs), Ethernet-based Packet Data Networks (PDNs), and any combinations thereof and the like. In this example, the bus 26 is a universal serial bus, although other bus types and links may be used, such as PCI-Express or hyper-transport bus.
Each of the plurality of client computing devices 12(1)-12(n) includes a central processing unit (CPU) or processor, a memory, and an I/O system, which are coupled together by a bus or other link, although other numbers and types of network devices could be used. The plurality of client computing devices 12(1)-12(n) communicates with the storage management computing device 14 for storage management, although the client computing devices 12(1)-12(n) can interact with the storage management computing device 14 for other purposes. By way of example, the plurality of client computing devices 12(1)-12(n) may run application(s) that may provide an interface to make requests to access, modify, delete, edit, read or write data within storage management computing device 14 or the plurality of storage devices 16(1)-16(n) via the communication network 30.
Each of the plurality of storage devices 16(1)-16(n) includes a central processing unit (CPU) or processor, and an I/O system, which are coupled together by a bus or other link, although other numbers and types of network devices could be used. Each plurality of storage devices 16(1)-16(n) assists with storing data, although the plurality of storage devices 16(1)-16(n) can assist with other types of operations such as storing of files or data. Various network processing applications, such as CIFS applications, NFS applications, HTTP Web Data storage device applications, and/or FTP applications, may be operating on the plurality of storage devices 16(1)-16(n) and transmitting data (e.g., files or web pages) in response to requests from the storage management computing device 14 and the plurality of client computing devices 12(1)-12(n). It is to be understood that the plurality of storage devices 16(1)-16(n) may be hardware such as hard disk drives, solid state devices (SSD), or magnetic tapes, or software or may represent a system with multiple external resource servers, which may include internal or external networks. While in this example it is illustrated that the plurality of storage devices 16(1)-16(n) are outside the storage management computing device 14, the plurality of storage devices 16(1)-16(n) can reside within the storage management computing device 14 in other examples.
Although the exemplary network environment 10 includes the plurality of client computing devices 12(1)-12(n), the storage management computing device 14, and the plurality of storage devices 16(1)-16(n) described and illustrated herein, other types and numbers of systems, devices, components, and/or other elements in other topologies can be used. It is to be understood that the systems of the examples described herein are for exemplary purposes, as many variations of the specific hardware and software used to implement the examples are possible, as will be appreciated by those of ordinary skill in the art.
In addition, two or more computing systems or devices can be substituted for any one of the systems or devices in any example. Accordingly, principles and advantages of distributed processing, such as redundancy and replication also can be implemented, as desired, to increase the robustness and performance of the devices and systems of the examples. The examples may also be implemented on computer system(s) that extend across any suitable network using any suitable interface mechanisms and traffic technologies, including by way of example only teletraffic in any suitable form (e.g., voice and modem), wireless traffic media, wireless traffic networks, cellular traffic networks, G3 traffic networks, Public Switched Telephone Network (PSTNs), Packet Data Networks (PDNs), the Internet, intranets, and combinations thereof.
The examples also may be embodied as a non-transitory computer readable medium having instructions stored thereon for one or more aspects of the present technology as described and illustrated by way of the examples herein, as described herein, which when executed by the processor, cause the processor to carry out the steps necessary to implement the methods of this technology as described and illustrated with the examples herein.
An example of a method for minimizing fragmentation in SSD within a storage system will now be described herein with reference to FIGS. 1-3 . The exemplary method begins at step 305 where the storage management computing device 14 receives a data operation request, such as a read request or a write request by way of example, from one of the plurality of client computing devices 12(1)-12(n), although the storage management computing device 14 can receive other types and/or numbers of requests from other devices. In this particular example, the received data operation request includes a write request on an object, although the storage management computing device 14 can receive other types and/or numbers of data operation requests from other devices.
Next in step 310, the storage management computing device 14 caches the received write data operation into the cache within memory 20 of the storage management computing device 14, although the storage management computing device 14 can cache or otherwise store the operations at other memory locations.
Next in step 315, upon caching the received write data operation the storage management computing device 14 may determine when the cache within memory 20 has reached a threshold size. In this particular example, the cache within the memory 20 has a stored or otherwise set threshold size after which the cache cannot receive further data. Accordingly, when the storage management computing device 14 determines that the cache has not yet reached the threshold size, then the No branch is taken back to step 305 where the storage management computing device 14 continues to receive the data operation request from the plurality of client computing devices 12(1)-12(n).
However, if back in step 315 when the storage management computing device 14 determines that the cache is full, then the Yes branch is taken to step 320. In step 320, the storage management computing device 14 identifies an allocation area in one of the plurality of storage devices 16(1)-16(n) which is a solid state drive (SSD), although the storage management computing device 14 can identify the allocation area at other memory locations, such as in memory 20. In this example, allocation area relates to the memory space in the SSD where the contents from the cache could be transferred. Additionally in this example, the storage management computing device 14 includes metadata stored in the memory 20 that includes the address range in the memory strip within the SSD that can stage the data being flushed out of the cache within the memory 20. The storage management computing device 14 may use the metadata to identify the allocation area, although the storage management computing device 14 can use other types and/or amounts of information to identify the allocation area. Further in this example, the size allocation area identified by the storage management computing device 14 is at least equal to the size of the data in the cache that is going to be transferred, although the size of the allocation area can be more than the size of the data being transferred. Furthermore, the address range of the identified allocation area can include a combination of free space and previously stored data (due to fragmented data that was previously stored) in this example.
Next in step 325, the storage management computing device 14 determines or identifies the amount of data in the cache that can be transferred to the identified allocation area sequentially from the starting of the address range of the identified allocation area to the end of the address range of the identified allocation area.
In step 330, during the traversing the storage management computing device 14 determines when there is data stored in the address range of the identified allocation area, although the storage management computing device 14 can use other techniques to make the determination. In this example, the allocation area identified by the storage management computing device 14 can include data that was previously transferred from the cache within memory 20, although the identified allocation area can include other types and/or amounts of data. Accordingly, when the storage management computing device 14 determines that the data is present in the address range, then the Yes branch is taken in to step 335.
In step 335, the storage management computing device 14 reads the previously stored data in the address range of the allocation area. In this example, the storage management computing device 14 is required to read the previously stored data within the address range in the allocation area as the data has to be first read before writing the data back into the SSD.
Next in step 340, the storage management computing device 14 combines the read data from step 335 with the cache data that is required to be transferred to the allocation area, although the storage management computing device 14 can add other types and/or amounts of information.
Next in step 345, the storage management computing device 14 writes or transfers the data present in the cache within the memory 20 and the previously stored data that was read back into the identified allocation area within the SSD. Additionally back in step 330, when the storage management computing device 14 determines that there was no data found during the traversing, then the No branch is taken to this step 345. When the exemplary flow proceeds to step 345 with no data in the address range of the identified allocation area, then the storage management computing device 14 only writes or transfers the data from the cache within memory 20 to the identified allocation area. In this example, the storage management computing device 14 begins the write or transfers the data from the start of the address range of the identified allocation and proceeds to write or transfer the data sequentially in the address range. By writing the data into the allocation area sequentially, the technology is able to reduce the fragmentation in the SSD within the plurality of storage devices 16(1)-16(n) and the exemplary method ends at step 350.
Accordingly, as illustrated and described by way of the examples herein, this technology provides a number of advantages including providing methods, non-transitory computer readable media and devices for minimizing fragmentation in SSD within an enterprise storage system. By writing the data sequentially into the allocation area in the SSD, the technology able to significantly reduce the fragmentation in SSD thereby increases the memory utilization of the SSD.
Having thus described the basic concept of the technology, it will be rather apparent to those skilled in the art that the foregoing detailed disclosure is intended to be presented by way of example only, and is not limiting. Various alterations, improvements, and modifications will occur and are intended to those skilled in the art, though not expressly stated herein. These alterations, improvements, and modifications are intended to be suggested hereby, and are within the spirit and scope of the technology. Additionally, the recited order of processing elements or sequences, or the use of numbers, letters, or other designations therefore, is not intended to limit the claimed processes to any order except as may be specified in the claims. Accordingly, the invention is limited only by the following claims and equivalents thereto.

Claims (20)

What is claimed is:
1. A method for minimizing fragmentation in solid-state drives within a storage system, the method comprising:
identifying, upon determining that a cache cannot accept any more data, an allocation area in a solid-state drive of a plurality of solid-state drives to which content of the cache can be transferred;
determining a size for the allocation area that is equal to or great than a size of the content within the cache that is to be transferred;
determining an address range in a memory strip of the solid-state drive for the allocation area; and
initiating a sequential transfer of the content of the cache to a start of the address range in the allocation area.
2. The method of claim 1, comprising:
storing a write data operation into the cache.
3. The method of claim 1, comprising:
reading data from the allocation area at the address range.
4. The method of claim 3, wherein the data is read from the allocation area based upon the cache reaching a threshold size.
5. The method of claim 3, comprising:
sequentially writing cached data from the cache and the data read from the allocation area back into the allocation area.
6. The method of claim 5, wherein the cached data and the data read from the allocation area are written sequentially beginning from the start of the address range in the allocation area.
7. The method of claim 5, comprising:
writing the cached data to the allocation area sequentially based upon a determination that the allocation area lacks the data read from the allocation area.
8. A non-transitory machine readable medium comprising instructions for performing a method, which when executed by a machine, causes the machine to:
in response to determining that a cache cannot accept any more data, identify an allocation area in a solid-state drive of a plurality of solid-state drives to which content of the cache can be transferred;
determine a size for the allocation area that is equal to or great than a size of the content within the cache that is to be transferred;
determine an address range in a memory strip of the solid-state drive for the allocation area; and
initiate a sequential transfer of the content of the cache to a start of the address range in the allocation area.
9. The non-transitory machine readable medium of claim 8, wherein the instructions cause the machine to:
read data from the allocation area at the address range.
10. The non-transitory machine readable medium of claim 9, wherein the data is read from the allocation area based upon the cache reaching a threshold size.
11. The non-transitory machine readable medium of claim 9, wherein the instructions cause the machine to:
sequentially write cached data from the cache and the data read from the allocation area back into the allocation area.
12. The non-transitory machine readable medium of claim 11, wherein the cached data and the data read from the allocation area are written sequentially beginning from the start of the address range in the allocation area.
13. The non-transitory machine readable medium of claim 11, wherein the instructions cause the machine to:
write the cached data to the allocation area sequentially based upon a determination that the allocation area lacks the data read from the allocation area.
14. The non-transitory machine readable medium of claim 8, wherein the instructions cause the machine to:
store a write data operation into the cache.
15. A computing device comprising:
a cache into which write data operations from client devices are cached;
a plurality of solid-state drives; and
a processor configured to execute instructions to cause the computing device to:
in response to determining that a cache cannot accept any more data, identify an allocation area in a solid-state drive of the plurality of solid-state drives to which content of the cache can be transferred;
determine a size for the allocation area that is equal to or great than a size of the content within the cache that is to be transferred;
determine an address range in a memory strip of the solid-state drive for the allocation area; and
initiate a sequential transfer of the content of the cache to a start of the address range in the allocation area.
16. The computing device of claim 15, wherein the instructions cause the computing device to:
read data from the allocation area at the address range.
17. The computing device of claim 16, wherein the data is read from the allocation area based upon the cache reaching a threshold size.
18. The computing device of claim 16, wherein the instructions cause the computing device to:
sequentially write cached data from the cache and the data read from the allocation area back into the allocation area.
19. The computing device of claim 18, wherein the cached data and the data read from the allocation area are written sequentially beginning from the start of the address range in the allocation area.
20. The computing device of claim 18, wherein the instructions cause the computing device to:
write the cached data to the allocation area sequentially based upon a determination that the allocation area lacks the data read from the allocation area.
US17/486,160 2016-06-28 2021-09-27 Methods for minimizing fragmentation in SSD within a storage system and devices thereof Active US11592986B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/486,160 US11592986B2 (en) 2016-06-28 2021-09-27 Methods for minimizing fragmentation in SSD within a storage system and devices thereof

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US15/195,093 US10430081B2 (en) 2016-06-28 2016-06-28 Methods for minimizing fragmentation in SSD within a storage system and devices thereof
US16/584,025 US11132129B2 (en) 2016-06-28 2019-09-26 Methods for minimizing fragmentation in SSD within a storage system and devices thereof
US17/486,160 US11592986B2 (en) 2016-06-28 2021-09-27 Methods for minimizing fragmentation in SSD within a storage system and devices thereof

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/584,025 Continuation US11132129B2 (en) 2016-06-28 2019-09-26 Methods for minimizing fragmentation in SSD within a storage system and devices thereof

Publications (2)

Publication Number Publication Date
US20220083232A1 US20220083232A1 (en) 2022-03-17
US11592986B2 true US11592986B2 (en) 2023-02-28

Family

ID=59091609

Family Applications (3)

Application Number Title Priority Date Filing Date
US15/195,093 Active 2036-12-17 US10430081B2 (en) 2016-06-28 2016-06-28 Methods for minimizing fragmentation in SSD within a storage system and devices thereof
US16/584,025 Active 2036-09-03 US11132129B2 (en) 2016-06-28 2019-09-26 Methods for minimizing fragmentation in SSD within a storage system and devices thereof
US17/486,160 Active US11592986B2 (en) 2016-06-28 2021-09-27 Methods for minimizing fragmentation in SSD within a storage system and devices thereof

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/195,093 Active 2036-12-17 US10430081B2 (en) 2016-06-28 2016-06-28 Methods for minimizing fragmentation in SSD within a storage system and devices thereof
US16/584,025 Active 2036-09-03 US11132129B2 (en) 2016-06-28 2019-09-26 Methods for minimizing fragmentation in SSD within a storage system and devices thereof

Country Status (4)

Country Link
US (3) US10430081B2 (en)
CN (1) CN109416663B (en)
DE (1) DE112017003232T5 (en)
WO (1) WO2018005041A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111045732B (en) * 2019-12-05 2023-06-09 腾讯科技(深圳)有限公司 Data processing method, chip, device and storage medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6418510B1 (en) * 2000-09-14 2002-07-09 International Business Machines Corporation Cooperative cache and rotational positioning optimization (RPO) scheme for a direct access storage device (DASD)
US20070214309A1 (en) * 2006-03-07 2007-09-13 Matsushita Electric Industrial Co., Ltd. Nonvolatile storage device and data writing method thereof
US20080250195A1 (en) 2000-01-06 2008-10-09 Super Talent Electronics Inc. Multi-Operation Write Aggregator Using a Page Buffer and a Scratch Flash Block in Each of Multiple Channels of a Large Array of Flash Memory to Reduce Block Wear
US20110099326A1 (en) 2009-10-27 2011-04-28 Samsung Electronics Co., Ltd. Flash memory system and defragmentation method
US20130019057A1 (en) * 2011-07-15 2013-01-17 Violin Memory, Inc. Flash disk array and controller
US20140173178A1 (en) 2012-12-19 2014-06-19 Apple Inc. Joint Logical and Physical Address Remapping in Non-volatile Memory
US20150106410A1 (en) 2013-10-10 2015-04-16 Apple Inc. Memory System with Shared File System
US20160085674A1 (en) 2014-09-24 2016-03-24 Netapp, Inc. Methods and systems for dynamically controlled caching
US20160274792A1 (en) 2013-11-22 2016-09-22 Nec Corporation Storage apparatus, method, and program
US20160371035A1 (en) 2014-03-27 2016-12-22 Hitachi, Ltd. Storage system
US20170031994A1 (en) 2015-07-27 2017-02-02 Datrium, Inc. System and Methods for Storage Data Deduplication

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8135933B2 (en) * 2007-01-10 2012-03-13 Mobile Semiconductor Corporation Adaptive memory system for enhancing the performance of an external computing device
KR101717644B1 (en) * 2009-09-08 2017-03-27 샌디스크 테크놀로지스 엘엘씨 Apparatus, system, and method for caching data on a solid-state storage device
US9003159B2 (en) * 2009-10-05 2015-04-07 Marvell World Trade Ltd. Data caching in non-volatile memory
US8407403B2 (en) * 2009-12-07 2013-03-26 Microsoft Corporation Extending SSD lifetime using hybrid storage
US20110320733A1 (en) 2010-06-04 2011-12-29 Steven Ted Sanford Cache management and acceleration of storage media
KR101824295B1 (en) * 2011-08-12 2018-01-31 샌디스크 테크놀로지스 엘엘씨 Cache management including solid state device virtualization
CN103827970B (en) 2012-09-25 2016-05-18 株式会社东芝 To solid-state drive storage device, storage control and the method for configuration data again
CN102968380B (en) * 2012-11-02 2016-12-07 深圳市同洲电子股份有限公司 The management method of memory partitioning and device in memory file system

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080250195A1 (en) 2000-01-06 2008-10-09 Super Talent Electronics Inc. Multi-Operation Write Aggregator Using a Page Buffer and a Scratch Flash Block in Each of Multiple Channels of a Large Array of Flash Memory to Reduce Block Wear
US6418510B1 (en) * 2000-09-14 2002-07-09 International Business Machines Corporation Cooperative cache and rotational positioning optimization (RPO) scheme for a direct access storage device (DASD)
US20070214309A1 (en) * 2006-03-07 2007-09-13 Matsushita Electric Industrial Co., Ltd. Nonvolatile storage device and data writing method thereof
US20110099326A1 (en) 2009-10-27 2011-04-28 Samsung Electronics Co., Ltd. Flash memory system and defragmentation method
US20130019057A1 (en) * 2011-07-15 2013-01-17 Violin Memory, Inc. Flash disk array and controller
US20140173178A1 (en) 2012-12-19 2014-06-19 Apple Inc. Joint Logical and Physical Address Remapping in Non-volatile Memory
US20150106410A1 (en) 2013-10-10 2015-04-16 Apple Inc. Memory System with Shared File System
US20160274792A1 (en) 2013-11-22 2016-09-22 Nec Corporation Storage apparatus, method, and program
US20160371035A1 (en) 2014-03-27 2016-12-22 Hitachi, Ltd. Storage system
US20160085674A1 (en) 2014-09-24 2016-03-24 Netapp, Inc. Methods and systems for dynamically controlled caching
US20170031994A1 (en) 2015-07-27 2017-02-02 Datrium, Inc. System and Methods for Storage Data Deduplication

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Heuristic User Behavior Based Hard Disk Defragmentation Mechanism", 2010, Varun Mittal and Chinmay Sharma, UKSim Fourth European Modeling Symposium on Computer Modeling and Simulation, pp. 43-45.
"Smart Data Management—Atonomous Disk De-fragmentation", Sep. 2014, ip.com, an IP.com Prior Art Database Technical Disclosure 3 pgs.
Int. Search Report/Written Opinion cited in the International Application No. PCT/US036680 dated Jun. 9, 2017, 10 pgs.
Oh, Yongseok, et al. "Caching less for better performance: balancing cache size and update cost of flash memory cache in hybrid storage systems." FAST. vol. 12. 2012. (Year: 2012). *

Also Published As

Publication number Publication date
WO2018005041A1 (en) 2018-01-04
CN109416663B (en) 2023-07-21
CN109416663A (en) 2019-03-01
US20220083232A1 (en) 2022-03-17
US10430081B2 (en) 2019-10-01
US20200019321A1 (en) 2020-01-16
US11132129B2 (en) 2021-09-28
US20170371556A1 (en) 2017-12-28
DE112017003232T5 (en) 2019-03-28

Similar Documents

Publication Publication Date Title
CN109725840B (en) Throttling writes with asynchronous flushing
JP5087467B2 (en) Method and apparatus for managing data compression and integrity in a computer storage system
US20060242264A1 (en) Network file sharing method and system
US7596659B2 (en) Method and system for balanced striping of objects
CN110109915B (en) Method, apparatus and computer program product for managing hash tables
TWI476676B (en) File system for storage device which uses different cluster sizes
TWI668626B (en) Storage apparatus and method for autonomous space compaction
US20050086386A1 (en) Shared running-buffer-based caching system
US10721332B2 (en) System and method for process migration in a content centric network
US9936017B2 (en) Method for logical mirroring in a memory-based file system
CN107832423B (en) File reading and writing method for distributed file system
WO2017042978A1 (en) Computer system, storage device, and data management method
US11592986B2 (en) Methods for minimizing fragmentation in SSD within a storage system and devices thereof
US6614709B2 (en) Method and apparatus for processing commands in a queue coupled to a system or memory
US10387043B2 (en) Writing target file including determination of whether to apply duplication elimination
WO2014153931A1 (en) File storage method and device, access client and metadata server system
WO2018028218A1 (en) Data writing method and apparatus
US9984112B1 (en) Dynamically adjustable transaction log
US20170344578A1 (en) Compressed data layout for optimizing data transactions
US9652393B1 (en) Managing shared memory in database server
EP4087212A1 (en) Method and apparatus for cloning file system
US10684786B2 (en) Methods for performing global deduplication on data blocks and devices thereof
CN111600943A (en) Method and equipment for acquiring target data
JP5494363B2 (en) File management program, file management method, and file management apparatus
US20170187804A1 (en) Methods for transferring data based on actual size of a data operation and devices thereof

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE