US20190107976A1 - Apparatus and method for assigning velocities to write data - Google Patents

Apparatus and method for assigning velocities to write data Download PDF

Info

Publication number
US20190107976A1
US20190107976A1 US16/213,274 US201816213274A US2019107976A1 US 20190107976 A1 US20190107976 A1 US 20190107976A1 US 201816213274 A US201816213274 A US 201816213274A US 2019107976 A1 US2019107976 A1 US 2019107976A1
Authority
US
United States
Prior art keywords
application
data
velocity
data type
different
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/213,274
Other versions
US11231873B2 (en
Inventor
Sanjeev N. Trika
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SK Hynix NAND Product Solutions Corp
Original Assignee
Intel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corp filed Critical Intel Corp
Priority to US16/213,274 priority Critical patent/US11231873B2/en
Assigned to INTEL CORPORATION reassignment INTEL CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TRIKA, SANJEEV N.
Publication of US20190107976A1 publication Critical patent/US20190107976A1/en
Application granted granted Critical
Publication of US11231873B2 publication Critical patent/US11231873B2/en
Assigned to SK HYNIX NAND PRODUCT SOLUTIONS CORP. reassignment SK HYNIX NAND PRODUCT SOLUTIONS CORP. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTEL CORPORATION
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0655Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
    • G06F3/0659Command handling arrangements, e.g. command buffers, queues, command scheduling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/061Improving I/O performance
    • G06F3/0611Improving I/O performance in relation to response time
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • G06F3/0616Improving the reliability of storage systems in relation to life time, e.g. increasing Mean Time Between Failures [MTBF]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device
    • G06F3/0679Non-volatile semiconductor memory device, e.g. flash memory, one time programmable memory [OTP]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Debugging And Monitoring (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

An apparatus is described. The apparatus includes velocity assignment logic to assign a velocity to data that is to be written to a non volatile storage medium. The velocity assignment logic is to accept input information pertaining to an identity of an application that is writing the data, the data type of the data and the state of the application in order to determine the velocity.

Description

    FIELD OF INVENTION
  • The field of invention pertains generally to the computing sciences, and, more particularly, to an apparatus and method for assigning velocities to write data.
  • BACKGROUND
  • Writing data to flash memory is a complicated process. As data cannot be directly overwritten on flash, old data must first be erased, after which new data may be written. The process by which the flash reclaims the physical areas that no longer have valid data is called “garbage collection.” This process entails consolidating valid data from a flash block, writing it to a different block, and then erasing the original block which then removes the invalid data and frees up the space it was consuming.
  • The garbage collection process impacts performance the more varied the lifetimes of the data within a stored block of data (where, lifetime corresponds to the amount of time the data sits in flash without change). That is, if the lifetimes of the data items within an average block of data greatly vary, the garbage collection process will spend more time re-writing valid items of data to other blocks prior to erasing the original block of data. By contrast, if the average block of data contains data with uniform/same lifetimes, ideally, little/no time is spent re-writing valid data during garbage collection because all of the block's data becomes invalid within a same time window, and, the entire block can be erased without re-writing any valid data (the entire block is invalid).
  • Present and future Non Volatile Memory Express (NVMe) specifications (which are open logical device interface specifications for accessing non volatile storage media over a Peripheral Component Interface (PCI) Express (PCIe) link), therefore include “stream” technology in which the host appends a respective velocity to an item of data to be written into an NVMe compliant device (e.g., a solid state drive (SSD)). The velocity value essentially indicates the data item's expected lifetime. The non volatile storage device will then store items of data having similar/same lifetimes in a same data block so as to minimize the garbage collection overhead.
  • FIGURES
  • A better understanding of the present invention can be obtained from the following detailed description in conjunction with the following drawings, in which:
  • FIG. 1 shows a velocity assignment function;
  • FIG. 2 shows a method performed by the velocity assignment function;
  • FIGS. 3a and 3b show examples of application velocity and file velocity tables, respectively.
  • FIG. 4 shows a computing system.
  • DETAILED DESCRIPTION
  • With stream technology already being integrated on NVMe compliant storage devices, a remaining challenge is the host's ability to accurately assign velocities to the data items it stores on a storage device having streams functionality. FIG. 1 shows a velocity assignment function 100 that initially assigns velocities to data items based on respective characteristics of the data items and default values assigned to these characteristics. Over the runtime of the storage device 110, statistics related to actual data lifetimes are tracked and tweaks to the characteristics' initial default values are made (if warranted).
  • In various embodiments, a velocity determination is made for a data item to be written in a storage device 110 having streams functionality by velocity determination logic 101 within the velocity assignment function 100. The characteristics that the velocity determination logic 101 bases its velocity determination upon include: 1) the identity of the application whose data is being written; 2) the type of data that is being written (e.g., filetype); and, 3) the state of the application whose data it being written.
  • As observed in FIG. 1, the velocity determination logic 101 receives information from an application velocity table (AVT) 102 and a file type velocity table (FVT) 103. The AVT 102 lists different applications and assigns a respective velocity for each listed application. The FVT 103 lists different data types and assigns a respective velocity for each listed data type. Initially, the values of these tables are set to default values. Over time, the default values can be modified based on statistics that pertain to data lifetime collected over the runtime of the storage device 110.
  • As is known in the art, a write request 106, e.g., as submitted from a higher level of software to a lower level of software/firmware, includes the data item and an address for the data item (e.g., a logical block address (LBA)). If the velocity assignment function 100 is a component of software that directly receives the write request 106 through an application programming interface (API) 107 from the application whose data is being written, the identity of the application can be discernable directly from the write request 106 itself (e.g., because the application identifies itself as part of invoking the API 107), or, the application expressly identifies itself in the structure of the request 106 (e.g., a special field is reserved in the format of the request 106 where the identity of the application is provided). The later approach may be particularly useful, e.g., where one or more layers of software exist between the application whose data is being written and the layer where the velocity function 100 resides.
  • Regardless, in any of the aforementioned embodiments, the state of the application whose data is being written is provided in the write request 106, the API 107, and/or is provided/reported by some other technique (such as a process-to-process communication (e.g., the application communicates to the storage device driver directly)). As just one possible implementation of application state being provided in the write request 106 and/or API 107, the syntax structure of the write request 106 and/or API 107 is designed to include information that indicates the present state of the application (e.g., editing, closing, downloading, etc.). For example, a special field is reserved in the format of the write request 106 and/or API 107 where the state of the application is provided.
  • Thus, referring to FIG. 1, as part of the processing of the write request 106, the identity of the application whose data is being written is used as a look-up parameter into the AVT 102 which, in turn, provides a first parameter 111 based on the identity of the application (the application parameter 111). Additionally, the type of data that is being written is used as a look-up parameter into the FVT 103, which, in turn provides a second parameter 112 based on the data type (the data type parameter 112). Here, in the case of a file, the file name's extension (e.g., .doc, .docx, .pdf, .rtf, .html, .exe, etc.) can be taken from the write request and used to identify the data type (the file name and it's extension is included in the write request such as being a component of the data item's LBA).
  • As will be explained in more detail below, both the application parameter 111 and the data type parameter 112 are values (e.g., numerical values) that reflect an expected lifetime of the data being written and are used by the velocity determination logic 101 to determine a velocity value for the data.
  • Finally, the state of the application whose data is being written is provided to the velocity determination logic 101 as a third input that the determined velocity is based on. According to a first approach, the state of the application may be provided to the velocity determination logic 101 directly and the velocity determination logic 101 is hardcoded and/or hardwired and/or programmed to account for the application state (FIG. 1 indicates this approach). Or, according to second approach, as with the application identity and the data type identity, a table is instantiated having multiple entries where each entry provides a specific parameter for a specific application state. In this later case, the application state is identified in the write request 106 and/or API 107 and is used as a look-up parameter in the table which returns a third parameter (the application state parameter) for the application's particular state. The application state parameter, in turn, is a value that reflects the application's state effect on the lifetime of the data being written and is provided to the velocity determination logic 101.
  • The application parameter 111 from the AVT 102, the data type parameter 112 from the FVT 103 and the application state (or an application state parameter) are provided to the velocity determination logic 101 which determines a velocity for the data being written based on these inputs. The determined velocity is then provided to the storage device 110 as part of the write process.
  • In various implementation the velocity assignment function 100 is partially or wholly integrated in higher levels of software above the storage device driver (such as a virtual machine monitor, operating system and/or user application software program).
  • In other implementations, the velocity assignment function 100 is part of the device driver software for the storage device 110. Here, as is known in the art, when higher level software (e.g., a specific user application, an operating system instance, a virtual machine monitor) desires to read/write data from/to the storage device 110, the higher level software provides a request through the device driver's API. In this case, the API 107 of FIG. 1 corresponds to the device driver API.
  • In other implementations, the velocity assignment function 100 is implemented partially in the storage device's device driver software and partially in the storage device's hardware and/or firmware. In still other implementations, the velocity assignment function 100 is implemented in the storage device hardware and/or firmware, entirely, and/or two or more layers of software (which the storage device driver may or may not be included within).
  • In any of these approaches, any of the velocity determination logic 101 and look-up tables 101, 102 (and statistics tracking logic 105 as described more fully below) may be implemented in hardware and/or firmware of the host computer and/or storage device 110. Here, the hardware may entail dedicated, custom hardwired logic circuitry (e.g., an application specific integrated circuit (ASIC), state machine logic circuitry, etc.), programmable logic circuitry (e.g., field programmable logic array (FPGA)) or logic circuitry that executes program code (e.g., an embedded processor, embedded controller, etc.) or any combination of the above.
  • As discussed above, the initial application parameter 111 and data type parameter 112 values are default values. Over the runtime of the storage device, however, these values 111, 112 are updated based on monitored statistics of the storage device's actual usage. In particular, statistics tracking logic 105 tracks the lifetime of data on a per application basis and on a per data type basis. According to one embodiment, lifetime statistics include both the number of total writes made and the number of those writes that overwrite existing data. That is, a first count corresponds to the total amount of data written and a second count corresponds to the total amount of data that overwrites a previous version of the data. Such counts are maintained by the statistics tracking logic 105 which updates counts in the AVT 102 and FVT 103 for each application and each data type.
  • Here, the first count may be determined, e.g., by combining the total amount of data associated with all PUT operations and all MODIFY operations for both a particular application and a particular data type (as is understood in the art, a PUT operation corresponds to the creation of a new data item in storage, while, a MODIFY operation corresponds to the overwrite of a data item that already exists in storage). The second count may be determined, e.g., by counting the total amount of data associated with all MODIFY operations associated with a particular application and a particular data type. As such, for any PUT operation, the first count is incremented for both the data's application and data type, and, for any MODIFY operation, both the first and second counts are incremented for the data's application and data type. Here, the statistics tracking logic increments count values in both the AVT 102 and FVT 103 for both PUT operations and MODIFY operations.
  • In an embodiment, both the application parameter 111 and the data type parameter 112 are expressed as the percentage of total written data that is overwritten data, or information from which such a percentage can be calculated. According to one approach, the AVT table 102 not only includes the default value for each application but also the first and second counts for each application. Likewise, the FVT table 103 not only includes the default value for each data type but also the first and second counts for each data types.
  • As such, in order to determine the percentages, the velocity determination logic 101 normalizes the second count by the first count (i.e., parameter=100*(first count)/(second count)). In other embodiments, the percentage is kept in the tables 101, 102 and the velocity determination logic 101 accepts them without having to perform percentage calculations (in these approaches, the application and data type parameters 111, 112 correspond to percentage calculations rather than count values). For convenience the remainder of the present specification will assume that count values are kept in tables 102, 103.
  • Regardless, higher percentages of overwrites (e.g., application or data type parameter >50) corresponds to data that has a shorter lifetime, whereas, lower percentages of overwrites (e.g., application or data type parameter <50) corresponds to data that has a longer lifetime.
  • FIG. 2 shows a process 200 for determining an appropriate velocity for data that is to be written into a storage device 110. According to the method of FIG. 2, the AVT table 102 is accessed 201 to retrieve the default value and the first and second counts for the application whose data is being written. The application parameter 111 is then determined 203. If the first count (total written data) is beneath some threshold the default value is accepted for the application parameter. Contra-wise, if the first count is above some threshold, the application parameter is calculated as 100*(first count)/(second count). The first and second application count values can be incremented to account for the data that is being written by the instant write request.
  • Likewise, the FVT table 103 is also accessed 202 to retrieve the default value and the first and second counts for the type of data that is being written. The retrieved default value and first and second counts are provided to the velocity determination logic 101 so that the data type parameter can be determined 204. If the first count (total written data) is beneath some threshold the default value is accepted for the data type parameter. Contra-wise, if the first count is above some threshold, the data type parameter is calculated as 100*(first count)/(second count). The first and second data type count values can be incremented to account for the data that is being written by the instant write request.
  • With the application and data type parameters being determined 203, 204 and provided to the velocity determination logic 101, the velocity determination logic 101 also takes into account the state of the application whose data is being written to finally determine the velocity 205. As mentioned above, the application state can be provided directly to the velocity determination logic 101 from the write request, or, the velocity determination logic 101 can receive an application state parameter provided by a third look-up table in response to the application state being presented to the look-up table.
  • According to one approach, in determining 205 the appropriate velocity for the data being written, the application state is applied as an offset to the average of the application and data type parameters. That is, the average of the application and data type parameters is calculated and the resultant is adjusted up or down based on the application state. Here, the application parameter may have one influence on the data's lifetime while the type of data parameter may exhibit another influence. In the case where the two influences are deemed to be equal, the first and second parameters are averaged with equal weights applied to both parameters. In other approaches, the application and data type parameters may have less than equal influences and a weighted average is calculated (e.g., by factoring the lesser of the two influences by a fraction (e.g., a value between 0 and 1)
  • Adjusting the average of the application and data type parameters up or down in view of the application state can be implement a significant adjustment to velocity in situations where the application state provides significant guidance into the expected lifetime of the data that is being written. As an example, if the application whose data is being written is known to be in an editing mode (e.g., a word processing application is active and executing in a document editing state), an offset value of some amount is added to the average of the application and data type parameters to indicate that the data being written is apt to have a shorter lifetime (because the data's application is in an editing mode, the data is apt to be overwritten in the near future).
  • By contrast, if the application is known to be closing (shutting down), in all likelihood the data is being saved as a final save prior to the shut down of the application. As such, the average can be adjusted lower (an offset is subtracted from the average) to indicate that the data is expected to have a longer lifetime (it will not be modified until the application presently being closed is re-opened). As such, more generally, the application state is applied as a positive offset to the average of the application and data type parameters if the application state is more likely to modify the data in the near future, or, the application state is applied as a negative offset to the average of the application and data type parameters if the application state is less likely to modify the data in the near future.
  • Other types of application state conditions that can affect the final velocity determination 205 include whether data is being downloaded or not (if data is being downloaded, it is apt remain intact in the near future which corresponds to a negative offset).
  • After the velocity has been determined 205, it is communicated to the storage device as part of the data's write process (e.g., is included in the write request for the data that is provided to the storage device 110). The count values of the AVT and FVT tables 102, 103 are then updated to reflect the amount of data being written.
  • FIGS. 3a and 3b show exemplary AVT and FVT table structures, respectively. Here, as observed in FIGS. 3a and 3b , both tables include a default value, a total write data amount and a total overwrite data amount for each table entry. Each entry corresponds to a different application (in the case of the AVT table of FIG. 3a ) or a different type of data (in the case of the FVT table of FIG. 3b ).
  • Although the values are exemplary, referring the exemplary AVT table of FIG. 3a , note that applications that are designed to edit information (such as a word processing application (e.g., Microsoft Word) or a spreadsheet application (e.g., Microsoft Excel) have higher default values than applications that, for the most part, merely display or render information (e.g., a web browser or media viewer). Likewise, referring to the exemplary FVT table of FIG. 3b , note that data files that are frequently updated (e.g., text documents, spreadsheets, page files (e.g., cached data being written back to storage from main memory)) have higher velocity values whereas data types that are primarily used as read-only data structures (e.g., HTML files) have lower velocity values.
  • FIG. 4 provides an exemplary depiction of a host computing system 400 (e.g., a smartphone, a tablet computer, a laptop computer, a desktop computer, a server computer, etc.). As observed in FIG. 4, the basic computing system 400 may include a central processing unit 401 (which may include, e.g., a plurality of general purpose processing cores 415_1 through 415_X) and a main memory controller 417 disposed on a multi-core processor or applications processor, system memory 402, a display 403 (e.g., touchscreen, flat-panel), a local wired point-to-point link (e.g., USB) interface 404, various network I/O functions 405 (such as an Ethernet interface and/or cellular modem subsystem), a wireless local area network (e.g., WiFi) interface 406, a wireless point-to-point link (e.g., Bluetooth) interface 407 and a Global Positioning System interface 408, various sensors 409_1 through 409_Y, one or more cameras 410, a battery 411, a power management control unit 412, a speaker and microphone 413 and an audio coder/decoder 414.
  • An applications processor or multi-core processor 450 may include one or more general purpose processing cores 415 within its CPU 401, one or more graphical processing units 416, a memory management function 417 (e.g., a memory controller) and an I/O control function 418. The general purpose processing cores 415 typically execute the operating system and application software of the computing system. The graphics processing unit 416 typically executes graphics intensive functions to, e.g., generate graphics information that is presented on the display 403. The memory control function 417 interfaces with the system memory 402 (also referred to as main memory) to write/read data to/from system memory 402. The power management control unit 412 generally controls the power consumption of the system 400.
  • Each of the touchscreen display 403, the communication interfaces 404-407, the GPS interface 408, the sensors 409, the camera(s) 410, and the speaker/ microphone codec 413, 414 all can be viewed as various forms of I/O (input and/or output) relative to the overall computing system including, where appropriate, an integrated peripheral device as well (e.g., the one or more cameras 410). Depending on implementation, various ones of these I/O components may be integrated on the applications processor/multi-core processor 450 or may be located off the die or outside the package of the applications processor/multi-core processor 450. The computing system also includes non-volatile storage 420 which may be the mass storage component of the system.
  • The computing system may include functionality to assign velocities to write data of the non-volatile storage 420 as described at length above.
  • Embodiments of the invention may include various processes as set forth above. The processes may be embodied in machine-executable instructions. The instructions can be used to cause a general-purpose or special-purpose processor to perform certain processes. Alternatively, these processes may be performed by specific/custom hardware components that contain hardwired logic circuitry or programmable logic circuitry (e.g., field programmable gate array (FPGA), programmable logic device (PLD)) for performing the processes, or by any combination of programmed computer components and custom hardware components.
  • Elements of the present invention may also be provided as a machine-readable medium for storing the machine-executable instructions. The machine-readable medium may include, but is not limited to, floppy diskettes, optical disks, CD-ROMs, and magneto-optical disks, FLASH memory, ROMs, RAMs, EPROMs, EEPROMs, magnetic or optical cards, propagation media or other type of media/machine-readable medium suitable for storing electronic instructions. For example, the present invention may be downloaded as a computer program which may be transferred from a remote computer (e.g., a server) to a requesting computer (e.g., a client) by way of data signals embodied in a carrier wave or other propagation medium via a communication link (e.g., a modem or network connection).
  • In the foregoing specification, the invention has been described with reference to specific exemplary embodiments thereof. It will, however, be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention as set forth in the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims (20)

1. An apparatus, comprising:
velocity assignment logic to assign a velocity to data that is to be written to a non volatile storage medium, the velocity assignment logic to accept input information pertaining to an identity of an application that is writing the data, the data type of the data and the state of the application in order to determine the velocity.
2. The apparatus of claim 1 wherein the apparatus further comprises an application look-up table having different entries for different application identities.
3. The apparatus of claim 2 wherein the different entries include a respective default value for an application parameter to be accepted by the velocity assignment logic as a component of the input information.
4. The apparatus of claim 2 wherein the different entries include a respective write count value for a respective application.
5. The apparatus of claim 1 wherein the apparatus further comprises a data type look-up table having different entries for different application identities.
6. The apparatus of claim 5 wherein the different entries include a respective default value for a data type parameter to be accepted by the velocity assignment logic as a component of the input information.
7. The apparatus of claim 5 wherein the different entries include a respective write count value for a respective data type.
8. The apparatus of claim 1 wherein the velocity assignment logic determines a value from the input information associated with the identity of the application and the data type, and, adjusts the value based on the application state.
9. A computing system, comprising:
a plurality of processing cores;
a non volatile storage device; and,
a machine readable storage medium containing program code that when processed by the computing system causes a method to be performed, the method comprising:
receiving first information pertaining to an identity of an application that is writing data to the non volatile device;
receiving second information pertaining to the data's data type;
receiving third information pertaining to a state of the application; and,
determining a velocity for the data from the first, second and third information and providing the velocity to the non volatile storage device.
10. The computing system of claim 9 wherein the method further comprises using the identity of the application as an input to an application look-up table having different entries for different application identities.
11. The computing system of claim 10 wherein the different entries include a respective default value for an application parameter to be accepted by velocity assignment logic.
12. The computing system of claim 10 wherein the different entries include a respective write count value for a respective application.
13. The computing system of claim 9 wherein the method further comprises using the data's data type as an input to a data type look-up table having different entries for different application identities.
14. The computing system of claim 13 wherein the different entries include a respective default value for a data type parameter to be accepted by velocity assignment logic.
15. The computing system of claim 13 wherein the different entries include a respective write count value for a respective data type.
16. The computing system of claim 9 wherein the determining a velocity further comprises determining a value from the first and second information, and, adjusting the value based on the third information.
17. A machine readable storage medium containing program code that when processed by one or more processors causes a method to be performed, the method comprising:
receiving first information pertaining to an identity of an application that is writing data to the non volatile device;
receiving second information pertaining to the data's data type;
receiving third information pertaining to a state of the application; and,
determining a velocity for the data from the first, second and third information and providing the velocity to the non volatile storage device.
18. The machine readable storage medium of claim 17 wherein the method further comprises using the identity of the application as an input to an application look-up table having different entries for different application identities.
19. The machine readable storage medium of claim 18 wherein the method further comprises using the data's data type as an input to a data type look-up table having different entries for different application identities.
20. The machine readable storage medium of claim 17 wherein the determining a velocity further comprises determining a value from the first and second information, and, adjusting the value based on the third information.
US16/213,274 2018-12-07 2018-12-07 Apparatus and method for assigning velocities to write data Active 2038-12-26 US11231873B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/213,274 US11231873B2 (en) 2018-12-07 2018-12-07 Apparatus and method for assigning velocities to write data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US16/213,274 US11231873B2 (en) 2018-12-07 2018-12-07 Apparatus and method for assigning velocities to write data

Publications (2)

Publication Number Publication Date
US20190107976A1 true US20190107976A1 (en) 2019-04-11
US11231873B2 US11231873B2 (en) 2022-01-25

Family

ID=65993970

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/213,274 Active 2038-12-26 US11231873B2 (en) 2018-12-07 2018-12-07 Apparatus and method for assigning velocities to write data

Country Status (1)

Country Link
US (1) US11231873B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114238232A (en) * 2021-12-15 2022-03-25 南方电网电力科技股份有限公司 Intelligent ammeter file reading and writing system, method, equipment and medium
US11531615B2 (en) * 2018-12-14 2022-12-20 SK Hynix Inc. Controller and memory system for performing garbage collection operation, and operating method thereof

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020152060A1 (en) * 1998-08-31 2002-10-17 Tseng Ping-Sheng Inter-chip communication system
US6487713B1 (en) * 1999-09-24 2002-11-26 Phoenix Technologies Ltd. Software development system that presents a logical view of project components, facilitates their selection, and signals missing links prior to compilation
US20030105772A1 (en) * 1999-02-11 2003-06-05 Harlan Sexton Write-barrier maintenance in a garbage collector
US20030144828A1 (en) * 2001-07-30 2003-07-31 Lin Sharon Sheau-Pyng Hub array system and method
US6910053B1 (en) * 1999-06-18 2005-06-21 Sap Aktiengesellschaft Method for data maintenance in a network of partially replicated database systems
US20100064111A1 (en) * 2008-09-09 2010-03-11 Kabushiki Kaisha Toshiba Information processing device including memory management device managing access from processor to memory and memory management method
US20120191900A1 (en) * 2009-07-17 2012-07-26 Atsushi Kunimatsu Memory management device
US8639024B2 (en) * 1997-04-15 2014-01-28 Intel Corporation Generation of a disparity result with low latency
US20160179376A1 (en) * 2014-12-23 2016-06-23 Intel Corporation Method and apparatus for improving read performance of a solid state drive
US20160283116A1 (en) * 2015-03-27 2016-09-29 Intel Corporation Sequential write stream management
US20160313943A1 (en) * 2015-04-24 2016-10-27 Kabushiki Kaisha Toshiba Storage device that secures a block for a stream or namespace and system having the storage device
US20180074709A1 (en) * 2016-09-15 2018-03-15 Western Digital Technologies, Inc. Stream management for storage devices
US20180302443A1 (en) * 2012-08-02 2018-10-18 CellSec, Inc. Dividing a data processing device into separate security domains
US20190155521A1 (en) * 2017-11-21 2019-05-23 Micron Technology, Inc. Data categorization based on invalidation velocities

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW201445429A (en) * 2013-05-22 2014-12-01 Asmedia Technology Inc Disk array system and data processing method
US10509770B2 (en) * 2015-07-13 2019-12-17 Samsung Electronics Co., Ltd. Heuristic interface for enabling a computer device to utilize data property-based data placement inside a nonvolatile memory device
US10120573B2 (en) * 2015-09-14 2018-11-06 Microsoft Technology Licensing, Llc. Modular sequential writing of data to data storage devices
TWI610179B (en) * 2016-12-07 2018-01-01 慧榮科技股份有限公司 Host device and methods for controlling a data transfer speed
CN108170370B (en) * 2016-12-07 2021-01-26 慧荣科技股份有限公司 Data storage device and data transmission rate control method
US11081163B2 (en) * 2019-04-12 2021-08-03 Kioxia Corporation Information processing apparatus, information processing system, and semiconductor storage device

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8639024B2 (en) * 1997-04-15 2014-01-28 Intel Corporation Generation of a disparity result with low latency
US20020152060A1 (en) * 1998-08-31 2002-10-17 Tseng Ping-Sheng Inter-chip communication system
US20030105772A1 (en) * 1999-02-11 2003-06-05 Harlan Sexton Write-barrier maintenance in a garbage collector
US6910053B1 (en) * 1999-06-18 2005-06-21 Sap Aktiengesellschaft Method for data maintenance in a network of partially replicated database systems
US20050203946A1 (en) * 1999-06-18 2005-09-15 Sap Aktiengesellschaft Method for data maintenance in a network of partially replicated database systems
US7587432B2 (en) * 1999-06-18 2009-09-08 Sap Ag Method for data maintenance in a network of partially replicated database systems
US6487713B1 (en) * 1999-09-24 2002-11-26 Phoenix Technologies Ltd. Software development system that presents a logical view of project components, facilitates their selection, and signals missing links prior to compilation
US20030144828A1 (en) * 2001-07-30 2003-07-31 Lin Sharon Sheau-Pyng Hub array system and method
US20100064111A1 (en) * 2008-09-09 2010-03-11 Kabushiki Kaisha Toshiba Information processing device including memory management device managing access from processor to memory and memory management method
US20120191900A1 (en) * 2009-07-17 2012-07-26 Atsushi Kunimatsu Memory management device
US20180302443A1 (en) * 2012-08-02 2018-10-18 CellSec, Inc. Dividing a data processing device into separate security domains
US20160179376A1 (en) * 2014-12-23 2016-06-23 Intel Corporation Method and apparatus for improving read performance of a solid state drive
US20160283116A1 (en) * 2015-03-27 2016-09-29 Intel Corporation Sequential write stream management
US20160313943A1 (en) * 2015-04-24 2016-10-27 Kabushiki Kaisha Toshiba Storage device that secures a block for a stream or namespace and system having the storage device
US9696935B2 (en) * 2015-04-24 2017-07-04 Kabushiki Kaisha Toshiba Storage device that secures a block for a stream or namespace and system having the storage device
US20190138212A1 (en) * 2015-04-24 2019-05-09 Toshiba Memory Corporation Storage device that secures a block for a stream or namespace and system having the storage device
US20180074709A1 (en) * 2016-09-15 2018-03-15 Western Digital Technologies, Inc. Stream management for storage devices
US10031689B2 (en) * 2016-09-15 2018-07-24 Western Digital Technologies, Inc. Stream management for storage devices
US20190155521A1 (en) * 2017-11-21 2019-05-23 Micron Technology, Inc. Data categorization based on invalidation velocities

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11531615B2 (en) * 2018-12-14 2022-12-20 SK Hynix Inc. Controller and memory system for performing garbage collection operation, and operating method thereof
CN114238232A (en) * 2021-12-15 2022-03-25 南方电网电力科技股份有限公司 Intelligent ammeter file reading and writing system, method, equipment and medium

Also Published As

Publication number Publication date
US11231873B2 (en) 2022-01-25

Similar Documents

Publication Publication Date Title
US11960393B2 (en) Data processing method and apparatus, and flash device
US10275162B2 (en) Methods and systems for managing data migration in solid state non-volatile memory
US10732905B2 (en) Automatic I/O stream selection for storage devices
US8918581B2 (en) Enhancing the lifetime and performance of flash-based storage
EP3118745B1 (en) A heuristic interface for enabling a computer device to utilize data property-based data placement inside a nonvolatile memory device
US11030094B2 (en) Apparatus and method for performing garbage collection by predicting required time
US20180121351A1 (en) Storage system, storage management apparatus, storage device, hybrid storage apparatus, and storage management method
US9384123B2 (en) Memory system
US9507732B1 (en) System and method for cache management
US8898410B1 (en) Efficient garbage collection in a data storage device
US10048876B2 (en) Method for providing nonvolatile storage write bandwidth using a caching namespace
US20150370700A1 (en) Managing storage devices
US9891833B2 (en) Eliminating garbage collection in nand flash devices
US10140031B2 (en) Hierarchical flash translation layer structure and method for designing the same
US10120583B2 (en) Performance penalty avoidance for solid state drive
US11231873B2 (en) Apparatus and method for assigning velocities to write data
CN111033478A (en) Dynamic TRIM processing using disk cache
JP2015001909A (en) Information processor, control circuit, control program, and control method
CN114096952A (en) Adjusting block pool size in a storage system
US11138104B2 (en) Selection of mass storage device streams for garbage collection based on logical saturation
US10628301B1 (en) System and method for optimizing write amplification of non-volatile memory storage media
US10108350B2 (en) Method for providing nonvolatile storage write bandwidth using a caching namespace
US11093408B1 (en) System and method for optimizing write amplification of non-volatile memory storage media
US20230195351A1 (en) Automatic deletion in a persistent storage device
US11074173B1 (en) Method and system to determine an optimal over-provisioning ratio

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTEL CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TRIKA, SANJEEV N.;REEL/FRAME:047708/0561

Effective date: 20181206

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: SK HYNIX NAND PRODUCT SOLUTIONS CORP., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTEL CORPORATION;REEL/FRAME:063815/0490

Effective date: 20211229