WO2014137842A1 - Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive - Google Patents

Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive Download PDF

Info

Publication number
WO2014137842A1
WO2014137842A1 PCT/US2014/019681 US2014019681W WO2014137842A1 WO 2014137842 A1 WO2014137842 A1 WO 2014137842A1 US 2014019681 W US2014019681 W US 2014019681W WO 2014137842 A1 WO2014137842 A1 WO 2014137842A1
Authority
WO
WIPO (PCT)
Prior art keywords
page
map
logical
volatile memory
data
Prior art date
Application number
PCT/US2014/019681
Other languages
French (fr)
Inventor
Andrew J. Tomlin
Justin Jones
Rodney N. Mullendore
Original Assignee
Western Digital Technologies, Inc.
Skyera, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Western Digital Technologies, Inc., Skyera, Inc. filed Critical Western Digital Technologies, Inc.
Priority to AU2014226246A priority Critical patent/AU2014226246B2/en
Priority to KR1020187031705A priority patent/KR101965549B1/en
Priority to EP14760216.3A priority patent/EP2965208B1/en
Priority to CN201480011821.3A priority patent/CN105009094B/en
Priority to KR1020157027180A priority patent/KR101916206B1/en
Publication of WO2014137842A1 publication Critical patent/WO2014137842A1/en
Priority to HK16103997.9A priority patent/HK1216197A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/023Free address space management
    • G06F12/0238Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory
    • G06F12/0246Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory in block erasable memory, e.g. flash memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0604Improving or facilitating administration, e.g. storage management
    • G06F3/0607Improving or facilitating administration, e.g. storage management by facilitating the process of upgrading existing storage systems, e.g. for improving compatibility between host and storage device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0655Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
    • G06F3/0659Command handling arrangements, e.g. command buffers, queues, command scheduling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • G06F3/0688Non-volatile semiconductor memory arrays
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/0292User address space allocation, e.g. contiguous or non contiguous base addressing using tables or multilevel address translation means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/72Details relating to flash memory management
    • G06F2212/7201Logical to physical mapping or translation of blocks or pages
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/72Details relating to flash memory management
    • G06F2212/7205Cleaning, compaction, garbage collection, erase control

Definitions

  • SSDs solid state drives
  • a page in an SSD is typically 8-16 kilobytes (KB) in size and a block consists of a large number of pages (e.g., 256 or 512).
  • KB kilobytes
  • a particular physical location in an SSD e.g., a page
  • address indirection is needed.
  • L2P Logical-to-Physical mapping system known as logical block addressing (LBA) that is part of the Flash translation layer (FTL).
  • LBA logical block addressing
  • FTL Flash translation layer
  • a large L2P map table maps logical entries to physical address locations on an SSD.
  • This large L2P map table is usually saved in small sections as writes come in. For example, if random writing occurs, although the system may have to update only one entry, it may nonetheless have to save the entire table or a portion thereof, including entries that have not been updated, which is inherently inefficient.
  • Fig. 1 shows aspects of a conventional Logical Block Addressing (LBA) scheme for data storage devices.
  • LBA Logical Block Addressing
  • a map table 104 contains one entry for every logical block 102 defined for the data storage device's Flash memory 106.
  • a 64 GB data storage device that supports 512 byte logical blocks may present itself to the host as having 125,000,000 logical blocks.
  • One entry in the map table 104 contains the current location of each of the 125,000 logical blocks in the Flash memory 106.
  • a Flash page holds an integer number of logical blocks (i.e., a logical block does not span across Flash pages). In this conventional example, an 8 KB Flash page would hold 16 logical blocks (of size 512 bytes).
  • each entry in the logical-to-physical map table 1 04 contains a field 108 identifying the die on which the LBA is stored, a field 1 1 0 identifying the flash block on which the LBA is stored, another field 1 12 identifying the flash page within the flash block and a field 1 14 identifying the offset within the flash page that identifies where the LBA data begins in the identified Flash page.
  • the large size of the map table 104 prevents the table from being held inside the SSD controller. Conventionally, the large map table 1 04 is held in an external DRAM connected to the SSD controller. As the map table 104 is stored in volatile DRAM, it must be restored when the SSD powers up, which can take a long time, due to the large size of the table.
  • the corresponding entry in the map table 104 is updated to reflect the new location of the logical block.
  • the corresponding entry in the map table 104 is read to determine the location in Flash memory to be read.
  • a read is then performed to the Flash page specified in the corresponding entry in the map table 104.
  • the read data is available for the Flash page, the data at the offset specified by the Map Entry is transferred from the Flash device to the host.
  • the Flash memory holding the "old" version of the data becomes "garbage" (i.e., data that is no longer valid).
  • the Flash memory when a logical block is written, the Flash memory will initially contain at least two versions of the logical block; namely, the valid, most recently written version (pointed to by the map table 104) and at least one other, older version thereof that is stale and is no longer pointed to by any entry in the map table 104.
  • These "stale" entries are referred to as garbage, which occupies space that must be accounted for, collected, erased and made available for future use. This process is known as "garbage collection”.
  • the conventional method of rebuilding the map table 104 includes first initializing all memory locations within the volatile memory to zero, as RAM powers on in a random state (i.e., without valid ECC). Thereafter, the L2P map may be rebuilt from mapping information saved in the non-volatile memory while, at the same time, carrying out free space accounting. These operations involve lengthy random reads and random writes, which take a significant time, which ultimately delays the drive's time-to-ready for host reads and writes on power up. Conventionally, both the map rebuilding and the free space accounting must be completed before the drive is able to report to the host that it is ready to service data access commands.
  • Fig. 1 shows aspects of a conventional Logical Block Addressing scheme for SSDs.
  • FIG. 2 is a block diagram of a data storage device according to one embodiment.
  • FIG. 3 is a diagram showing aspects of the physical and logical data organization of a data storage device according to one embodiment.
  • Fig. 4 shows a logical-to-physical address translation map and illustrative entries thereof, according to one embodiment.
  • Fig. 5 shows aspects of a method for updating a logical-to- physical address translation map and for creating an S-Journal entry, according to one embodiment.
  • Fig. 6 is a block diagram of an S-Journal, according to one embodiment.
  • Fig. 7 shows an exemplary organization of one entry of an S- Journal, according to one embodiment.
  • Fig. 8 is a block diagram of a superblock (S-Block), according to one embodiment.
  • Fig. 9 shows another view of a Super page (S-page), according to one embodiment.
  • Fig. 10A shows relationships between the logical-to-physical address translation map, S-Journals and S-Blocks, according to one embodiment.
  • Fig. 10B is a block diagram of an S-Journal Map, according to one embodiment.
  • Fig. 1 1 is a flowchart of a method of carrying out a two-stage power-on logical-to-physical map rebuild, according to one embodiment.
  • FIG. 12 is a block diagram illustrating further aspects of a method of carrying out a two-stage power-on logical-to-physical map rebuild, according to one embodiment.
  • Fig. 13 is a representation of a valid size information table, according to one embodiment.
  • Fig. 14 is a block diagram illustrating aspects of a method of carrying out a two-stage power-on logical-to-physical map rebuild with free space accounting, according to one embodiment.
  • Fig. 15 is a block diagram illustrating aspects of garbage collection, according to one embodiment.
  • FIG. 1 6 is a block diagram illustrating further aspects of garbage collection, according to one embodiment.
  • FIG. 17 is a block diagram illustrating still further aspects of garbage collection, according to one embodiment.
  • FIG. 18 is a block diagram illustrating yet further aspects of garbage collection, according to one embodiment.
  • Fig. 2 is a diagram showing aspects of the physical and logical data organization of a data storage device according to one embodiment.
  • the data storage device is an SSD.
  • the data storage device is a hybrid drive including Flash memory and rotating magnetic storage media.
  • the disclosure is applicable to both SSD and hybrid implementations, but for the sake of simplicity the various embodiments are described with reference to SSD-based implementations.
  • a data storage device controller 202 may be configured to be coupled to a host, as shown at reference numeral 21 8.
  • the controller may comprise one or more processors that execute some or all of the functions described below as being performed by the controller.
  • the host 218 may utilize a logical block addressing (LBA) scheme.
  • LBA logical block addressing
  • the host can vary the size of the LBA dynamically.
  • the LBA size may vary by interface and interface mode. Indeed, while 512 bytes is most common, 4 KB is also becoming more common, as are 512+ (520, 528, etc.) and 4 KB+ (4 KB+8, 4K+16, etc.) formats.
  • the data storage device controller 202 may comprise or be coupled to a page register 204.
  • the page register 204 may be configured to enable the controller 202 to read data from and store data to the data storage device.
  • the controller 202 may be configured to program and read data from an array of flash memory devices responsive to data access commands from the host 218.
  • the array of memory devices may comprise other types of non-volatile memory devices such as flash integrated circuits, Chalcogenide RAM (C-RAM), Phase Change Memory (PC-RAM or PRAM), Programmable Metallization Cell RAM (PMC-RAM or PMCm), Ovonic Unified Memory (OUM), Resistance RAM (RRAM), NAND memory (e.g., single-level cell (SLC) memory, multi-level cell (MLC) memory, or any combination thereof), NOR memory, EEPROM, Ferroelectric Memory (FeRAM), Magnetoresistive RAM (MRAM), other discrete NVM (non-volatile memory) chips, or any combination thereof.
  • flash integrated circuits e.g., Chalcogenide RAM (C-RAM), Phase Change Memory (PC-RAM or PRAM), Programmable Metallization Cell RAM (PMC-RAM or PMCm), Ovonic Unified Memory (OUM), Resistance RAM (RRAM), NAND memory (e.g., single-level cell (SLC) memory, multi-level cell (MLC) memory,
  • the page register 204 may be configured to enable the controller 202 to read data from and store data to the array.
  • the array of flash memory devices may comprise a plurality of nonvolatile memory devices in die (e.g., 128 dies), each of which comprises a plurality of blocks, such as shown at 206 in Fig. 2.
  • Other page registers 204 may be coupled to blocks on other die.
  • a combination of Flash blocks, grouped together, may be called a Superblock or S-Block.
  • the individual blocks that form an S-Block may be chosen from one or more dies, planes or other levels of granularity.
  • An S-Block may comprise a plurality of Flash blocks, spread across one or more die, that are combined together.
  • the S-Block may form a unit on which the Flash Management System (FMS) operates.
  • the individual blocks that form an S-Block may be chosen according to a different granularity than at the die level, such as the case when the memory devices include dies that are sub-divided into structures such as planes (i.e., blocks may be taken from individual planes).
  • allocation, erasure and garbage collection may be carried out at the S-Block level.
  • the FMS may perform data operations according to other logical groupings such as pages, blocks, planes, dies, etc.
  • each of the Flash blocks 206 comprises a plurality of Flash pages (F-Pages) 208.
  • Each F-Page may be of a fixed size such as, for example, 16 KB.
  • the F-Page is the size of the minimum unit of program for a given Flash device.
  • each F-Page 208 may be configured to accommodate a plurality of physical pages, hereinafter referred to as E-Pages 210.
  • the term ⁇ -Page refers to a data structure stored in Flash memory on which an error correcting code (ECC) has been applied.
  • E-Page 210 may form the basis for physical addressing within the data storage device and may constitute the minimum unit of Flash read data transfer.
  • the E-Page 210 may be (but need not be) of a predetermined fixed size (such as 2 KB, for example) and determine the size of the payload (e.g., host data) of the ECC system.
  • each F- Page 208 may be configured to fit a predetermined plurality of E-Pages 210 within its boundaries. For example, given 16 KB size F-Pages 208 and a fixed size of 2 KB per E-Page 21 0, eight E-Pages 210 fit within a single F-Page 208, as shown in Fig. 3.
  • a power of 2 multiple of E-Pages 210, including ECC may be configured to fit into an F-Page 208.
  • Each E-Page 21 0 may comprise a data portion 214 and, depending on where the E-Page 210 is located, may also comprise an ECC portion 21 6. Neither the data portion 214 nor the ECC portion 216 need be fixed in size.
  • the address of an E-Page uniquely identifies the location of the E-Page within the Flash memory. For example, the E- Page's address may specify the Flash channel, a particular die within the identified Flash channel, a particular block within the die, a particular F-Page and, finally, the E-Page within the identified F-Page.
  • L-Page logical page
  • An L-Page, denoted in Fig. 3 at reference numeral 212 may comprise the minimum unit of address translation used by the FMS.
  • Each L-Page may be associated with an L-Page number.
  • the L-Page numbers of L-Pages 212 therefore, may be configured to enable the controller 202 to logically reference host data stored in one or more of the physical pages, such as the E-Pages 21 0.
  • the L-Page 212 may also be utilized as the basic unit of compression.
  • L-Pages 212 are not fixed in size and may vary in size, due to variability in the compression of data to be stored. Since the compressibility of data varies, a 4 KB amount of data of one type may be compressed into a 2 KB L-Page while a 4 KB amount of data of a different type may be compressed into a 1 KB L-Page, for example. Due to such compression, therefore, the size of L-Pages may vary within a range defined by a minimum compressed size of, for example, 24 bytes to a maximum uncompressed size of, for example, 4 KB or 4 KB+. Other sizes and ranges may be implemented. As shown in Fig.
  • L-Pages 212 need not be aligned with the boundaries of E-Page 210. Indeed, L-Pages 212 may be configured to have a starting address that is aligned with an F-Page 208 and/or E-Page 210 boundary, but also may be configured to be unaligned with either of the boundaries of an F- Page 208 or E-Page 210. That is, an L-Page starting address may be located at a non-zero offset from either the start or ending addresses of the F-Pages 208 or the start or ending addresses of the E-Pages 210, as shown in Fig. 3.
  • L-Pages 212 are not fixed in size and may be smaller than the fixed-size E-Pages 210, more than one L-Page 212 may fit within a single E-Page 210. Similarly, as the L-Pages 212 may be larger in size than the E-Pages 21 0, L-Pages 212 may span more than one E-Page, and may even cross the boundaries of F-Pages 208, shown in Fig. 3 at numeral 21 7.
  • LBA size is 51 2 or 512+ bytes
  • a maximum of, for example, eight sequential LBAs may be packed into a 4 KB L-Page 212, given that an uncompressed L-Page 212 may be 4 KB to 4 KB+.
  • the exact logical size of an L-Page 212 is unimportant as, after compression, the physical size may span from few bytes at minimum size to thousands of bytes at full size. For example, for 4TB SSD device, 30 bits of addressing may be used to address each L-Page 212 that could potentially be present in such a SSD.
  • Fig. 4 shows a logical-to-physical address translation map and illustrative entries thereof, according to one embodiment.
  • a logical-to-physical address translation map is required to enable the controller 202 to associate an L-Page number of an L-Page 212 to one or more E-Pages 210.
  • Such a logical-to-physical address translation map is shown in Fig. 4 at 302 and, in one embodiment, is a linear array having one entry per L-Page 212.
  • Such a logical-to-physical address translation map 302 may be stored in a volatile memory, such as a DRAM or SRAM.
  • Fig. 4 also shows the entries in the logical-to-physical address translation map for four different L-Pages 21 2, which L-Pages 212 in Fig. 4 are associated with L-Page numbers denoted as L-Page 1 , L-Page 2, L-Page 3 and L-Page 4.
  • each L-Page stored in the data storage device may be pointed to by a single and unique entry in the logical-to-physical address translation map 302. Accordingly, in the example being developed herewith, four entries are shown.
  • each entry in the map 302 may comprise an L-Page number, which may comprise an identification of the physical page (e.g., E-Page) containing the start address of the L-Page being referenced, the offset of the start address within the physical page (e.g., E-Page) and the length of the L-Page.
  • a plurality of ECC bits may provide error correction functionality for the map entry. For example, and as shown in Fig. 4, and assuming an E-Page size of 2 KB, L-Page 1 may be referenced in the logical-to-physical address translation map 302 as follows: E-Page 1 003, offset 800, length 1 624, followed by a predetermined number of ECC bits (not shown).
  • the start of L-Page 1 is within (not aligned with) E-Page 1 003, and is located at an offset from the starting physical location of the E-Page 1003 that is equal to 800 bytes.
  • Compressed L-Page 1 furthermore, extends 1 ,624 bytes, thereby crossing an E-Page boundary to E-Page 1004. Therefore, E-Pages 1 003 and 1004 each store a portion of the L-Page 212 denoted by L-Page number L-Page 1 .
  • the compressed L-Page referenced by L-Page number L-Page 2 is stored entirely within E-Page 1004, and begins at an offset therein of 400 bytes and extends only 696 bytes within E-Page 1 004.
  • the compressed L-Page associated with L-Page number L-Page 3 starts within E-Page 1004 at an offset of 1 ,120 bytes (just 24 bytes away from the boundary of L-Page 2) and extends 4,096 bytes past E-Page 1005 and into E-Page 1006. Therefore, the L- Page associated with L-Page number L-Page 3 spans a portion of E-Page 1004, all of E-Page 1005 and a portion of E-Page 1006.
  • L-Page associated with L-Page number L-Page 4 begins within E-Page 1006 at an offset of 1 ,144 bytes, and extends 3, 128 bytes to fully span E-Page 1007, crossing an F-Page boundary into E- Page 1 008 of the next F-Page.
  • each of these constituent identifier fields (E-Page, offset, length and ECC) making up each entry of the logical-to-physical address translation map 302 may be, for example, 8 bytes in size. That is, for an exemplary 4 TB drive, the address of the E-Page may be 32 bits in size, the offset may be 12 bits (for E-Page data portions up to 4 KB) in size, the length may be 10 bits in size and the ECC field may be provided. Other organizations and bit-widths are possible. Such an 8 byte entry may be created each time an L-Page is written or modified, to enable the controller 202 to keep track of the host data, written in L-Pages, within the Flash storage.
  • This 8-byte entry in the logical-to-physical address translation map may be indexed by an L-Page number or LPN.
  • the L-Page number functions as an index into the logical-to-physical address translation map 302.
  • the LBA is the same as the LPN.
  • the LPN therefore, may constitute the address of the entry within the volatile memory.
  • the controller 202 receives a read command from the host 218, the LPN may be derived from the supplied LBA and used to index into the logical-to-physical address translation map 302 to extract the location of the data to be read in the Flash memory.
  • the LPN When the controller 202 receives a write command from the host, the LPN may be constructed from the LBA and the logical-to-physical address translation map 302 may be modified. For example, a new entry therein may be created. Depending upon the size of the volatile memory storing the logical-to-physical address translation map 302, the LPN may be stored in a single entry or broken into, for example, a first entry identifying the E-Page containing the starting address of the L-Page in question (plus ECC bits) and a second entry identifying the offset and length (plus ECC bits). According to one embodiment, therefore, these two entries may together correspond and point to a single L-Page within the Flash memory. In other embodiments, the specific format of the logical-to-physical address translation map entries may be different from the examples shown above.
  • the logical-to-physical address translation map 302 may be stored in a volatile memory, it necessarily must be rebuilt upon startup or any other loss of power to the volatile memory. This, therefore, requires some mechanism and information to be stored in a non-volatile memory that will enable the controller 202 to reconstruct the logical-to-physical address translation map 302 before the controller can "know" where the L-Pages are stored in the non-volatile memory after startup or after a power-fail event. According to one embodiment, such mechanism and information may be embodied in a construct that may be called a System Journal, or S-Journal.
  • the controller 202 may be configured to maintain, in the plurality of non-volatile memory devices (e.g., in one or more of the blocks 206 in one or more die, channel or plane), a plurality of S- Journals defining physical-to-logical address correspondences.
  • each S-Journal may cover a pre-determined range of physical pages (e.g., E-Pages).
  • each S-Journal may comprise a plurality of journal entries, with each entry being configured to associate one or more physical pages, such as E-Pages, to the L-Page number of each L-Page.
  • Fig. 5 shows aspects of a method for updating a logical-to- physical address translation map and for creating an S-Journal entry, according to one embodiment.
  • the logical-to-physical address translation map 302 may be updated as shown at B52.
  • an S-Journal entry may also be created, storing therein information pointing to the location of the updated L-Page.
  • both the logical-to-physical address translation map 302 and the S- Journals are updated when new writes occur (e.g., as the host issues writes to nonvolatile memory, as garbage collection/wear leveling occurs, etc.).
  • Write operations to the non-volatile memory devices to maintain a power-safe copy of address translation data may be configured, therefore, to be triggered by newly created S-Journal entries (which may be just a few bytes in size) instead of re-saving all or a portion of the logical-to-physical address translation map, such that Write Amplification (WA) is reduced.
  • WA Write Amplification
  • the updating of the S-Journals ensure that the controller 202 can access a newly updated L-Page and that the logical-to-physical address translation map 302 may be reconstructed upon restart or other information- erasing power event affecting the non-volatile memory in which the logical-to- physical address translation map is stored.
  • the S-Journals are useful in enabling effective Garbage Collection (GC). Indeed, the S-Journals may contain the last-in-time update to all L-Page numbers, and may also contain stale entries, entries that do not point to a valid L-Page.
  • the S-Journal may constitute the main flash management data written to the media.
  • S- Journals may contain mapping information for a given S-Block and may contain the Physical-to-Logical (P2L) information for a given S-Block.
  • Fig. 6 is a block diagram showing aspects of an S-Journal, according to one embodiment. As shown therein and according to one embodiment, each S-Journal 602 covers a predetermined physical region of the non-volatile memory such as, for example, 32 E-Pages as shown at 606, which are addressable using 5 bits. Each S-Journal 602 may be identified by an S-Journal Number 604.
  • the S-Journal Number 604 used for storing P2L information for host data may comprise a portion of the address of the first physical page covered by the S-Journal.
  • the S-Journal Number of S- Journal 602 may comprise, for example, the 27 MSbs of the first E-Page covered by this S-Journal 602.
  • Fig. 7 shows an exemplary organization of one entry 702 of an S-Journal 602, according to one embodiment.
  • Each entry 702 of the S-Journal 602 may point to the starting address of one L-Page, which is physically addressed in E- Pages.
  • Each entry 702 may comprise, for example, a number (5, for example) of LSbs of the E-Page containing the starting E-Page of the L-Page.
  • the full E-Page address may be obtained by concatenating these 5 LSbs with the 27 MSbs of the S- Journal Number in the header.
  • the entry 702 may then comprise the L-Page number, its offset within the identified E-Page and its size.
  • each entry 702 of a data band S-Journal 602 may comprise the 5 LSbs of the first E-Page covered by this S-Journal entry, 30 bits of L-Page number, 9 bits of E-Page offset and 10 bits of L-Page size, adding up to an overall size of about 7 bytes.
  • Various other internal journal entry formats may be used in other embodiments.
  • a variable number of L-Pages may be stored in a physical area, such as a physical area equal to 32 E-Pages, as shown at 606 in Fig. 6.
  • S-Journals 602 may comprise a variable number of entries.
  • an L-Page may be 24 bytes in size and an S-Journal 602 may comprise over 2,500 entries, referencing an equal number of L-Pages, one L- Page per S-Journal entry 702.
  • S-Journals 602 may be configured to contain mapping information for a given S-Block and may contain the P2L information for a given S-Block. More precisely, according to one embodiment, S-Journals 602 may contain the mapping information for a predetermined range of E-Pages within a given S-Block.
  • Fig. 8 is a block diagram of a superblock (S-Block), according to one embodiment. As shown therein, an S-Block 802 may comprise one Flash block (F- Block) 804 (as also shown at 206 in Fig. 2) per die.
  • F- Block Flash block
  • An S-Block 802 may be thought of as a collection of F-Blocks 804, one F-Block per die, that are combined together to form a unit of the Flash Management System.
  • allocation, erasure and GC may be managed at the Superblock level.
  • Each F-Block 804, as shown in Fig. 8, may comprise a plurality of Flash pages (F- Page) such as, for example, 256 or 512 F-Pages.
  • An F-Page may be the size of the minimum unit of program for a given non-volatile memory device.
  • Fig. 9 shows a Super Page (S-Page), according to one embodiment.
  • an S-Page 803 may comprise one F-Page per block of an S-Block, meaning that an S-Page 803 spans across an entire S-Block 802.
  • Fig. 10A shows relationships between the logical-to-physical address translation map, S-Journals and S-Blocks, according to one embodiment.
  • Reference 902 denotes the logical-to-physical address translation map.
  • the logical-to-physical address translation map 902 may be indexed by L-Page number, in that there may be one entry in the logical-to-physical address translation map 902 per L-Page in the logical-to-physical address translation map.
  • the physical address of the start of the L-Page in the Flash memory and the size thereof may be given in the map entry; namely by E-Page address, offset within the E-Page and the size of the L-Page.
  • the L-Page depending upon its size, may span one or more E-Pages and may span F-Pages and blocks as well.
  • the volatile memory may also store an S-Journal map.
  • An entry in the S-Journal map 904 stores information related to where an S-Journal is physically located in the non-volatile memory. For example, the 27 MSbs of the E-Page physical address where the start of the L-Page is stored may constitute the S-Journal Number.
  • the S-Journal map 904 in the volatile memory may also include the address of the S-Journal in non-volatile memory, referenced in system E-Pages. From the S-Journal map 904 in volatile memory, System S-Block Information 908 may be extracted.
  • the System S-Block Information 908 may be indexed by System S-Block (S-Block in the System Band) and may comprise, among other information regarding the S-Block, the size of any free or used space in the System S-Block. Also from the S-Journal map 904, the physical location of the S-Journals in non-volatile memory 910 may be extracted.
  • the System Band does not contain L-Page data and may contain all File Management System (FMS) meta-data and information.
  • the System Band may be configured as lower-page only for reliability and power fail simplification. During normal operation, the System Band need not be read except during Garbage Collection.
  • the System Band may be provided with significantly higher overprovisioning than the data band for overall WA optimization.
  • Other bands may include the Hot Band, which may contain L-Page data and is frequently updated, and the Cold Band, which is a physical area of memory storing static data retained from the garbage collection process, which may be infrequently updated.
  • the System, Hot and Cold Bands may be allocated by controller firmware on an S-Block basis.
  • each of these S-Journals in non-volatile memory may comprise a collection of S-Journal entries and cover, for example, 32 E-Pages worth of data.
  • These S-Journals in non-volatile memory 910 enable the controller 202 to access the S-Journals entries in non-volatile memory upon startup, enable the controller 202 to rebuild in volatile memory not only the logical-to-physical address translation map 902, but also the S-Journal map 904, the S-Block Information 906, and the System S-Block Information 908.
  • the S-Journals in non-volatile memory may also contain all of the stale L-Page information, thereby enabling the controller 202 to GC after the logical-to-physical address translation map 902 and the S-Journal Map 904 in volatile memory are rebuilt.
  • the S-Journals therefore, may be said to contain a sequential history of all updates, over time, to the logical-to-physical address translation map 902.
  • Fig. 10B is a block diagram of another view of an S-Journal Map 904, according to one embodiment.
  • the S-Journal Map 904 may reference a plurality of S-Journal entries for each S-Block.
  • the S- Block Number may be the MSb of the S-Journal Number.
  • the size of the S-Journal map 904 may be correlated to the number of S-Blocks times the number of S- Journal entries per S-Block. Indexing into the S-Journal Map 904, therefore, may be carried out by referencing the S-Block Number (the MSb of the S-Journal Number) and the S-Journal entry for that S-Block number.
  • the controller 202 may be further configured to build or rebuild a map of the S-Journals and store the resulting S-Journal Map 904 in volatile memory. For example, upon restart or upon the occurrence of another event in which power fails or after a restart subsequent to error recovery, the controller 202 may read the plurality of S-Journals in a predetermined sequential order, build a map of the S-Journals stored in the nonvolatile memory devices based upon the sequentially read plurality of S-Journals, and store the built S-Journal Map 904 in the volatile memory.
  • the rebuilt S-Journal Map 904 may be configured to contain the physical location for the most recently-written version of each S-Journal.
  • the S-Journal Map 904 may be rebuilt by the processor 202 based upon read S-Journals that are determined to be valid.
  • the address translation map rebuilding and the free space accounting may be split into two distinct stages.
  • the address translation map may be rebuilt from the S-Journals stored in the System Band of the non-volatile memory.
  • the System Band may be that portion of the non-volatile memory that contains the FMS meta-data and information.
  • the System Band may be configured for lower page only operation, thereby avoiding any possibility of lower page corruption of MLC Flash memory.
  • the System Band may be significantly more over-provisioned than the Data Band (which stores all L-Page data and contains both the Hot and Cold Bands) for overall WA optimization.
  • all of the S-Journals may be read. Reading all of the S-Journals enables the controller 202 to extract, from each entry of each S-Journal, the size and exact physical location of each L- Page associated with each L-Page Number. In one embodiment, after all S-Journals are read and the physical-to-logical information extracted therefrom and is stored in the logical-to-physical address translation map 902 in volatile memory, the controller 202 will have accounted for and mapped all of the L-pages stored in the entire nonvolatile memory.
  • S-Journal construct may be used to store information related to a virtual address range and for storing other file management tables in the volatile memory.
  • the S-Journals may be configured to collectively contain all of the physical-to-logical information, and since many L-pages are likely to have been updated many times, the order in which the S- Journals are read is significant. Indeed, as the S-Journals may be configured to collectively store a complete history of updates to the L-Pages (at least since a last garbage collecting of the S-Block in which the L-Pages are stored), the S-Journals are read in one embodiment in the order in which they were created, as indicated at Block B1 51 . This is graphically shown in Fig.
  • the controller 202 may be configured, at startup, to read each of the plurality of S-Journals in the order in which they were created as shown at B1 51 and to rebuild the logical-to- physical address translation map 902 in the volatile memory from the read plurality of journals. That is, the controller 202 may, from the read S-Journal entries, populate the logical-to-physical address translation map 902 with the physical location information (e.g., a length and starting address within an E-Page in non-volatile memory) of each L-Page, such as also shown at 302 in Fig. 4.
  • the physical location information e.g., a length and starting address within an E-Page in non-volatile memory
  • each L-Page stored in the data storage device may be pointed to by a single and unique entry in the logical-to-physical address translation map 902.
  • the controller 202 may carry out a write operation (e.g., an 8 byte write) to the volatile memory for each entry in each of the read S-Journals. It is to be understood, however, that embodiments are not limited to the exemplary 8 byte size shown and described herein and that other implementations are possible.
  • the data storage device may report its readiness to process data access commands to the host(s) 218, as shown at B152. Note that the data storage device, at the completion of this first stage, has not yet carried out any free space accounting operations and thus may not yet be ready to optimally carry out garbage collecting operations.
  • the controller 202 at the end of this first stage, does not yet know which is the optimal S-Block to garbage collect, as the free space accounting has not yet been carried out.
  • One of the considerations in choosing an S-Block for garbage collection activities may comprise ascertaining the amount of free space available in each S-Block. For example, the controller 202 may select an S-Block having the greatest amount of free space as the next S-Block on which to carry out garbage collection activities. Note that, according to one embodiment, the controller 202 may also identify the next best S- Block to garbage collect, as well as the next best S-Block after that, and so on.
  • the second stage of the present two-stage power on procedure may be carried out.
  • one or more tables may be rebuilt from the rebuilt logical-to-physical address translation map 902 and, based on such rebuilt table(s), one or more of the plurality of S-Blocks may be selected for garbage collection, as shown at B153 in Fig. 1 1 .
  • the controller 202 may, among other possible actions, rebuild a free space table containing an amount of free space in each of the plurality of S-Blocks after the logical-to-physical address translation map 902 has been rebuilt. To rebuild such free space table(s), the amount of free space in each of the plurality of blocks may be calculated. Instead of the controller 202 tracking the amount of free space in each S-Block, the controller 202 may, according to one embodiment, track the amount of valid data in each S-Block. Such information for each S-Block may be readily available in the logical-to-physical address translation map 902, which may be configured to contain a record of the size of each L-page in the non-volatile memory.
  • the controller 202 may calculate the amount of free space in each S-Block by subtracting the size of the valid data in each S-Block (obtained from reading the logical-to-physical address translation map 902) from the predetermined size of the S-Blocks.
  • the difference between the amount of valid data stored in an S-Block and its size may be equal to or related to the amount of free space within each S-Block.
  • the amount of free space in each S-Block may be calculated by the controller 202.
  • the accumulated (i.e., summed across entries in the logical-to-physical address translation map 902) amount of valid space in each S-Block may be stored, for example, in a table maintained in the volatile memory.
  • the system block information table 908 (see, for example, Fig. 10A) may be used for that purpose.
  • the controller 202 may create and populate a separate valid size information table with the accumulated size of valid data for each S-Block.
  • S- Block 1 has a total of 450 MB of valid data stored therein and S-Block 2 has a total of 1 .5 TB of valid data stored therein.
  • the amount of free space in each S-Block may be readily calculated.
  • the amount of free space calculation is performed by directly summing the amount of free space based on the information from the logical-to-physical translation map 902.
  • the data storage device reports that it is ready to process host commands shortly after having completed the rebuild of the logical-to-physical address translation map 902 (and optionally other housekeeping activities such as populating the S-Journal map 904 and other tables in volatile memory).
  • the data storage device is configured to carry out free space accounting operations (including, e.g., rebuilding the free space table(s)) while and/or after processing host (i.e., data access) commands.
  • Such incoming host commands may alter the free space accounting of the S-Blocks while the second stage is performed.
  • One embodiment accounts for such changes in the amount of valid data that is present in each S-Block.
  • the map 902 may be sequentially read, from a beginning to the end thereof.
  • the controller 202 may be configured to track a pointer during the reading of the now-coherent logical-to-physical address translation map 902, as shown at 1902 in Fig. 14.
  • the controller 202 may be configured to act differently depending whether the L-Page that is modified as a result of a write command is associated with an L-Page Number that has already been processed for free space accounting by the controller 202 or that is associated with an L-Page Number that has not yet been read by the controller 202 as it sequentially reads the logical-to-physical address translation map 902. According to one embodiment, if the L-Page that is modified as a result of a write command is associated with an L-Page Number that has already been processed for free space by the controller 202, as suggested by the "New Write 1 " at reference numeral 1904 in Fig. 14, the controller 202 may carry out free space accounting in the manner described above.
  • This free space accounting may be necessary in this case, as the controller 202 has already processed the entry corresponding to the now-updated L-Page for free space and has added the length thereof to its running count of valid data for the S-Block where the updated L-Page is stored. It is possible that the updated length is the same as the previous length of the L-page, in which case the running count of the valid data would be unchanged if the L-Page is written to the same S-Block. However, the accumulated size of the data in the S-Block would indeed change if the executed write resulted in a change to the length of the L-page or if written to a different S-Block. To maintain an accurate accounting of the free space in each S-Block, therefore, requires that the controller 202 go back and properly account for the free space made available as a result of the update to the L- Page.
  • the controller 202 allows the controller 202 to refrain from carrying out free space accounting activities and continue with its sequential read of the logical-to-physical address translation map 902.
  • the blocks 206 of the non-volatile memory may comprise Multi-Level Cells (MLC) that store, for example, 2 bits per cell.
  • MLC Multi-Level Cells
  • For an MLC device there is at least a lower page and an upper page, which may not be contiguous.
  • the lower page of an MLC is typically programmed before the upper page thereof.
  • the charge stored by the MLC is gradually increased, in a step-wise manner, until the desired voltage indicative of the desired logic state is reached. This is the reason that, according to one embodiment, the System Band where the S-Journals are stored may be configured to only use the lower pages in a block for increased reliability.
  • one embodiment utilizes a non-volatile buffer to temporarily store updated L-Pages at least until both the lower and upper pages of each MLC are programmed.
  • the non-volatile store may be configured to be sufficiently large as to store both lower and upper pages of a MLC. That is, the non-volatile buffer may be configured to be at least sufficiently large to enable recovery from lower page corruption after a power loss to the data storage device.
  • Such a non-volatile buffer is shown in Fig. 2 at reference numeral 21 1 .
  • the nonvolatile buffer 21 1 may be configured as a circular buffer.
  • the controller 202 may be configured, according to one embodiment, to accumulate data to be written to the non-volatile memory in the non-volatile buffer 21 1 . To do so, the controller 202 may allocate space in the buffer 21 1 and write the accumulated data to the allocated space in the buffer 21 1 . At least a portion of the accumulated data may then be stored in a lower page of the non-volatile memory and in the allocated space in the buffer 21 1 . At some later point in time, the upper page corresponding to the previously-programmed lower page may itself be programmed, whereupon the previously-allocated space in the non-volatile buffer 21 1 may be de-allocated, to make room for further incoming data.
  • the provision of such nonvolatile buffer 21 1 and the temporary storage therein of newly-updated L-pages enables the controller 202 to generate and send a write acknowledgement to a host 218 after the accumulated data is written to the allocated space in the non-volatile buffer 21 1 .
  • This acknowledgement therefore, may be sent sooner than would otherwise be the case were the acknowledgment sent after the data was programmed in the non-volatile memory, while avoiding the use of such backup- power mechanisms such as super capacitors or battery-backed NAND.
  • the non-volatile buffer 21 1 may comprise Magnetic Random Access Memory (MRAM), which operates at speeds comparable to DRAM while being storing data in a non-volatile manner.
  • MRAM Magnetic Random Access Memory
  • non-volatile buffer 21 1 Using such non-volatile buffer 21 1 enables the controller, after the power is restored to the data storage device after a loss of power, to read data from the non-volatile buffer 21 1 and to store at least a portion of the read data in the non-volatile memory. As shown in Fig. 2 at 209, data may be accumulated, written and stored in the non-volatile buffer 21 1 in units of S-Pages.
  • the non-volatile buffer 21 1 enables the controller 202 to read L-pages stored in the non-volatile buffer 21 1 and to rebuild the logical-to-physical address translation map 902 using the logical pages read therefrom.
  • the controller 202 may be configured to read S-Journals in the System Band that correspond to L-pages stored in the Cold Band.
  • the Cold Band may be that portion of the non-volatile memory allocated by the controller 202 that is designated for storage of recently garbage collected L-Pages. It is possible, however, that the Cold Band stores "unsafe" L-page data at lower pages for which the corresponding upper page has not yet been programmed.
  • the controller 202 may be configured to read S-Journals corresponding to the Cold Band no further than entries thereof corresponding to safe L-pages stored in a lower page where the corresponding upper page has been programmed.
  • the controller 202 may be further configured to stop processing S-Journals containing entries corresponding to such unsafe L- Pages and to instead read the unsafe L-Pages from the non-volatile buffer 21 1 and write them to non-volatile memory. This, in turn, generates updates to the logical-to- physical address translation map 902 and corresponding S-Journal entries and/or new S-Journals.
  • This procedure may be carried out for one or more last-written S- Blocks referenced by the S-Journal(s) containing such entries (e.g., entries of S- Journals corresponding to the Cold Band that includes unsafe L-pages stored in a lower page where the corresponding upper page has not been programmed).
  • Such one or more S-Blocks may then be garbage collected.
  • the nonvolatile buffer 21 1 may be used to recover potentially unsafe L-Pages during reconstruction of the logical-to-physical address translation map 902 (stage 1 map rebuild of Fig. 1 1 ) and to properly schedule garbage collection (stage 2 of map rebuild of Fig. 1 1 ).
  • the additional processing related to unsafe pages is performed for other type of data (in lieu of the Cold Band data.
  • one embodiment does not require reading the volatile memory beforehand and thus does not require reading the volatile memory until all entries have been written.
  • the data storage device may be TRIMed (all physical pages of the subjected to the TRIM command or some functional equivalent), to thereby pre-populate the logical-to- physical address translation map 902 with valid entries.
  • 16 GB of random writes to the volatile memory may be carried out to process all of the S- Journals from the non-volatile memory to the now-rebuilt and coherent logical-to- physical address translation map 902.
  • the controller 202 may report that it is ready to service host data access commands.
  • the controller 202 may carry out 6.4 GB of sequential volatile memory reads to update the valid data table 1702 (Fig. 1 3) as part of its free space accounting activities. Thereafter, while or after carrying out data access commands, the data storage device may select the best S-Block(s) to garbage collect. According to one embodiment, the controller 202 may be configured to choose such S-Block(s) for garbage collection based at least in part upon the amount of free space available therein. According to one embodiment, the controller 202 may select that S-Block(s) having the most free space for garbage collection.
  • such free space accounting and block picking may be carried out in hardware and on a drive that is accepting and executing data access commands.
  • Figs. 15-18 are block diagrams illustrating aspects of garbage collection, according to one embodiment.
  • the user S-Block information table 906 (Fig. 10A) may be scanned to select the "best" S-Block to garbage collect.
  • the best S-Block to garbage collect may be that S-Block having the largest amount of free space and the lowest Program Erase (PE) count.
  • PE Program Erase
  • these and/or other criteria may be weighed to select the S-Block to be garbage collected.
  • the S-Block selected to be to garbage collected in Figs. 15-18 is S-Block 1 5, shown at 1502.
  • the user S-Block information table 906 may comprise, among other items of information, a running count of the number of PE cycles undergone by each tracked S-Block, which may be evaluated in deciding which S-Block to garbage collect. As shown at 1502, S-Block 15 has a mix of valid data (hashed blocks) and invalid data (non-hashed blocks).
  • the S-Journal Map (see 904 in Fig. 10B) may be consulted (e.g., indexed into by the S-Block number) to find the location in non-volatile memory of the corresponding S-Journal(s) for that S-Block.
  • the S-Journal pointed to by the S-Journal Map 904 is then located using the header (e.g., 27 LSB) of S-Journal Number, and read into the buffer 1009, as shown in Fig. 15. That is, the E-Page in the System S-Block 1804 pointed to by S-Journal Map 904 (Fig.
  • each physical to logical entry in the S-Journal in the buffer 1009 may then be compared to the corresponding entry in the logical-to-physical address translation map 1802. That is, the address of the L-Page in the S-Journal may be compared with the address of the L-Page in the logical-to-physical address translation map 1802. For each entry in the S-Journal in the buffer 1009, it may be determined whether the address for the L-Page of that entry matches the address of the L-Page in the corresponding entry in the logical-to- physical address translation map 1 802. If the two match, that entry is valid.
  • the S-Journal in System S-Block 1804 is shown as being hashed, indicating that it is now stale.
  • the logical-to-physical address translation map 1802 may then be updated, generating a new E-Page starting address for the valid data read into the buffer 1009. It is to be noted that during the update of the logical-to-physical translation map, the map 1802 may be rechecked for valid entries and may be locked during the map update process to guarantee atomicity.
  • the valid data may then be written out to the Cold S-Block 1801 (the Hot Block being used for recently written host data, not garbage collected data), as shown at Fig. 17.
  • S-Journal for the Cold S-Block 1801 , as shown at 1302 in Fig. 18.
  • S- Journal 1302 may be written out to the System S-Block 1804 in the System Band.
  • S-Block 15 has now been garbage collected and its space may now be added to the free space accounting for S-Block 15 in the data S-Block information 908.
  • System S-Block Information 908 now indicates that the entire S-Block 15 is free space, which may thereafter be erased, its PE count updated and made available for host data. It is to be noted that an invalid S-Journal is still present in S-Block 15. The space in Flash memory in the System Band occupied by this invalid S-Journal may be garbage collected at some later time, erased and re-used.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Techniques For Improving Reliability Of Storages (AREA)
  • Memory System (AREA)
  • Memory System Of A Hierarchy Structure (AREA)

Abstract

A data storage device comprises a non-volatile memory comprising a plurality of blocks, each configured to store a plurality of physical pages at predetermined physical locations. A controller programs and reads data stored in a plurality of logical pages. A volatile memory comprises a logical-to-physical address translation map configured to enabling determination of the physical location, within one or more physical pages, of the data stored in each logical page. A plurality of journals may be stored, each comprising a plurality of entries associating one or more physical pages to each logical page. At startup, the controller may read at least some of the plurality of journals in an order and rebuild the map; indicate a readiness to service data access commands after the map is rebuilt; rebuild a table from the map and, based thereon, select block(s) for garbage collection after having indicated the readiness to process the commands.

Description

METHODS, DEVICES AND SYSTEMS FOR TWO STAGE POWER-ON MAP REBUILD WITH FREE SPACE ACCOUNTING IN A SOLID STATE DRIVE
BACKGROUND
[0001] Due to the nature of flash memory in solid state drives (SSDs), data is typically programmed by pages and erased by blocks. A page in an SSD is typically 8-16 kilobytes (KB) in size and a block consists of a large number of pages (e.g., 256 or 512). Thus, a particular physical location in an SSD (e.g., a page) cannot be directly overwritten without overwriting data in pages within the same block, as is possible in a magnetic hard disk drive. As such, address indirection is needed. Conventional data storage device controllers, which manage the Flash memory on the data storage device and interfaces with the host system, use a Logical-to-Physical (L2P) mapping system known as logical block addressing (LBA) that is part of the Flash translation layer (FTL). When new data comes in replacing older data already written, the data storage device controller causes the new data to be written in a new location (as the data storage device cannot directly overwrite the old data) and update the logical mapping to point to the new physical location. At this juncture, the old physical location no longer holds valid data. As such, the old physical location will eventually need to be erased before it can be written again.
[0002] Conventionally, a large L2P map table maps logical entries to physical address locations on an SSD. This large L2P map table is usually saved in small sections as writes come in. For example, if random writing occurs, although the system may have to update only one entry, it may nonetheless have to save the entire table or a portion thereof, including entries that have not been updated, which is inherently inefficient.
[0003] Fig. 1 shows aspects of a conventional Logical Block Addressing (LBA) scheme for data storage devices. As shown therein, a map table 104 contains one entry for every logical block 102 defined for the data storage device's Flash memory 106. For example, a 64 GB data storage device that supports 512 byte logical blocks may present itself to the host as having 125,000,000 logical blocks. One entry in the map table 104 contains the current location of each of the 125,000 logical blocks in the Flash memory 106. In a conventional data storage device, a Flash page holds an integer number of logical blocks (i.e., a logical block does not span across Flash pages). In this conventional example, an 8 KB Flash page would hold 16 logical blocks (of size 512 bytes). Therefore, each entry in the logical-to-physical map table 1 04 contains a field 108 identifying the die on which the LBA is stored, a field 1 1 0 identifying the flash block on which the LBA is stored, another field 1 12 identifying the flash page within the flash block and a field 1 14 identifying the offset within the flash page that identifies where the LBA data begins in the identified Flash page. The large size of the map table 104 prevents the table from being held inside the SSD controller. Conventionally, the large map table 1 04 is held in an external DRAM connected to the SSD controller. As the map table 104 is stored in volatile DRAM, it must be restored when the SSD powers up, which can take a long time, due to the large size of the table.
[0004] When a logical block is written, the corresponding entry in the map table 104 is updated to reflect the new location of the logical block. When a logical block is read, the corresponding entry in the map table 104 is read to determine the location in Flash memory to be read. A read is then performed to the Flash page specified in the corresponding entry in the map table 104. When the read data is available for the Flash page, the data at the offset specified by the Map Entry is transferred from the Flash device to the host. When a logical block is written, the Flash memory holding the "old" version of the data becomes "garbage" (i.e., data that is no longer valid). It is to be noted that when a logical block is written, the Flash memory will initially contain at least two versions of the logical block; namely, the valid, most recently written version (pointed to by the map table 104) and at least one other, older version thereof that is stale and is no longer pointed to by any entry in the map table 104. These "stale" entries are referred to as garbage, which occupies space that must be accounted for, collected, erased and made available for future use. This process is known as "garbage collection".
[0005] Upon shutdown or other power-interrupting event, the contents of volatile memory in which the map table 104 is stored, are no longer valid. The map table 104, therefore, must be rebuilt. The goals of this process are 1 ) to create a coherent map of the logical to physical relationships and 2) to enable normal operation of the memory system. This normal operation should enable the servicing of data access commands by time-to-ready constraints and should enable the memory system to pick the best block for garbage collection operations. Picking the best block for garbage collection, in turn, involves accurately accounting for the free space within the memory blocks, among other possible factors.
[0006] The conventional method of rebuilding the map table 104 includes first initializing all memory locations within the volatile memory to zero, as RAM powers on in a random state (i.e., without valid ECC). Thereafter, the L2P map may be rebuilt from mapping information saved in the non-volatile memory while, at the same time, carrying out free space accounting. These operations involve lengthy random reads and random writes, which take a significant time, which ultimately delays the drive's time-to-ready for host reads and writes on power up. Conventionally, both the map rebuilding and the free space accounting must be completed before the drive is able to report to the host that it is ready to service data access commands.
BRIEF DESCRIPTION OF THE DRAWINGS
[0007] Fig. 1 shows aspects of a conventional Logical Block Addressing scheme for SSDs.
[0008] Fig. 2 is a block diagram of a data storage device according to one embodiment.
[0009] Fig. 3 is a diagram showing aspects of the physical and logical data organization of a data storage device according to one embodiment.
[0010] Fig. 4 shows a logical-to-physical address translation map and illustrative entries thereof, according to one embodiment.
[0011] Fig. 5 shows aspects of a method for updating a logical-to- physical address translation map and for creating an S-Journal entry, according to one embodiment.
[0012] Fig. 6 is a block diagram of an S-Journal, according to one embodiment.
[0013] Fig. 7 shows an exemplary organization of one entry of an S- Journal, according to one embodiment.
[0014] Fig. 8 is a block diagram of a superblock (S-Block), according to one embodiment.
[0015] Fig. 9 shows another view of a Super page (S-page), according to one embodiment.
[0016] Fig. 10A shows relationships between the logical-to-physical address translation map, S-Journals and S-Blocks, according to one embodiment.
[0017] Fig. 10B is a block diagram of an S-Journal Map, according to one embodiment.
[0018] Fig. 1 1 is a flowchart of a method of carrying out a two-stage power-on logical-to-physical map rebuild, according to one embodiment.
[0019] Fig. 12 is a block diagram illustrating further aspects of a method of carrying out a two-stage power-on logical-to-physical map rebuild, according to one embodiment.
[0020] Fig. 13 is a representation of a valid size information table, according to one embodiment.
[0021] Fig. 14 is a block diagram illustrating aspects of a method of carrying out a two-stage power-on logical-to-physical map rebuild with free space accounting, according to one embodiment.
[0022] Fig. 15 is a block diagram illustrating aspects of garbage collection, according to one embodiment.
[0023] Fig. 1 6 is a block diagram illustrating further aspects of garbage collection, according to one embodiment.
[0024] Fig. 17 is a block diagram illustrating still further aspects of garbage collection, according to one embodiment.
[0025] Fig. 18 is a block diagram illustrating yet further aspects of garbage collection, according to one embodiment.
DETAILED DESCRIPTION
System Overview
[0026] Fig. 2 is a diagram showing aspects of the physical and logical data organization of a data storage device according to one embodiment. In one embodiment, the data storage device is an SSD. In another embodiment, the data storage device is a hybrid drive including Flash memory and rotating magnetic storage media. The disclosure is applicable to both SSD and hybrid implementations, but for the sake of simplicity the various embodiments are described with reference to SSD-based implementations. A data storage device controller 202 according to one embodiment may be configured to be coupled to a host, as shown at reference numeral 21 8. The controller may comprise one or more processors that execute some or all of the functions described below as being performed by the controller. The host 218 may utilize a logical block addressing (LBA) scheme. While the LBA size is normally fixed, the host can vary the size of the LBA dynamically. For example, the LBA size may vary by interface and interface mode. Indeed, while 512 bytes is most common, 4 KB is also becoming more common, as are 512+ (520, 528, etc.) and 4 KB+ (4 KB+8, 4K+16, etc.) formats. As shown therein, the data storage device controller 202 may comprise or be coupled to a page register 204. The page register 204 may be configured to enable the controller 202 to read data from and store data to the data storage device. The controller 202 may be configured to program and read data from an array of flash memory devices responsive to data access commands from the host 218. While the description herein refers to flash memory, it is understood that the array of memory devices may comprise other types of non-volatile memory devices such as flash integrated circuits, Chalcogenide RAM (C-RAM), Phase Change Memory (PC-RAM or PRAM), Programmable Metallization Cell RAM (PMC-RAM or PMCm), Ovonic Unified Memory (OUM), Resistance RAM (RRAM), NAND memory (e.g., single-level cell (SLC) memory, multi-level cell (MLC) memory, or any combination thereof), NOR memory, EEPROM, Ferroelectric Memory (FeRAM), Magnetoresistive RAM (MRAM), other discrete NVM (non-volatile memory) chips, or any combination thereof.
[0027] The page register 204 may be configured to enable the controller 202 to read data from and store data to the array. According to one embodiment, the array of flash memory devices may comprise a plurality of nonvolatile memory devices in die (e.g., 128 dies), each of which comprises a plurality of blocks, such as shown at 206 in Fig. 2. Other page registers 204 (not shown), may be coupled to blocks on other die. A combination of Flash blocks, grouped together, may be called a Superblock or S-Block. In some embodiments, the individual blocks that form an S-Block may be chosen from one or more dies, planes or other levels of granularity. An S-Block, therefore, may comprise a plurality of Flash blocks, spread across one or more die, that are combined together. In this manner, the S-Block may form a unit on which the Flash Management System (FMS) operates. In some embodiments, the individual blocks that form an S-Block may be chosen according to a different granularity than at the die level, such as the case when the memory devices include dies that are sub-divided into structures such as planes (i.e., blocks may be taken from individual planes). According to one embodiment, allocation, erasure and garbage collection may be carried out at the S-Block level. In other embodiments, the FMS may perform data operations according to other logical groupings such as pages, blocks, planes, dies, etc.
[0028] In turn, each of the Flash blocks 206 comprises a plurality of Flash pages (F-Pages) 208. Each F-Page may be of a fixed size such as, for example, 16 KB. The F-Page, according to one embodiment, is the size of the minimum unit of program for a given Flash device. As shown in Fig. 3, each F-Page 208 may be configured to accommodate a plurality of physical pages, hereinafter referred to as E-Pages 210. The term Έ-Page" refers to a data structure stored in Flash memory on which an error correcting code (ECC) has been applied. According to one embodiment, the E-Page 210 may form the basis for physical addressing within the data storage device and may constitute the minimum unit of Flash read data transfer. The E-Page 210, therefore, may be (but need not be) of a predetermined fixed size (such as 2 KB, for example) and determine the size of the payload (e.g., host data) of the ECC system. According to one embodiment, each F- Page 208 may be configured to fit a predetermined plurality of E-Pages 210 within its boundaries. For example, given 16 KB size F-Pages 208 and a fixed size of 2 KB per E-Page 21 0, eight E-Pages 210 fit within a single F-Page 208, as shown in Fig. 3. In any event, according to one embodiment, a power of 2 multiple of E-Pages 210, including ECC, may be configured to fit into an F-Page 208. Each E-Page 21 0 may comprise a data portion 214 and, depending on where the E-Page 210 is located, may also comprise an ECC portion 21 6. Neither the data portion 214 nor the ECC portion 216 need be fixed in size. The address of an E-Page uniquely identifies the location of the E-Page within the Flash memory. For example, the E- Page's address may specify the Flash channel, a particular die within the identified Flash channel, a particular block within the die, a particular F-Page and, finally, the E-Page within the identified F-Page.
[0029] To bridge between physical addressing on the data storage device and logical block addressing by the host, a logical page (L-Page) construct is introduced. An L-Page, denoted in Fig. 3 at reference numeral 212 may comprise the minimum unit of address translation used by the FMS. Each L-Page, according to one embodiment, may be associated with an L-Page number. The L-Page numbers of L-Pages 212, therefore, may be configured to enable the controller 202 to logically reference host data stored in one or more of the physical pages, such as the E-Pages 21 0. The L-Page 212 may also be utilized as the basic unit of compression. According to one embodiment, unlike F-Pages 208 and E-Pages 210, L-Pages 212 are not fixed in size and may vary in size, due to variability in the compression of data to be stored. Since the compressibility of data varies, a 4 KB amount of data of one type may be compressed into a 2 KB L-Page while a 4 KB amount of data of a different type may be compressed into a 1 KB L-Page, for example. Due to such compression, therefore, the size of L-Pages may vary within a range defined by a minimum compressed size of, for example, 24 bytes to a maximum uncompressed size of, for example, 4 KB or 4 KB+. Other sizes and ranges may be implemented. As shown in Fig. 3, L-Pages 212 need not be aligned with the boundaries of E-Page 210. Indeed, L-Pages 212 may be configured to have a starting address that is aligned with an F-Page 208 and/or E-Page 210 boundary, but also may be configured to be unaligned with either of the boundaries of an F- Page 208 or E-Page 210. That is, an L-Page starting address may be located at a non-zero offset from either the start or ending addresses of the F-Pages 208 or the start or ending addresses of the E-Pages 210, as shown in Fig. 3. As the L-Pages 212 are not fixed in size and may be smaller than the fixed-size E-Pages 210, more than one L-Page 212 may fit within a single E-Page 210. Similarly, as the L-Pages 212 may be larger in size than the E-Pages 21 0, L-Pages 212 may span more than one E-Page, and may even cross the boundaries of F-Pages 208, shown in Fig. 3 at numeral 21 7.
[0030] For example, where the LBA size is 51 2 or 512+ bytes, a maximum of, for example, eight sequential LBAs may be packed into a 4 KB L-Page 212, given that an uncompressed L-Page 212 may be 4 KB to 4 KB+. It is to be noted that, according to one embodiment, the exact logical size of an L-Page 212 is unimportant as, after compression, the physical size may span from few bytes at minimum size to thousands of bytes at full size. For example, for 4TB SSD device, 30 bits of addressing may be used to address each L-Page 212 that could potentially be present in such a SSD.
Address Translation Map and Related Data Structures
[0031] Fig. 4 shows a logical-to-physical address translation map and illustrative entries thereof, according to one embodiment. As the host data is referenced by the host in L-Pages 21 2 and as the data storage device stores the L- Pages 212 in one or more contiguous E-Pages 21 0, a logical-to-physical address translation map is required to enable the controller 202 to associate an L-Page number of an L-Page 212 to one or more E-Pages 210. Such a logical-to-physical address translation map is shown in Fig. 4 at 302 and, in one embodiment, is a linear array having one entry per L-Page 212. Such a logical-to-physical address translation map 302 may be stored in a volatile memory, such as a DRAM or SRAM. Fig. 4 also shows the entries in the logical-to-physical address translation map for four different L-Pages 21 2, which L-Pages 212 in Fig. 4 are associated with L-Page numbers denoted as L-Page 1 , L-Page 2, L-Page 3 and L-Page 4. According to one embodiment, each L-Page stored in the data storage device may be pointed to by a single and unique entry in the logical-to-physical address translation map 302. Accordingly, in the example being developed herewith, four entries are shown. As shown at 302, each entry in the map 302 may comprise an L-Page number, which may comprise an identification of the physical page (e.g., E-Page) containing the start address of the L-Page being referenced, the offset of the start address within the physical page (e.g., E-Page) and the length of the L-Page. In addition, a plurality of ECC bits may provide error correction functionality for the map entry. For example, and as shown in Fig. 4, and assuming an E-Page size of 2 KB, L-Page 1 may be referenced in the logical-to-physical address translation map 302 as follows: E-Page 1 003, offset 800, length 1 624, followed by a predetermined number of ECC bits (not shown). That is, in physical address terms, the start of L-Page 1 is within (not aligned with) E-Page 1 003, and is located at an offset from the starting physical location of the E-Page 1003 that is equal to 800 bytes. Compressed L-Page 1 , furthermore, extends 1 ,624 bytes, thereby crossing an E-Page boundary to E-Page 1004. Therefore, E-Pages 1 003 and 1004 each store a portion of the L-Page 212 denoted by L-Page number L-Page 1 . Similarly, the compressed L-Page referenced by L-Page number L-Page 2 is stored entirely within E-Page 1004, and begins at an offset therein of 400 bytes and extends only 696 bytes within E-Page 1 004. The compressed L-Page associated with L-Page number L-Page 3 starts within E-Page 1004 at an offset of 1 ,120 bytes (just 24 bytes away from the boundary of L-Page 2) and extends 4,096 bytes past E-Page 1005 and into E-Page 1006. Therefore, the L- Page associated with L-Page number L-Page 3 spans a portion of E-Page 1004, all of E-Page 1005 and a portion of E-Page 1006. Finally, the L-Page associated with L-Page number L-Page 4 begins within E-Page 1006 at an offset of 1 ,144 bytes, and extends 3, 128 bytes to fully span E-Page 1007, crossing an F-Page boundary into E- Page 1 008 of the next F-Page.
[0032] Collectively, each of these constituent identifier fields (E-Page, offset, length and ECC) making up each entry of the logical-to-physical address translation map 302 may be, for example, 8 bytes in size. That is, for an exemplary 4 TB drive, the address of the E-Page may be 32 bits in size, the offset may be 12 bits (for E-Page data portions up to 4 KB) in size, the length may be 10 bits in size and the ECC field may be provided. Other organizations and bit-widths are possible. Such an 8 byte entry may be created each time an L-Page is written or modified, to enable the controller 202 to keep track of the host data, written in L-Pages, within the Flash storage. This 8-byte entry in the logical-to-physical address translation map may be indexed by an L-Page number or LPN. In other words, according to one embodiment, the L-Page number functions as an index into the logical-to-physical address translation map 302. It is to be noted that, in the case of a 4 KB sector size, the LBA is the same as the LPN. The LPN, therefore, may constitute the address of the entry within the volatile memory. When the controller 202 receives a read command from the host 218, the LPN may be derived from the supplied LBA and used to index into the logical-to-physical address translation map 302 to extract the location of the data to be read in the Flash memory. When the controller 202 receives a write command from the host, the LPN may be constructed from the LBA and the logical-to-physical address translation map 302 may be modified. For example, a new entry therein may be created. Depending upon the size of the volatile memory storing the logical-to-physical address translation map 302, the LPN may be stored in a single entry or broken into, for example, a first entry identifying the E-Page containing the starting address of the L-Page in question (plus ECC bits) and a second entry identifying the offset and length (plus ECC bits). According to one embodiment, therefore, these two entries may together correspond and point to a single L-Page within the Flash memory. In other embodiments, the specific format of the logical-to-physical address translation map entries may be different from the examples shown above.
[0033] As the logical-to-physical address translation map 302 may be stored in a volatile memory, it necessarily must be rebuilt upon startup or any other loss of power to the volatile memory. This, therefore, requires some mechanism and information to be stored in a non-volatile memory that will enable the controller 202 to reconstruct the logical-to-physical address translation map 302 before the controller can "know" where the L-Pages are stored in the non-volatile memory after startup or after a power-fail event. According to one embodiment, such mechanism and information may be embodied in a construct that may be called a System Journal, or S-Journal. According to one embodiment, the controller 202 may be configured to maintain, in the plurality of non-volatile memory devices (e.g., in one or more of the blocks 206 in one or more die, channel or plane), a plurality of S- Journals defining physical-to-logical address correspondences. According to one embodiment, each S-Journal may cover a pre-determined range of physical pages (e.g., E-Pages). According to one embodiment, each S-Journal may comprise a plurality of journal entries, with each entry being configured to associate one or more physical pages, such as E-Pages, to the L-Page number of each L-Page. According to one embodiment, each time the controller 202 restarts or whenever the logical-to- physical address translation map 302 must be rebuilt, the controller 202 reads the S- Journals and, from the information read from the S-Journal entries, rebuilds the logical-to-physical address translation map 302.
[0034] Fig. 5 shows aspects of a method for updating a logical-to- physical address translation map and for creating an S-Journal entry, according to one embodiment. As shown therein, to ensure that the logical-to-physical address translation map 302 is kept up-to-date, whenever an L-Page is written or otherwise updated as shown at block B51 , the logical-to-physical address translation map 302 may be updated as shown at B52. As shown at B53, an S-Journal entry may also be created, storing therein information pointing to the location of the updated L-Page. In this manner, both the logical-to-physical address translation map 302 and the S- Journals are updated when new writes occur (e.g., as the host issues writes to nonvolatile memory, as garbage collection/wear leveling occurs, etc.). Write operations to the non-volatile memory devices to maintain a power-safe copy of address translation data may be configured, therefore, to be triggered by newly created S-Journal entries (which may be just a few bytes in size) instead of re-saving all or a portion of the logical-to-physical address translation map, such that Write Amplification (WA) is reduced. The updating of the S-Journals ensure that the controller 202 can access a newly updated L-Page and that the logical-to-physical address translation map 302 may be reconstructed upon restart or other information- erasing power event affecting the non-volatile memory in which the logical-to- physical address translation map is stored. Moreover, in addition to their utility in rebuilding the logical-to-physical address translation map 302, the S-Journals are useful in enabling effective Garbage Collection (GC). Indeed, the S-Journals may contain the last-in-time update to all L-Page numbers, and may also contain stale entries, entries that do not point to a valid L-Page.
[0035] According to one embodiment, the S-Journal may constitute the main flash management data written to the media. According to one embodiment, S- Journals may contain mapping information for a given S-Block and may contain the Physical-to-Logical (P2L) information for a given S-Block. Fig. 6 is a block diagram showing aspects of an S-Journal, according to one embodiment. As shown therein and according to one embodiment, each S-Journal 602 covers a predetermined physical region of the non-volatile memory such as, for example, 32 E-Pages as shown at 606, which are addressable using 5 bits. Each S-Journal 602 may be identified by an S-Journal Number 604. The S-Journal Number 604 used for storing P2L information for host data may comprise a portion of the address of the first physical page covered by the S-Journal. For example, the S-Journal Number of S- Journal 602 may comprise, for example, the 27 MSbs of the first E-Page covered by this S-Journal 602.
[0036] Fig. 7 shows an exemplary organization of one entry 702 of an S-Journal 602, according to one embodiment. Each entry 702 of the S-Journal 602 may point to the starting address of one L-Page, which is physically addressed in E- Pages. Each entry 702 may comprise, for example, a number (5, for example) of LSbs of the E-Page containing the starting E-Page of the L-Page. The full E-Page address may be obtained by concatenating these 5 LSbs with the 27 MSbs of the S- Journal Number in the header. The entry 702 may then comprise the L-Page number, its offset within the identified E-Page and its size. For example, each entry 702 of a data band S-Journal 602 may comprise the 5 LSbs of the first E-Page covered by this S-Journal entry, 30 bits of L-Page number, 9 bits of E-Page offset and 10 bits of L-Page size, adding up to an overall size of about 7 bytes. Various other internal journal entry formats may be used in other embodiments.
[0037] According to one embodiment, due to the variability in the compression or the host configuration of the data stored in L-Pages, a variable number of L-Pages may be stored in a physical area, such as a physical area equal to 32 E-Pages, as shown at 606 in Fig. 6. As a result of the use of compression and the consequent variability in the sizes of L-Pages, S-Journals 602 may comprise a variable number of entries. For example, according to one embodiment, at maximum compression, an L-Page may be 24 bytes in size and an S-Journal 602 may comprise over 2,500 entries, referencing an equal number of L-Pages, one L- Page per S-Journal entry 702.
[0038] As noted above, S-Journals 602 may be configured to contain mapping information for a given S-Block and may contain the P2L information for a given S-Block. More precisely, according to one embodiment, S-Journals 602 may contain the mapping information for a predetermined range of E-Pages within a given S-Block. Fig. 8 is a block diagram of a superblock (S-Block), according to one embodiment. As shown therein, an S-Block 802 may comprise one Flash block (F- Block) 804 (as also shown at 206 in Fig. 2) per die. An S-Block 802, therefore, may be thought of as a collection of F-Blocks 804, one F-Block per die, that are combined together to form a unit of the Flash Management System. According to one embodiment, allocation, erasure and GC may be managed at the Superblock level. Each F-Block 804, as shown in Fig. 8, may comprise a plurality of Flash pages (F- Page) such as, for example, 256 or 512 F-Pages. An F-Page, according to one embodiment, may be the size of the minimum unit of program for a given non-volatile memory device. Fig. 9 shows a Super Page (S-Page), according to one embodiment. As shown therein, an S-Page 803 may comprise one F-Page per block of an S-Block, meaning that an S-Page 803 spans across an entire S-Block 802.
[0039] Fig. 10A shows relationships between the logical-to-physical address translation map, S-Journals and S-Blocks, according to one embodiment. Reference 902 denotes the logical-to-physical address translation map. According to one embodiment, the logical-to-physical address translation map 902 may be indexed by L-Page number, in that there may be one entry in the logical-to-physical address translation map 902 per L-Page in the logical-to-physical address translation map. The physical address of the start of the L-Page in the Flash memory and the size thereof may be given in the map entry; namely by E-Page address, offset within the E-Page and the size of the L-Page. As noted earlier, the L-Page, depending upon its size, may span one or more E-Pages and may span F-Pages and blocks as well.
[0040] As shown at 904, the volatile memory (e.g., DRAM) may also store an S-Journal map. An entry in the S-Journal map 904 stores information related to where an S-Journal is physically located in the non-volatile memory. For example, the 27 MSbs of the E-Page physical address where the start of the L-Page is stored may constitute the S-Journal Number. The S-Journal map 904 in the volatile memory may also include the address of the S-Journal in non-volatile memory, referenced in system E-Pages. From the S-Journal map 904 in volatile memory, System S-Block Information 908 may be extracted. The System S-Block Information 908 may be indexed by System S-Block (S-Block in the System Band) and may comprise, among other information regarding the S-Block, the size of any free or used space in the System S-Block. Also from the S-Journal map 904, the physical location of the S-Journals in non-volatile memory 910 may be extracted.
[0041] The System Band, according to one embodiment, does not contain L-Page data and may contain all File Management System (FMS) meta-data and information. The System Band may be configured as lower-page only for reliability and power fail simplification. During normal operation, the System Band need not be read except during Garbage Collection. According to one embodiment, the System Band may be provided with significantly higher overprovisioning than the data band for overall WA optimization. Other bands may include the Hot Band, which may contain L-Page data and is frequently updated, and the Cold Band, which is a physical area of memory storing static data retained from the garbage collection process, which may be infrequently updated. According to one embodiment, the System, Hot and Cold Bands may be allocated by controller firmware on an S-Block basis.
[0042] As noted above, each of these S-Journals in non-volatile memory may comprise a collection of S-Journal entries and cover, for example, 32 E-Pages worth of data. These S-Journals in non-volatile memory 910 enable the controller 202 to access the S-Journals entries in non-volatile memory upon startup, enable the controller 202 to rebuild in volatile memory not only the logical-to-physical address translation map 902, but also the S-Journal map 904, the S-Block Information 906, and the System S-Block Information 908.
[0043] The S-Journals in non-volatile memory may also contain all of the stale L-Page information, thereby enabling the controller 202 to GC after the logical-to-physical address translation map 902 and the S-Journal Map 904 in volatile memory are rebuilt. The S-Journals, therefore, may be said to contain a sequential history of all updates, over time, to the logical-to-physical address translation map 902.
[0044] Fig. 10B is a block diagram of another view of an S-Journal Map 904, according to one embodiment. The S-Journal Map 904 may reference a plurality of S-Journal entries for each S-Block. According to one embodiment, the S- Block Number may be the MSb of the S-Journal Number. The size of the S-Journal map 904 may be correlated to the number of S-Blocks times the number of S- Journal entries per S-Block. Indexing into the S-Journal Map 904, therefore, may be carried out by referencing the S-Block Number (the MSb of the S-Journal Number) and the S-Journal entry for that S-Block number. The controller 202 may be further configured to build or rebuild a map of the S-Journals and store the resulting S-Journal Map 904 in volatile memory. For example, upon restart or upon the occurrence of another event in which power fails or after a restart subsequent to error recovery, the controller 202 may read the plurality of S-Journals in a predetermined sequential order, build a map of the S-Journals stored in the nonvolatile memory devices based upon the sequentially read plurality of S-Journals, and store the built S-Journal Map 904 in the volatile memory. In particular, the rebuilt S-Journal Map 904 may be configured to contain the physical location for the most recently-written version of each S-Journal. Indeed, according to one embodiment, in rebuilding the S-Journal Map 904, the physical location of older S- Journals may be overwritten when a newer S-Journal is found. Stated differently, according to one embodiment, the S-Journal Map 904 may be rebuilt by the processor 202 based upon read S-Journals that are determined to be valid. Two-Stage Map Rebuilding
[0045] According to one embodiment, at power-on, the address translation map rebuilding and the free space accounting may be split into two distinct stages. In a first stage, the address translation map may be rebuilt from the S-Journals stored in the System Band of the non-volatile memory. As noted above, the System Band may be that portion of the non-volatile memory that contains the FMS meta-data and information. For reliability and for simplification of post-power power-fail operations, the System Band may be configured for lower page only operation, thereby avoiding any possibility of lower page corruption of MLC Flash memory. Moreover, the System Band may be significantly more over-provisioned than the Data Band (which stores all L-Page data and contains both the Hot and Cold Bands) for overall WA optimization.
[0046] According to one embodiment and as shown at B151 of Fig. 1 1 , to rebuild the logical-to-physical address translation map 902, all of the S-Journals may be read. Reading all of the S-Journals enables the controller 202 to extract, from each entry of each S-Journal, the size and exact physical location of each L- Page associated with each L-Page Number. In one embodiment, after all S-Journals are read and the physical-to-logical information extracted therefrom and is stored in the logical-to-physical address translation map 902 in volatile memory, the controller 202 will have accounted for and mapped all of the L-pages stored in the entire nonvolatile memory. How many S-Journals are read depends at least in part upon the range of physical pages covered by each S-Journal (e.g., 32 E-Pages, according to one embodiment), as well as the size of the non-volatile memory. Moreover and according to one embodiment, the S-Journal construct may be used to store information related to a virtual address range and for storing other file management tables in the volatile memory.
[0047] As the S-Journals, according to one embodiment, may be configured to collectively contain all of the physical-to-logical information, and since many L-pages are likely to have been updated many times, the order in which the S- Journals are read is significant. Indeed, as the S-Journals may be configured to collectively store a complete history of updates to the L-Pages (at least since a last garbage collecting of the S-Block in which the L-Pages are stored), the S-Journals are read in one embodiment in the order in which they were created, as indicated at Block B1 51 . This is graphically shown in Fig. 12, in which a L2P translation map entry 1602 generated from an earlier in time S-Journal entry is sequentially replaced by a corresponding map entry 1602 in a comparatively later-created S-Journal entry for the same L-Page, as the S-Journals are read in order. In this manner, should (as is likely) there have been several updates to a particular L-Page, reading the S- Journals in the order in which they were created ensures that invalid S-Journal entries (i.e., S-Journal entries that do not point to the current location of the L-Page) are read before valid entries (i.e., S-Journal entries that do point to the current location of the L-Page). This in turn ensures that the latest in time S-Journal entry for a particular L-Page is used as a basis of creating a valid map entry 1602 for that L-Page. This sequential reading of the S-Journals in the order in which they were created and ensuring that the last-read S-Journal entry for an L-Page is the most current one, ensure that the rebuilt logical-to-physical address translation map 902 is coherent; that is, accurately maps the logical pages to the correct physical locations within the non-volatile memory. According to one embodiment, as the S-Journals stored in non-volatile memory are read, the S-Journal Map 904 (see Fig. 10B) may also be reconstructed, from the S-Journal Number and address of each read S- Journal. The S-Journal Map 904 may be stored in the volatile memory.
[0048] According to one embodiment, in a first stage of one embodiment of the two stage process shown and described herein, the controller 202 may be configured, at startup, to read each of the plurality of S-Journals in the order in which they were created as shown at B1 51 and to rebuild the logical-to- physical address translation map 902 in the volatile memory from the read plurality of journals. That is, the controller 202 may, from the read S-Journal entries, populate the logical-to-physical address translation map 902 with the physical location information (e.g., a length and starting address within an E-Page in non-volatile memory) of each L-Page, such as also shown at 302 in Fig. 4. In this manner, each L-Page stored in the data storage device may be pointed to by a single and unique entry in the logical-to-physical address translation map 902. According to one embodiment, the controller 202 may carry out a write operation (e.g., an 8 byte write) to the volatile memory for each entry in each of the read S-Journals. It is to be understood, however, that embodiments are not limited to the exemplary 8 byte size shown and described herein and that other implementations are possible. [0049] According to one embodiment, after all S-Journals have been read and as the logical-to-physical address translation map 902, the S-Journal map 904 and any required tables have been rebuilt in volatile memory, the data storage device may report its readiness to process data access commands to the host(s) 218, as shown at B152. Note that the data storage device, at the completion of this first stage, has not yet carried out any free space accounting operations and thus may not yet be ready to optimally carry out garbage collecting operations. Indeed, in an embodiment where optimal block selection for garbage collection depends at least in part on free space within the individual S-Blocks, the controller 202, at the end of this first stage, does not yet know which is the optimal S-Block to garbage collect, as the free space accounting has not yet been carried out. One of the considerations in choosing an S-Block for garbage collection activities may comprise ascertaining the amount of free space available in each S-Block. For example, the controller 202 may select an S-Block having the greatest amount of free space as the next S-Block on which to carry out garbage collection activities. Note that, according to one embodiment, the controller 202 may also identify the next best S- Block to garbage collect, as well as the next best S-Block after that, and so on.
[0050] According to one embodiment, after at least the logical-to- physical address translation map 902 has been rebuilt, and after the data storage device has reported its readiness to process data access commands, the second stage of the present two-stage power on procedure may be carried out. According to one embodiment, after at least the logical-to-physical address translation map 902 has been rebuilt, and after the data storage device has reported its readiness to process data access commands, one or more tables may be rebuilt from the rebuilt logical-to-physical address translation map 902 and, based on such rebuilt table(s), one or more of the plurality of S-Blocks may be selected for garbage collection, as shown at B153 in Fig. 1 1 . To do so, the controller 202 may, among other possible actions, rebuild a free space table containing an amount of free space in each of the plurality of S-Blocks after the logical-to-physical address translation map 902 has been rebuilt. To rebuild such free space table(s), the amount of free space in each of the plurality of blocks may be calculated. Instead of the controller 202 tracking the amount of free space in each S-Block, the controller 202 may, according to one embodiment, track the amount of valid data in each S-Block. Such information for each S-Block may be readily available in the logical-to-physical address translation map 902, which may be configured to contain a record of the size of each L-page in the non-volatile memory. From the amount of valid data in each S-Block and with the knowledge of the size of the S-Blocks, the controller 202 may calculate the amount of free space in each S-Block by subtracting the size of the valid data in each S-Block (obtained from reading the logical-to-physical address translation map 902) from the predetermined size of the S-Blocks. The difference between the amount of valid data stored in an S-Block and its size may be equal to or related to the amount of free space within each S-Block. Therefore, from a single sequential read of the logical-to-physical address translation map 902 or by updating the amount of valid data in each S-Block as the logical-to-physical address translation map 902 is being rebuilt, the amount of free space in each S-Block may be calculated by the controller 202. The accumulated (i.e., summed across entries in the logical-to-physical address translation map 902) amount of valid space in each S-Block may be stored, for example, in a table maintained in the volatile memory. For example, the system block information table 908 (see, for example, Fig. 10A) may be used for that purpose. Alternatively, the controller 202 may create and populate a separate valid size information table with the accumulated size of valid data for each S-Block. As shown in the table 1702 of Fig. 13, in this example, S- Block 1 has a total of 450 MB of valid data stored therein and S-Block 2 has a total of 1 .5 TB of valid data stored therein. As the size of the S-Blocks is known a priori, the amount of free space in each S-Block may be readily calculated. In another embodiment, the amount of free space calculation is performed by directly summing the amount of free space based on the information from the logical-to-physical translation map 902.
[0051] According to one embodiment, the data storage device reports that it is ready to process host commands shortly after having completed the rebuild of the logical-to-physical address translation map 902 (and optionally other housekeeping activities such as populating the S-Journal map 904 and other tables in volatile memory). In such an embodiment the data storage device is configured to carry out free space accounting operations (including, e.g., rebuilding the free space table(s)) while and/or after processing host (i.e., data access) commands. Such incoming host commands may alter the free space accounting of the S-Blocks while the second stage is performed. One embodiment accounts for such changes in the amount of valid data that is present in each S-Block.
[0052] For example, after the rebuilding of the logical-to-physical address translation map 902, the map 902 may be sequentially read, from a beginning to the end thereof. The controller 202 may be configured to track a pointer during the reading of the now-coherent logical-to-physical address translation map 902, as shown at 1902 in Fig. 14. According to one embodiment, in carrying out such free space accounting while processing host commands, the controller 202 may be configured to act differently depending whether the L-Page that is modified as a result of a write command is associated with an L-Page Number that has already been processed for free space accounting by the controller 202 or that is associated with an L-Page Number that has not yet been read by the controller 202 as it sequentially reads the logical-to-physical address translation map 902. According to one embodiment, if the L-Page that is modified as a result of a write command is associated with an L-Page Number that has already been processed for free space by the controller 202, as suggested by the "New Write 1 " at reference numeral 1904 in Fig. 14, the controller 202 may carry out free space accounting in the manner described above. This free space accounting may be necessary in this case, as the controller 202 has already processed the entry corresponding to the now-updated L-Page for free space and has added the length thereof to its running count of valid data for the S-Block where the updated L-Page is stored. It is possible that the updated length is the same as the previous length of the L-page, in which case the running count of the valid data would be unchanged if the L-Page is written to the same S-Block. However, the accumulated size of the data in the S-Block would indeed change if the executed write resulted in a change to the length of the L-page or if written to a different S-Block. To maintain an accurate accounting of the free space in each S-Block, therefore, requires that the controller 202 go back and properly account for the free space made available as a result of the update to the L- Page.
[0053] If, however, the L-Page that is modified as a result of a write command is associated with an L-Page Number that has not yet been processed for free space by the controller 202, one embodiment allows the controller 202 to refrain from carrying out free space accounting activities and continue with its sequential read of the logical-to-physical address translation map 902. Indeed, if an executed write command results in an update to an L-Page whose L-Page Number is yet-to- be-read by the controller 202, such as is the case with "New Write 2" referenced at numeral 1906, no separate processing for free space accounting need be performed, as the controller 202 will correctly accumulate the size of valid data in the table 1702 when its pointer 1902 reaches the L-Page Number of the now newly-updated L- Page.
[0054] According to one embodiment, the blocks 206 of the non-volatile memory may comprise Multi-Level Cells (MLC) that store, for example, 2 bits per cell. For an MLC device, there is at least a lower page and an upper page, which may not be contiguous. The lower page of an MLC is typically programmed before the upper page thereof. Because of the distribution of cell voltages in MLC devices, there is the possibility of corrupting the lower page of an MLC if power is interrupted upon or during the programming the upper page. Indeed, during programming, the charge stored by the MLC is gradually increased, in a step-wise manner, until the desired voltage indicative of the desired logic state is reached. This is the reason that, according to one embodiment, the System Band where the S-Journals are stored may be configured to only use the lower pages in a block for increased reliability.
[0055] To further mitigate against lower page corruption errors, one embodiment utilizes a non-volatile buffer to temporarily store updated L-Pages at least until both the lower and upper pages of each MLC are programmed. As the lower and upper pages are not contiguous, the non-volatile store may be configured to be sufficiently large as to store both lower and upper pages of a MLC. That is, the non-volatile buffer may be configured to be at least sufficiently large to enable recovery from lower page corruption after a power loss to the data storage device. Such a non-volatile buffer is shown in Fig. 2 at reference numeral 21 1 . The nonvolatile buffer 21 1 may be configured as a circular buffer. The controller 202 may be configured, according to one embodiment, to accumulate data to be written to the non-volatile memory in the non-volatile buffer 21 1 . To do so, the controller 202 may allocate space in the buffer 21 1 and write the accumulated data to the allocated space in the buffer 21 1 . At least a portion of the accumulated data may then be stored in a lower page of the non-volatile memory and in the allocated space in the buffer 21 1 . At some later point in time, the upper page corresponding to the previously-programmed lower page may itself be programmed, whereupon the previously-allocated space in the non-volatile buffer 21 1 may be de-allocated, to make room for further incoming data. Advantageously, the provision of such nonvolatile buffer 21 1 and the temporary storage therein of newly-updated L-pages enables the controller 202 to generate and send a write acknowledgement to a host 218 after the accumulated data is written to the allocated space in the non-volatile buffer 21 1 . This acknowledgement, therefore, may be sent sooner than would otherwise be the case were the acknowledgment sent after the data was programmed in the non-volatile memory, while avoiding the use of such backup- power mechanisms such as super capacitors or battery-backed NAND. For example, the non-volatile buffer 21 1 may comprise Magnetic Random Access Memory (MRAM), which operates at speeds comparable to DRAM while being storing data in a non-volatile manner. Using such non-volatile buffer 21 1 enables the controller, after the power is restored to the data storage device after a loss of power, to read data from the non-volatile buffer 21 1 and to store at least a portion of the read data in the non-volatile memory. As shown in Fig. 2 at 209, data may be accumulated, written and stored in the non-volatile buffer 21 1 in units of S-Pages.
[0056] According to one embodiment, the non-volatile buffer 21 1 enables the controller 202 to read L-pages stored in the non-volatile buffer 21 1 and to rebuild the logical-to-physical address translation map 902 using the logical pages read therefrom. During startup, when rebuilding the logical-to-physical address translation map 902, the controller 202 may be configured to read S-Journals in the System Band that correspond to L-pages stored in the Cold Band. The Cold Band may be that portion of the non-volatile memory allocated by the controller 202 that is designated for storage of recently garbage collected L-Pages. It is possible, however, that the Cold Band stores "unsafe" L-page data at lower pages for which the corresponding upper page has not yet been programmed. Such lower-page data could be at-risk of lower page corruption should a write command be executed by the controller, which results in the programming of the corresponding upper page. This lower-page corruption could, therefore, affect the coherency of the logical-to- physical address translation map 902, resulting in an error that could be potentially fatal to the data storage device. [0057] According to one embodiment in which the non-volatile memory comprises MLC, upon rebuilding the logical-to-physical address translation map 902, the controller 202 may be configured to read S-Journals corresponding to the Cold Band no further than entries thereof corresponding to safe L-pages stored in a lower page where the corresponding upper page has been programmed. This ensures that no later-executed write command could result in a lower page corruption of data used to rebuild the logical-to-physical address translation map 902. Also according to one embodiment, for entries of S-Journals corresponding to the Cold Band corresponding to unsafe L-pages stored in a lower page where the corresponding upper page has not been programmed, the controller 202 may be further configured to stop processing S-Journals containing entries corresponding to such unsafe L- Pages and to instead read the unsafe L-Pages from the non-volatile buffer 21 1 and write them to non-volatile memory. This, in turn, generates updates to the logical-to- physical address translation map 902 and corresponding S-Journal entries and/or new S-Journals. This procedure may be carried out for one or more last-written S- Blocks referenced by the S-Journal(s) containing such entries (e.g., entries of S- Journals corresponding to the Cold Band that includes unsafe L-pages stored in a lower page where the corresponding upper page has not been programmed). Such one or more S-Blocks may then be garbage collected. In this manner, the nonvolatile buffer 21 1 may be used to recover potentially unsafe L-Pages during reconstruction of the logical-to-physical address translation map 902 (stage 1 map rebuild of Fig. 1 1 ) and to properly schedule garbage collection (stage 2 of map rebuild of Fig. 1 1 ). In other embodiments, the additional processing related to unsafe pages is performed for other type of data (in lieu of the Cold Band data.
[0058] For example, consider a 3.2 TB data storage device. Conventional data storage devices may be required to read the volatile memory where the map table 104 (Fig. 1 ) is to be stored. This, in turn, requires an initialization procedure, as the ECC in the volatile memory has not been set. This may take on the order of, for example, 2 seconds, assuming volatile memory size on the order of 16 GB. Thereafter, the conventional data storage device may have to perform 16 GB of writes interleaved with 16 GB of random reads to rebuild the map table 104 and to update the free space information. Only thereafter may the conventional data storage device report that it is ready to accept and service host commands. In contrast, to carry out the present two-stage power-up procedure, one embodiment does not require reading the volatile memory beforehand and thus does not require reading the volatile memory until all entries have been written. Indeed, according to one embodiment, prior to leaving the manufacturing facility, the data storage device may be TRIMed (all physical pages of the subjected to the TRIM command or some functional equivalent), to thereby pre-populate the logical-to- physical address translation map 902 with valid entries. Thereafter, 16 GB of random writes to the volatile memory may be carried out to process all of the S- Journals from the non-volatile memory to the now-rebuilt and coherent logical-to- physical address translation map 902. At this early stage and before carrying out any free space accounting activities, the controller 202 may report that it is ready to service host data access commands. Thereafter, after having reported that it is ready to execute data access commands, the controller 202 may carry out 6.4 GB of sequential volatile memory reads to update the valid data table 1702 (Fig. 1 3) as part of its free space accounting activities. Thereafter, while or after carrying out data access commands, the data storage device may select the best S-Block(s) to garbage collect. According to one embodiment, the controller 202 may be configured to choose such S-Block(s) for garbage collection based at least in part upon the amount of free space available therein. According to one embodiment, the controller 202 may select that S-Block(s) having the most free space for garbage collection. Advantageously, according to one embodiment, such free space accounting and block picking may be carried out in hardware and on a drive that is accepting and executing data access commands.
Garbage Collection
[0059] Figs. 15-18 are block diagrams illustrating aspects of garbage collection, according to one embodiment. The user S-Block information table 906 (Fig. 10A) may be scanned to select the "best" S-Block to garbage collect. There are a number of criteria that may be evaluated to select which S-Block to garbage collect. For example, the best S-Block to garbage collect may be that S-Block having the largest amount of free space and the lowest Program Erase (PE) count. Alternatively, these and/or other criteria may be weighed to select the S-Block to be garbage collected. For purposes of example, the S-Block selected to be to garbage collected in Figs. 15-18 is S-Block 1 5, shown at 1502. It is to be noted that the user S-Block information table 906 (Fig. 10A) may comprise, among other items of information, a running count of the number of PE cycles undergone by each tracked S-Block, which may be evaluated in deciding which S-Block to garbage collect. As shown at 1502, S-Block 15 has a mix of valid data (hashed blocks) and invalid data (non-hashed blocks).
[0060] Now that S-Block 15 has been selected for GC, the S-Journal Map (see 904 in Fig. 10B) may be consulted (e.g., indexed into by the S-Block number) to find the location in non-volatile memory of the corresponding S-Journal(s) for that S-Block. The S-Journal pointed to by the S-Journal Map 904 is then located using the header (e.g., 27 LSB) of S-Journal Number, and read into the buffer 1009, as shown in Fig. 15. That is, the E-Page in the System S-Block 1804 pointed to by S-Journal Map 904 (Fig. 10B) may be accessed and the S-Journal stored beginning at that location may be read into the buffer 1009. Thereafter, each physical to logical entry in the S-Journal in the buffer 1009 may then be compared to the corresponding entry in the logical-to-physical address translation map 1802. That is, the address of the L-Page in the S-Journal may be compared with the address of the L-Page in the logical-to-physical address translation map 1802. For each entry in the S-Journal in the buffer 1009, it may be determined whether the address for the L-Page of that entry matches the address of the L-Page in the corresponding entry in the logical-to- physical address translation map 1 802. If the two match, that entry is valid. Conversely, if the address for the L-Page in the S-Journal does not match the entry for that L-Page in the logical-to-physical address translation map 1802, that entry in the S-Journal is not valid. According to one embodiment, as valid entries are found in the S-Journal whose entries are being parsed and compared, they may be written to the buffer 1009, as shown in Fig. 16. After processing each S-Journal for S-Block 15, as also shown in Fig. 16 at reference 1 502, S-Block 15 now contains only invalid data. As the entries in the S-Journal in System S-Block 1804 point to such invalid data, the S-Journal in System S-Block 1804 is shown as being hashed, indicating that it is now stale. The logical-to-physical address translation map 1802 may then be updated, generating a new E-Page starting address for the valid data read into the buffer 1009. It is to be noted that during the update of the logical-to-physical translation map, the map 1802 may be rechecked for valid entries and may be locked during the map update process to guarantee atomicity. The valid data may then be written out to the Cold S-Block 1801 (the Hot Block being used for recently written host data, not garbage collected data), as shown at Fig. 17. In turn, this generates a new S-Journal for the Cold S-Block 1801 , as shown at 1302 in Fig. 18. At some later time (e.g., after a sufficient number of entries have been populated), S- Journal 1302 may be written out to the System S-Block 1804 in the System Band. S-Block 15 has now been garbage collected and its space may now be added to the free space accounting for S-Block 15 in the data S-Block information 908. System S-Block Information 908 now indicates that the entire S-Block 15 is free space, which may thereafter be erased, its PE count updated and made available for host data. It is to be noted that an invalid S-Journal is still present in S-Block 15. The space in Flash memory in the System Band occupied by this invalid S-Journal may be garbage collected at some later time, erased and re-used.
[0061 ] While certain embodiments of the disclosure have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the disclosure. Indeed, the novel methods, devices and systems described herein may be embodied in a variety of other forms. Furthermore, various omissions, substitutions and changes in the form of the methods and systems described herein may be made without departing from the spirit of the disclosure. The accompanying claims and their equivalents are intended to cover such forms or modifications as would fall within the scope and spirit of the disclosure. For example, those skilled in the art will appreciate that in various embodiments, the actual physical and logical structures may differ from those shown in the figures. Depending on the embodiment, certain steps described in the example above may be removed, others may be added. Also, the features and attributes of the specific embodiments disclosed above may be combined in different ways to form additional embodiments, all of which fall within the scope of the present disclosure. Although the present disclosure provides certain preferred embodiments and applications, other embodiments that are apparent to those of ordinary skill in the art, including embodiments which do not provide all of the features and advantages set forth herein, are also within the scope of this disclosure. Accordingly, the scope of the present disclosure is intended to be defined only by reference to the appended claims.

Claims

WHAT IS CLAIMED IS:
1 . A data storage device controller configured to couple to a non-volatile memory comprising a plurality of blocks, each block being configured to store a plurality of physical pages at predetermined physical locations, the controller comprising:
a processor configured to program data to and read data from the non-volatile memory, the data being stored in a plurality of logical pages;
a volatile memory comprising a logical-to-physical address translation map configured to enable the processor to determine a physical location, within one or more physical pages, of the data stored in each logical page; and
a plurality of journals stored in the non-volatile memory, each journal comprising a plurality of journal entries associating one or more physical pages to each logical page,
wherein the controller is configured, upon startup, to:
read each of the plurality of journals in an order and to rebuild the map in the volatile memory from the read plurality of journals;
indicate a readiness to process data access commands after the map is rebuilt; and
rebuild a table from the map and select, based on the rebuilt table, at least one of the plurality of blocks for garbage collection after having indicated the readiness to process data access commands.
2. The data storage device controller of claim 1 , wherein the table is configured to specify free space in each of the plurality of blocks.
3. The data storage device of claim 1 , wherein the controller is configured to process a data access command as the table is being rebuilt.
4. The data storage device of claim 1 , wherein the controller is configured to garbage collect as the table is being rebuilt.
5. The data storage device controller of claim 2, wherein each of the blocks has a predetermined size and wherein the processor is configured to calculate the amount of free space by subtracting a size of valid data in each block obtained from reading the map from the predetermined size.
6. The data storage device controller of claim 5, wherein the processor is further configured to read sequentially through the map to determine, for each block, a size of valid data and to update a valid size information table with the determined size of valid data.
7. The data storage device controller of claim 6, wherein the processor is further configured to carry out a separate free space accounting operation to update a free space for a block storing a logical page corresponding to an already read entry of the map that is updated after having indicated the readiness to service data access commands.
8. The data storage controller of claim 6, wherein the processor is further configured to refrain from carrying out a separate free space accounting operation to update a free space for a block storing a logical page corresponding to a yet-to-be- read entry of the map that is updated after having indicated the readiness to service data access commands.
9. The data storage device controller of claim 1 , wherein the processor is further configured to build a journal map in the volatile memory as each of the plurality of journals is read, the journal map specifying a physical location in the nonvolatile memory of each of the read journals.
10. The data storage device controller of claim 1 , wherein the non-volatile memory comprises Multi-Level Cells (MLC) and wherein the processor is further configured to store each logical page in a non-volatile buffer at least until an upper page corresponding to a lower page of non-volatile memory where the logical page is stored has been programmed.
1 1 . The data storage device controller of claim 10, wherein the processor is further configured to read logical pages stored in the non-volatile buffer and to rebuild the map using the logical pages read from the non-volatile buffer.
12. The data storage device controller of claim 1 , wherein the non-volatile memory comprises Multi-Level Cells (MLC), wherein the non-volatile memory comprises a physical area comprising recently garbage-collected valid logical pages and wherein the processor is further configured to read journals corresponding to the physical area no further than entries thereof corresponding to safe logical pages stored in a lower page of non-volatile memory where the corresponding upper page has been programmed.
13. The data storage device controller of claim 12, wherein, for entries of journals corresponding to the physical area corresponding to unsafe logical pages stored in a lower page where the corresponding upper page has not been programmed, the processor is further configured to read journals associating one or more physical pages of a block from which the valid data was garbage collected.
14. A data storage device, comprising:
the non-volatile memory, and
the data storage device controller of claim 1 .
15. A method of controlling a data storage device comprising a non-volatile memory comprising a plurality of blocks, each block being configured to store a plurality of physical pages at predetermined physical locations, the method comprising:
programming data to and reading data from the non-volatile memory, the data being stored in a plurality of logical pages;
maintaining, in a volatile memory, a logical-to-physical address translation map configured to enable a determination of a physical location, within one or more physical pages, of the data stored in each logical page; and
maintaining, in the non-volatile memory, a plurality of journals, each journal comprising a plurality of journal entries associating one or more physical pages to each logical page and, upon startup: reading each of the plurality of journals in an order and rebuilding the map in the volatile memory from the read plurality of journals;
indicating a readiness to process data access commands after the map is rebuilt; and
rebuilding a table from the map and, based on the rebuilt table, selecting at least one of the plurality of blocks for garbage collection after having indicated the readiness to process data access commands.
16. The method of claim 15, wherein the table is configured to specify an amount of free space in each of the plurality of blocks.
17. The method of claim 15, further comprising processing a data access command as the table is being rebuilt.
18. The method of claim 15, further comprising garbage collecting as the table is being rebuilt.
19. The method of claim 16, wherein each of the blocks has a predetermined size and wherein the method further comprises calculating the amount of free space by subtracting a size of valid data in each block obtained from reading the map from the predetermined size.
20. The method of claim 19, further comprising reading sequentially through the map to determine, for each block, a size of valid data and updating a valid size information table with the determined size of valid data.
21 . The method of claim 20, further comprising carrying out a separate free space accounting operation to update a free space for a block storing a logical page corresponding to an already read entry of the map that is updated after having indicated the readiness to service data access commands.
22. The method of claim 19, further comprising refraining from carrying out a separate free space accounting operation to update a free space for a block storing a logical page corresponding to a yet-to-be-read entry of the map that is updated after having indicated the readiness to service data access commands.
23. The method of claim 15, further comprising building a journal map in the volatile memory as each of the plurality of journals is read, the journal map specifying a physical location in the non-volatile memory of each of the read journals.
24. The method of claim 15, wherein the non-volatile memory comprises Multi-Level Cells (MLC) and wherein the method further comprises storing each logical page in a non-volatile buffer at least until an upper page corresponding to a lower page of non-volatile memory where the logical page is stored has been programmed.
25. The method of claim 24, further comprising reading logical pages stored in the non-volatile buffer and rebuilding the map using the logical pages read from the non-volatile buffer.
26. The method of claim 15, wherein the non-volatile memory comprises Multi-Level Cells (MLC), wherein the non-volatile memory comprises a physical area of memory storing data retained from a garbage collection process and wherein the method further comprises reading journals corresponding to the physical area no further than entries thereof corresponding to safe logical pages stored in a lower page of non-volatile memory where the corresponding upper page has been programmed.
27. The method of claim 26 wherein, for entries of journals corresponding to the physical area corresponding to unsafe logical pages stored in a lower page where the corresponding upper page has not been programmed, the method further comprises reading journals associating one or more physical pages of a block from which the valid data was garbage collected.
PCT/US2014/019681 2013-03-05 2014-02-28 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive WO2014137842A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
AU2014226246A AU2014226246B2 (en) 2013-03-05 2014-02-28 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive
KR1020187031705A KR101965549B1 (en) 2013-03-05 2014-02-28 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive
EP14760216.3A EP2965208B1 (en) 2013-03-05 2014-02-28 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive
CN201480011821.3A CN105009094B (en) 2013-03-05 2014-02-28 The method, apparatus and system of two ranks power-up mapping are rebuild using space free count in solid state drive
KR1020157027180A KR101916206B1 (en) 2013-03-05 2014-02-28 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive
HK16103997.9A HK1216197A1 (en) 2013-03-05 2016-04-07 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/786,352 US9454474B2 (en) 2013-03-05 2013-03-05 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive
US13/786,352 2013-03-05

Publications (1)

Publication Number Publication Date
WO2014137842A1 true WO2014137842A1 (en) 2014-09-12

Family

ID=51489331

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2014/019681 WO2014137842A1 (en) 2013-03-05 2014-02-28 Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive

Country Status (7)

Country Link
US (2) US9454474B2 (en)
EP (1) EP2965208B1 (en)
KR (2) KR101916206B1 (en)
CN (1) CN105009094B (en)
AU (1) AU2014226246B2 (en)
HK (1) HK1216197A1 (en)
WO (1) WO2014137842A1 (en)

Families Citing this family (82)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9400744B2 (en) * 2012-10-30 2016-07-26 Mangstor, Inc. Magnetic random access memory journal for multi-level cell flash memory
US9495288B2 (en) * 2013-01-22 2016-11-15 Seagate Technology Llc Variable-size flash translation layer
EP2951700A4 (en) 2013-01-30 2016-08-31 Hewlett Packard Entpr Dev Lp Non-volatile memory write mechanism
US9823872B2 (en) * 2013-03-15 2017-11-21 Skyera, Llc Apparatus and method for translation from multi-dimensional to linear address space in storage
US9535828B1 (en) 2013-04-29 2017-01-03 Amazon Technologies, Inc. Leveraging non-volatile memory for persisting data
US9213633B2 (en) * 2013-04-30 2015-12-15 Seagate Technology Llc Flash translation layer with lower write amplification
US9170939B1 (en) * 2013-09-26 2015-10-27 Western Digital Technologies, Inc. Early de-allocation of write buffer in an SSD
US9612773B2 (en) * 2013-11-21 2017-04-04 Samsung Electronics Co., Ltd. User device having a host flash translation layer (FTL), a method for transferring an erase count thereof, a method for transferring reprogram information thereof, and a method for transferring a page offset of an open block thereof
US9772781B2 (en) * 2013-11-27 2017-09-26 Ngd Systems, Inc. System and method for supporting atomic writes in a flash translation layer
TWI522804B (en) * 2014-04-23 2016-02-21 威盛電子股份有限公司 Flash memory controller and data storage device and flash memory control method
US9703699B2 (en) * 2014-09-17 2017-07-11 Kabushiki Kaisha Toshiba Hybrid-HDD policy for what host-R/W data goes into NAND
TWI546666B (en) * 2014-11-03 2016-08-21 慧榮科技股份有限公司 Data storage device and flash memory control method
US10482008B2 (en) 2015-01-23 2019-11-19 Hewlett Packard Enterprise Development Lp Aligned variable reclamation
KR20160111222A (en) * 2015-03-16 2016-09-26 에스케이하이닉스 주식회사 Memory system including semiconductor memory device and memory controller, and operating method thereof
US20160291887A1 (en) * 2015-03-30 2016-10-06 Kabushiki Kaisha Toshiba Solid-state drive with non-volatile random access memory
US9880770B2 (en) 2015-09-02 2018-01-30 SK Hynix Inc. Supporting invalidation commands for non-volatile memory
TWI579696B (en) * 2015-11-06 2017-04-21 群聯電子股份有限公司 Method and system for data rebuilding and memory control circuit unit thereof
US10255004B2 (en) * 2015-11-16 2019-04-09 Marvell World Trade Ltd. Systems and methods for managing address-mapping data in memory devices
KR20170081118A (en) * 2015-12-30 2017-07-11 삼성전자주식회사 Storage device including nonvolatile memory device and controller, operating method of storage device, and access method for accessing storage device
US10162561B2 (en) 2016-03-21 2018-12-25 Apple Inc. Managing backup of logical-to-physical translation information to control boot-time and write amplification
US10126962B2 (en) * 2016-04-22 2018-11-13 Microsoft Technology Licensing, Llc Adapted block translation table (BTT)
US9940261B2 (en) * 2016-05-05 2018-04-10 Western Digital Technology, Inc. Zoning of logical to physical data address translation tables with parallelized log list replay
US20180024751A1 (en) * 2016-07-19 2018-01-25 Western Digital Technologies, Inc. Metadata management on a storage device
US10229000B2 (en) * 2016-08-09 2019-03-12 Seagate Llc Erasure codes to prevent lower page corruption in flash memory
KR20180019419A (en) 2016-08-16 2018-02-26 삼성전자주식회사 Memory Controller, Memory System and Operating Method thereof
US10540102B2 (en) * 2016-09-30 2020-01-21 Amazon Technologies, Inc. Physical media aware spacially coupled journaling and replay
US10459644B2 (en) * 2016-10-28 2019-10-29 Western Digital Techologies, Inc. Non-volatile storage system with integrated compute engine and optimized use of local fast memory
CN108121663B (en) * 2016-11-29 2022-05-03 群联电子股份有限公司 Data storage method, memory storage device and memory control circuit unit
US10380028B2 (en) 2016-12-30 2019-08-13 Western Digital Technologies, Inc. Recovery of validity data for a data storage system
US9811269B1 (en) * 2016-12-30 2017-11-07 Intel Corporation Achieving consistent read times in multi-level non-volatile memory
US20180224859A1 (en) * 2017-02-08 2018-08-09 Ford Global Technologies, Llc Tornado Detection Systems And Methods
KR20180128588A (en) * 2017-05-24 2018-12-04 에스케이하이닉스 주식회사 Memory system and method thereof
TWI633437B (en) * 2017-05-26 2018-08-21 慧榮科技股份有限公司 Data storage device and data storage method for detecting currently-used logical pages
US10275361B2 (en) * 2017-05-31 2019-04-30 Seagate Technology Llc Managing multiple namespaces in a non-volatile memory (NVM)
US11221956B2 (en) * 2017-05-31 2022-01-11 Seagate Technology Llc Hybrid storage device with three-level memory mapping
KR102458312B1 (en) * 2017-06-09 2022-10-24 삼성전자주식회사 Storage device and operating method thereof
CN107301073B (en) * 2017-06-23 2021-06-15 郑州云海信息技术有限公司 Configuration information retrieval method and device based on solid state disk system
JP2019050066A (en) * 2017-09-07 2019-03-28 株式会社東芝 Hard disk device and control method
KR102369402B1 (en) * 2017-09-20 2022-03-02 삼성전자주식회사 Storage device, method for operating the same storage system including the storage device
JP2019057172A (en) * 2017-09-21 2019-04-11 東芝メモリ株式会社 Memory system and control method
KR20190043863A (en) * 2017-10-19 2019-04-29 에스케이하이닉스 주식회사 Memory system and operating method thereof
JP6982468B2 (en) * 2017-10-27 2021-12-17 キオクシア株式会社 Memory system and control method
KR102469958B1 (en) * 2017-10-27 2022-11-25 삼성전자주식회사 Nonvolatile memory device configured to be accessed without block address and method of operating the same
KR102418856B1 (en) * 2017-11-02 2022-07-07 삼성전자주식회사 A data storage device
CN109800178B (en) * 2017-11-17 2023-05-16 爱思开海力士有限公司 Garbage collection method and memory system for hybrid address mapping
US10445230B2 (en) * 2017-12-08 2019-10-15 Macronix International Co., Ltd. Managing block arrangement of super blocks
US10552145B2 (en) * 2017-12-12 2020-02-04 Cypress Semiconductor Corporation Memory devices, systems, and methods for updating firmware with single memory device
US10725904B2 (en) * 2017-12-13 2020-07-28 Micron Technology, Inc. Synchronizing NAND logical-to-physical table region tracking
US10949113B2 (en) * 2018-01-10 2021-03-16 SK Hynix Inc. Retention aware block mapping in flash-based solid state drives
KR20190095825A (en) * 2018-02-07 2019-08-16 에스케이하이닉스 주식회사 Data storage device and operating method thereof
TWI714830B (en) * 2018-02-13 2021-01-01 緯穎科技服務股份有限公司 Management method of metadata and memory device using the same
US10592354B2 (en) * 2018-03-20 2020-03-17 Microsoft Technology Licensing, Llc Configurable recovery states
CN108681509B (en) * 2018-04-20 2022-04-08 江苏华存电子科技有限公司 Method for quickly establishing flash memory mapping table
KR20200015247A (en) * 2018-08-03 2020-02-12 에스케이하이닉스 주식회사 Memory system and operating method thereof
US10713163B2 (en) 2018-09-11 2020-07-14 Toshiba Memory Corporation Set aware system data and mapping tables
US10909030B2 (en) * 2018-09-11 2021-02-02 Toshiba Memory Corporation Enhanced trim command support for solid state drives
KR20200057473A (en) * 2018-11-16 2020-05-26 삼성전자주식회사 Storage Device And Storage System Including The Same
CN109634873A (en) * 2018-12-10 2019-04-16 浪潮(北京)电子信息产业有限公司 Solid-state disk host system log information method, apparatus, equipment and medium
US10871907B2 (en) * 2018-12-31 2020-12-22 Micron Technology, Inc. Sequential data optimized sub-regions in storage devices
US11175850B2 (en) * 2019-01-03 2021-11-16 Toshiba Memory Corporation Selective erasure of data in a SSD
US11061598B2 (en) * 2019-03-25 2021-07-13 Western Digital Technologies, Inc. Optimized handling of multiple copies in storage management
US11269552B2 (en) * 2019-06-14 2022-03-08 Micron Technology, Inc. Multi-pass data programming in a memory sub-system having multiple dies and planes
US11144448B2 (en) * 2019-08-05 2021-10-12 Micron Technology, Inc. Memory sub-system for managing flash translation layers table updates in response to unmap commands
KR20210034378A (en) 2019-09-20 2021-03-30 에스케이하이닉스 주식회사 Memory controller and operating method thereof
US11243709B2 (en) * 2019-08-22 2022-02-08 SK Hynix Inc. Data storage apparatus and operating method thereof
KR20210076497A (en) 2019-12-16 2021-06-24 에스케이하이닉스 주식회사 Storage device and operating method thereof
US11734175B2 (en) 2019-08-22 2023-08-22 SK Hynix Inc. Storage device and method of operating the same
US11436101B2 (en) 2019-09-19 2022-09-06 Samsung Electronics Co., Ltd. Rebuild directory/block allocation map from embedded data block information in file system disk
US11762769B2 (en) 2019-09-20 2023-09-19 SK Hynix Inc. Memory controller based on flush operation and method of operating the same
CN110688070B (en) * 2019-09-27 2022-10-11 江苏芯盛智能科技有限公司 Management method and device for solid state disk data table, storage medium and electronic equipment
US11086786B2 (en) * 2019-12-09 2021-08-10 Westem Digital Technologies, Inc. Storage system and method for caching a single mapping entry for a random read command
CN113127376B (en) * 2019-12-30 2024-02-27 阿里巴巴集团控股有限公司 Control method, device and equipment for solid state drive
US11194709B2 (en) * 2019-12-30 2021-12-07 Micron Technology, Inc. Asynchronous power loss recovery for memory devices
CN111581126B (en) * 2020-05-08 2023-08-01 苏州浪潮智能科技有限公司 SSD-based log data storage method, SSD-based log data storage device, SSD-based log data storage equipment and SSD-based log data storage medium
KR20220096329A (en) 2020-12-31 2022-07-07 에스케이하이닉스 주식회사 Memory system and operating method of memory system
CN112817879B (en) * 2021-01-11 2023-04-11 成都佰维存储科技有限公司 Garbage recycling method and device, readable storage medium and electronic equipment
JP2023001494A (en) 2021-06-21 2023-01-06 キオクシア株式会社 Memory system and control method
CN113590503B (en) * 2021-07-23 2024-03-22 合肥康芯威存储技术有限公司 Garbage recycling method and garbage recycling system for non-volatile memory
CN113590502B (en) * 2021-07-23 2024-03-22 合肥康芯威存储技术有限公司 Garbage recycling method and garbage recycling system for non-volatile memory storage equipment
US11615020B2 (en) 2021-08-12 2023-03-28 Micron Technology, Inc. Implementing mapping data structures to minimize sequentially written data accesses
CN114415942B (en) * 2021-12-21 2022-11-04 联芸科技(杭州)股份有限公司 L2P mapping table reconstruction method and solid state disk
US12079517B2 (en) * 2022-07-21 2024-09-03 Micron Technology, Inc. Buffer allocation for reducing block transit penalty

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009110305A1 (en) 2008-03-01 2009-09-11 Kabushiki Kaisha Toshiba Memory system
US20100250837A1 (en) 2008-05-28 2010-09-30 Hyperstone Gmbh Method for Addressing Page-Oriented Non-Volatile Memories
WO2012014140A2 (en) 2010-07-27 2012-02-02 International Business Machines Corporation Logical to physical address mapping in storage systems comprising solid state memory devices
US20120054465A1 (en) * 2010-08-31 2012-03-01 Apple Inc. Systems and methods for trimming logical block addresses corresponding to a data structure residing in non-volatile memory
US20120166759A1 (en) * 2005-06-08 2012-06-28 Micron Technology, Inc. Robust index storage for non-volatile memory
US20120198137A1 (en) * 2009-04-20 2012-08-02 Imation Corp. Logical-to-Physical Address Translation for a Removable Data Storage Device
US20120203958A1 (en) 2011-02-08 2012-08-09 SMART Storage Systems, Inc. Storage control system with change logging mechanism and method of operation thereof

Family Cites Families (90)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100449708B1 (en) * 2001-11-16 2004-09-22 삼성전자주식회사 Flash memory management method
US6856556B1 (en) 2003-04-03 2005-02-15 Siliconsystems, Inc. Storage subsystem with embedded circuit for protecting against anomalies in power signal from host
US20090193184A1 (en) * 2003-12-02 2009-07-30 Super Talent Electronics Inc. Hybrid 2-Level Mapping Tables for Hybrid Block- and Page-Mode Flash-Memory System
US7502256B2 (en) 2004-11-30 2009-03-10 Siliconsystems, Inc. Systems and methods for reducing unauthorized data recovery from solid-state storage devices
US7653778B2 (en) 2006-05-08 2010-01-26 Siliconsystems, Inc. Systems and methods for measuring the useful life of solid-state storage devices
US7765373B1 (en) 2006-06-27 2010-07-27 Siliconsystems, Inc. System for controlling use of a solid-state storage subsystem
US8108692B1 (en) 2006-06-27 2012-01-31 Siliconsystems, Inc. Solid-state storage subsystem security solution
US7509441B1 (en) 2006-06-30 2009-03-24 Siliconsystems, Inc. Systems and methods for segmenting and protecting a storage subsystem
US7447807B1 (en) 2006-06-30 2008-11-04 Siliconsystems, Inc. Systems and methods for storing data in segments of a storage subsystem
KR100855467B1 (en) * 2006-09-27 2008-09-01 삼성전자주식회사 Apparatus and method for mapping of nonvolatile non-volatile memory supporting separated cell type
US8161227B1 (en) 2006-10-30 2012-04-17 Siliconsystems, Inc. Storage subsystem capable of programming field-programmable devices of a target computer system
US8549236B2 (en) 2006-12-15 2013-10-01 Siliconsystems, Inc. Storage subsystem with multiple non-volatile memory arrays to protect against data losses
US7596643B2 (en) 2007-02-07 2009-09-29 Siliconsystems, Inc. Storage subsystem with configurable buffer
US7774525B2 (en) 2007-03-13 2010-08-10 Dell Products L.P. Zoned initialization of a solid state drive
US7685337B2 (en) 2007-05-24 2010-03-23 Siliconsystems, Inc. Solid state storage subsystem for embedded applications
US7685338B2 (en) 2007-05-24 2010-03-23 Siliconsystems, Inc. Solid state storage subsystem for embedded applications
US7685374B2 (en) 2007-07-26 2010-03-23 Siliconsystems, Inc. Multi-interface and multi-bus structured solid-state storage subsystem
US8095851B2 (en) 2007-09-06 2012-01-10 Siliconsystems, Inc. Storage subsystem capable of adjusting ECC settings based on monitored conditions
US8195912B2 (en) * 2007-12-06 2012-06-05 Fusion-io, Inc Apparatus, system, and method for efficient mapping of virtual and physical addresses
US8078918B2 (en) 2008-02-07 2011-12-13 Siliconsystems, Inc. Solid state storage subsystem that maintains and provides access to data reflective of a failure risk
US7962792B2 (en) 2008-02-11 2011-06-14 Siliconsystems, Inc. Interface for enabling a host computer to retrieve device monitor data from a solid state storage subsystem
US7733712B1 (en) 2008-05-20 2010-06-08 Siliconsystems, Inc. Storage subsystem with embedded circuit for protecting against anomalies in power signal from host
US8321652B2 (en) * 2008-08-01 2012-11-27 Infineon Technologies Ag Process and method for logical-to-physical address mapping using a volatile memory device in solid state disks
US8583835B1 (en) 2008-08-06 2013-11-12 Siliconsystems, Inc. Command portal for executing non-standard storage subsystem commands
US8375151B1 (en) 2009-02-12 2013-02-12 Siliconsystems, Inc. Command portal for securely communicating and executing non-standard storage subsystem commands
US9176859B2 (en) 2009-01-07 2015-11-03 Siliconsystems, Inc. Systems and methods for improving the performance of non-volatile memory operations
US8090899B1 (en) 2009-03-04 2012-01-03 Western Digital Technologies, Inc. Solid state drive power safe wear-leveling
US10079048B2 (en) 2009-03-24 2018-09-18 Western Digital Technologies, Inc. Adjusting access of non-volatile semiconductor memory based on access time
US8219776B2 (en) * 2009-09-23 2012-07-10 Lsi Corporation Logical-to-physical address translation for solid state disks
US8243525B1 (en) 2009-09-30 2012-08-14 Western Digital Technologies, Inc. Refreshing non-volatile semiconductor memory by reading without rewriting
US8254172B1 (en) 2009-09-30 2012-08-28 Western Digital Technologies, Inc. Wear leveling non-volatile semiconductor memory based on erase times and program times
US9753847B2 (en) 2009-10-27 2017-09-05 Western Digital Technologies, Inc. Non-volatile semiconductor memory segregating sequential, random, and system data to reduce garbage collection for page based mapping
US8261012B2 (en) 2009-10-30 2012-09-04 Western Digital Technologies, Inc. Non-volatile semiconductor memory comprising power fail circuitry for flushing write data in response to a power fail signal
US8135903B1 (en) 2009-10-30 2012-03-13 Western Digital Technologies, Inc. Non-volatile semiconductor memory compressing data to improve performance
TWI423024B (en) * 2009-11-23 2014-01-11 Phison Electronics Corp Data storing method for a flash memory, and flash memory controller and flash memory storage system using the same
US8397107B1 (en) 2009-12-11 2013-03-12 Western Digital Technologies, Inc. Data storage device employing data path protection using both LBA and PBA
US8443167B1 (en) 2009-12-16 2013-05-14 Western Digital Technologies, Inc. Data storage device employing a run-length mapping table and a single address mapping table
US8316176B1 (en) 2010-02-17 2012-11-20 Western Digital Technologies, Inc. Non-volatile semiconductor memory segregating sequential data during garbage collection to reduce write amplification
US8407449B1 (en) 2010-02-26 2013-03-26 Western Digital Technologies, Inc. Non-volatile semiconductor memory storing an inverse map for rebuilding a translation table
US8725931B1 (en) 2010-03-26 2014-05-13 Western Digital Technologies, Inc. System and method for managing the execution of memory commands in a solid-state memory
US8713066B1 (en) 2010-03-29 2014-04-29 Western Digital Technologies, Inc. Managing wear leveling and garbage collection operations in a solid-state memory using linked lists
US8782327B1 (en) 2010-05-11 2014-07-15 Western Digital Technologies, Inc. System and method for managing execution of internal commands and host commands in a solid-state memory
US9026716B2 (en) * 2010-05-12 2015-05-05 Western Digital Technologies, Inc. System and method for managing garbage collection in solid-state memory
US8886870B2 (en) 2010-05-25 2014-11-11 Marvell World Trade Ltd. Memory access table saving and restoring system and methods
US8341339B1 (en) 2010-06-14 2012-12-25 Western Digital Technologies, Inc. Hybrid drive garbage collecting a non-volatile semiconductor memory by migrating valid data to a disk
US8612669B1 (en) 2010-06-28 2013-12-17 Western Digital Technologies, Inc. System and method for performing data retention in solid-state memory using copy commands and validity and usage data
US8447920B1 (en) 2010-06-29 2013-05-21 Western Digital Technologies, Inc. System and method for managing data access in non-volatile memory
US8521972B1 (en) 2010-06-30 2013-08-27 Western Digital Technologies, Inc. System and method for optimizing garbage collection in data storage
US8639872B1 (en) 2010-08-13 2014-01-28 Western Digital Technologies, Inc. Hybrid drive comprising write cache spanning non-volatile semiconductor memory and disk
US8775720B1 (en) 2010-08-31 2014-07-08 Western Digital Technologies, Inc. Hybrid drive balancing execution times for non-volatile semiconductor memory and disk
US8638602B1 (en) 2010-09-10 2014-01-28 Western Digital Technologies, Inc. Background selection of voltage reference values for performing memory read operations
US8769190B1 (en) 2010-09-15 2014-07-01 Western Digital Technologies, Inc. System and method for reducing contentions in solid-state memory access
US8788779B1 (en) 2010-09-17 2014-07-22 Western Digital Technologies, Inc. Non-volatile storage subsystem with energy-based performance throttling
US8612804B1 (en) 2010-09-30 2013-12-17 Western Digital Technologies, Inc. System and method for improving wear-leveling performance in solid-state memory
US8601313B1 (en) 2010-12-13 2013-12-03 Western Digital Technologies, Inc. System and method for a data reliability scheme in a solid state memory
US8601311B2 (en) 2010-12-14 2013-12-03 Western Digital Technologies, Inc. System and method for using over-provisioned data capacity to maintain a data redundancy scheme in a solid state memory
US8615681B2 (en) 2010-12-14 2013-12-24 Western Digital Technologies, Inc. System and method for maintaining a data redundancy scheme in a solid state memory in the event of a power loss
US8458435B1 (en) 2010-12-20 2013-06-04 Western Digital Technologies, Inc. Sequential write thread detection
US8392635B2 (en) 2010-12-22 2013-03-05 Western Digital Technologies, Inc. Selectively enabling a host transfer interrupt
US8683113B2 (en) 2011-02-04 2014-03-25 Western Digital Technologies, Inc. Concurrently searching multiple devices of a non-volatile semiconductor memory
US8700950B1 (en) 2011-02-11 2014-04-15 Western Digital Technologies, Inc. System and method for data error recovery in a solid state subsystem
US8700951B1 (en) 2011-03-09 2014-04-15 Western Digital Technologies, Inc. System and method for improving a data redundancy scheme in a solid state subsystem with additional metadata
US8769232B2 (en) 2011-04-06 2014-07-01 Western Digital Technologies, Inc. Non-volatile semiconductor memory module enabling out of order host command chunk media access
US8862804B2 (en) 2011-04-29 2014-10-14 Western Digital Technologies, Inc. System and method for improved parity determination within a data redundancy scheme in a solid state memory
US8751728B1 (en) 2011-04-29 2014-06-10 Western Digital Technologies, Inc. Storage system bus transfer optimization
US9021178B2 (en) 2011-05-02 2015-04-28 Western Digital Technologies, Inc. High performance path for command processing
US8503237B1 (en) 2011-05-18 2013-08-06 Western Digital Technologies, Inc. System and method for data recovery in a solid state storage device
US8793429B1 (en) 2011-06-03 2014-07-29 Western Digital Technologies, Inc. Solid-state drive with reduced power up time
US8719531B2 (en) 2011-06-14 2014-05-06 Western Digital Technologies, Inc. System and method for performing data retention that incorporates environmental conditions
US8423722B1 (en) 2011-08-26 2013-04-16 Western Digital Technologies, Inc. System and method for high performance command processing in solid state drives
US8707104B1 (en) 2011-09-06 2014-04-22 Western Digital Technologies, Inc. Systems and methods for error injection in data storage systems
US8713357B1 (en) 2011-09-06 2014-04-29 Western Digital Technologies, Inc. Systems and methods for detailed error reporting in data storage systems
US8700834B2 (en) 2011-09-06 2014-04-15 Western Digital Technologies, Inc. Systems and methods for an enhanced controller architecture in data storage systems
US9690694B2 (en) * 2011-09-27 2017-06-27 Sandisk Technologies, Llc Apparatus, system, and method for an address translation layer
EP2764516A4 (en) * 2011-10-05 2015-07-08 Lsi Corp Self-journaling and hierarchical consistency for non-volatile storage
TWI432962B (en) * 2011-10-06 2014-04-01 Mstar Semiconductor Inc Electronic system and memory managing method thereof
US8977803B2 (en) 2011-11-21 2015-03-10 Western Digital Technologies, Inc. Disk drive data caching using a multi-tiered memory
US8724422B1 (en) 2012-02-29 2014-05-13 Western Digital Technologies, Inc. System and method for charging back-up charge storage element for data storage device using spindle phase switching elements
US9003224B2 (en) 2012-04-25 2015-04-07 Western Digital Technologies, Inc. Managing unreliable memory in data storage systems
US8788778B1 (en) 2012-06-04 2014-07-22 Western Digital Technologies, Inc. Garbage collection based on the inactivity level of stored data
US8966343B2 (en) 2012-08-21 2015-02-24 Western Digital Technologies, Inc. Solid-state drive retention monitor using reference blocks
US8788880B1 (en) 2012-08-22 2014-07-22 Western Digital Technologies, Inc. Efficient retry mechanism for solid-state memory failures
US9268682B2 (en) 2012-10-05 2016-02-23 Skyera, Llc Methods, devices and systems for physical-to-logical mapping in solid state drives
US8972826B2 (en) 2012-10-24 2015-03-03 Western Digital Technologies, Inc. Adaptive error correction codes for data storage systems
US9177638B2 (en) 2012-11-13 2015-11-03 Western Digital Technologies, Inc. Methods and devices for avoiding lower page corruption in data storage devices
US8954694B2 (en) 2012-11-15 2015-02-10 Western Digital Technologies, Inc. Methods, data storage devices and systems for fragmented firmware table rebuild in a solid state drive
US9021339B2 (en) 2012-11-29 2015-04-28 Western Digital Technologies, Inc. Data reliability schemes for data storage systems
US9059736B2 (en) 2012-12-03 2015-06-16 Western Digital Technologies, Inc. Methods, solid state drive controllers and data storage devices having a runtime variable raid protection scheme
US20140223255A1 (en) 2012-12-18 2014-08-07 Western Digital Technologies, Inc. Decoder having early decoding termination detection
US9430376B2 (en) 2012-12-26 2016-08-30 Western Digital Technologies, Inc. Priority-based garbage collection for data storage systems

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120166759A1 (en) * 2005-06-08 2012-06-28 Micron Technology, Inc. Robust index storage for non-volatile memory
WO2009110305A1 (en) 2008-03-01 2009-09-11 Kabushiki Kaisha Toshiba Memory system
US20100250837A1 (en) 2008-05-28 2010-09-30 Hyperstone Gmbh Method for Addressing Page-Oriented Non-Volatile Memories
US20120198137A1 (en) * 2009-04-20 2012-08-02 Imation Corp. Logical-to-Physical Address Translation for a Removable Data Storage Device
WO2012014140A2 (en) 2010-07-27 2012-02-02 International Business Machines Corporation Logical to physical address mapping in storage systems comprising solid state memory devices
US20120054465A1 (en) * 2010-08-31 2012-03-01 Apple Inc. Systems and methods for trimming logical block addresses corresponding to a data structure residing in non-volatile memory
US20120203958A1 (en) 2011-02-08 2012-08-09 SMART Storage Systems, Inc. Storage control system with change logging mechanism and method of operation thereof

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2965208A4

Also Published As

Publication number Publication date
CN105009094A (en) 2015-10-28
HK1216197A1 (en) 2016-10-21
EP2965208A1 (en) 2016-01-13
US9817577B2 (en) 2017-11-14
AU2014226246B2 (en) 2019-04-18
EP2965208A4 (en) 2016-12-14
KR101916206B1 (en) 2018-11-08
US20170010823A1 (en) 2017-01-12
AU2014226246A1 (en) 2015-09-17
EP2965208B1 (en) 2019-08-14
US20140258588A1 (en) 2014-09-11
KR20150128773A (en) 2015-11-18
CN105009094B (en) 2019-04-12
KR101965549B1 (en) 2019-04-03
US9454474B2 (en) 2016-09-27
KR20180121700A (en) 2018-11-07

Similar Documents

Publication Publication Date Title
US9817577B2 (en) Methods, devices and systems for two stage power-on map rebuild with free space accounting in a solid state drive
US10254983B2 (en) Atomic write command support in a solid state drive
US10055345B2 (en) Methods, devices and systems for solid state drive control
US9513831B2 (en) Method and system for atomically writing scattered information in a solid state storage device
EP2920697B1 (en) Methods, data storage devices and systems for fragmented firmware table rebuild in a solid state drive
US9268682B2 (en) Methods, devices and systems for physical-to-logical mapping in solid state drives
US9792067B2 (en) Trim command processing in a solid state drive
US9507523B1 (en) Methods, devices and systems for variable size logical page management in a solid state drive

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14760216

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2014760216

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2014226246

Country of ref document: AU

Date of ref document: 20140228

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20157027180

Country of ref document: KR

Kind code of ref document: A