US20220291996A1 - Systems, methods, and devices for fault resilient storage - Google Patents

Systems, methods, and devices for fault resilient storage Download PDF

Info

Publication number
US20220291996A1
US20220291996A1 US17/827,657 US202217827657A US2022291996A1 US 20220291996 A1 US20220291996 A1 US 20220291996A1 US 202217827657 A US202217827657 A US 202217827657A US 2022291996 A1 US2022291996 A1 US 2022291996A1
Authority
US
United States
Prior art keywords
storage device
storage
mode
fault resilient
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/827,657
Inventor
Yang Seok KI
Sungwook Ryu
Alain Tran
Changho Choi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Priority to US17/827,657 priority Critical patent/US20220291996A1/en
Publication of US20220291996A1 publication Critical patent/US20220291996A1/en
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TRAN, ALAIN, CHOI, CHANGHO, KI, YANG SEOK, RYU, SUNGWOOK
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/0703Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
    • G06F11/0793Remedial or corrective actions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1415Saving, restoring, recovering or retrying at system level
    • G06F11/142Reconfiguring to eliminate the error
    • G06F11/1428Reconfiguring to eliminate the error with loss of hardware functionality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/0703Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
    • G06F11/0706Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment
    • G06F11/0727Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment in a storage system, e.g. in a DASD or network based storage system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/0703Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
    • G06F11/0751Error or fault detection not based on redundancy
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1076Parity data used in redundant arrays of independent storages, e.g. in RAID systems
    • G06F11/1084Degraded mode, e.g. caused by single or multiple storage removals or disk failures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1415Saving, restoring, recovering or retrying at system level
    • G06F11/1417Boot up procedures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/1629Error detection by comparing the output of redundant processing systems
    • G06F11/165Error detection by comparing the output of redundant processing systems with continued operation after detection of the error
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0608Saving storage space on storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • G06F3/0619Improving the reliability of storage systems in relation to data integrity, e.g. data losses, bit errors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0625Power saving in storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0629Configuration or reconfiguration of storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0629Configuration or reconfiguration of storage systems
    • G06F3/0634Configuration or reconfiguration of storage systems by changing the state or mode of one or more devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0655Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
    • G06F3/0661Format or protocol conversion arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • G06F3/0689Disk arrays, e.g. RAID, JBOD
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/85Active fault masking without idle spares

Definitions

  • This disclosure relates generally to storage, and more specifically to systems, methods, and devices for fault resilient storage.
  • a storage device may encounter a fault condition that may affect the ability of the storage device to operate in a storage system.
  • a method of operating a storage device may include determining a fault condition of the storage device, selecting a fault resilient mode based on the fault condition of the storage device, and operating the storage device in the selected fault resilient mode.
  • the selected fault resilient mode may include a power cycle mode.
  • the selected fault resilient mode may include a reformat mode.
  • the selected fault resilient mode may include a reduced capacity read-only mode.
  • the selected fault resilient mode may include a reduced capacity mode.
  • the selected fault resilient mode may include a reduced performance mode.
  • the selected fault resilient mode may include a read-only mode.
  • the selected fault resilient mode may include a partial read-only mode.
  • the selected fault resilient mode may include a temporary read-only mode.
  • the selected fault resilient mode may include a temporary partial read-only mode.
  • the selected fault resilient mode may include a vulnerable mode.
  • the selected fault resilient mode may include a normal mode.
  • the storage device may be configured to perform a command received from a host.
  • the command may include a namespace capacity management command.
  • the namespace capacity management command may include a resize subcommand.
  • the namespace capacity management command may include a zero-size namespace command.
  • a storage device may include a storage medium, and a storage controller, wherein the storage controller is configured to determine a fault condition of the storage device, select a fault resilient mode based on the fault condition of the storage device, and operate the storage device in the selected fault resilient mode.
  • the selected resilient mode may include one of a power cycle mode, a reformat mode, a reduced capacity read-only mode, a reduced capacity mode, a reduced performance mode, a read-only mode, a partial read-only mode, a temporary read-only mode, a temporary partial read-only mode, or a vulnerable mode.
  • the storage device may be to perform a namespace capacity management command received from a host.
  • a system may include a host, and at least one storage device coupled to the host, wherein the storage device is configured to determine a fault condition of the storage device, select a fault resilient mode based on the fault condition of the storage device, operate in the selected fault resilient mode, and report the selected fault resilient mode to the host.
  • a method of operating a storage array may include determining a first fault resilient operating mode of a first fault resilient storage device of the storage array, determining a second fault resilient operating mode of a second fault resilient storage device of the storage array, allocating one or more rescue spaces of one or more additional fault resilient storage devices of the storage array, mapping user data from the first fault resilient storage device to the one or more rescue spaces, and mapping user data from the second fault resilient storage device to the one or more rescue spaces.
  • the method may further include reassigning at least one device identifier (ID) of the one or more additional fault resilient storage devices to a device ID of the first fault resilient storage device.
  • ID device identifier
  • the at least one device ID of the one or more additional fault resilient storage devices may be reassigned based on a current unaffected device ID and a current faulty device ID.
  • the method may further include redirecting one or more inputs and/or outputs (IOs) from the first fault resilient storage device to the one or more additional fault resilient storage devices.
  • the user data may include a strip of data.
  • the strip of data may be redirected to a target storage device of the one or more additional fault resilient storage devices based on a stripe ID of the user data.
  • Mapping user data from the first fault resilient storage device to the one or more rescue spaces may include maintaining a first mapping table.
  • Mapping user data from the second fault resilient storage device to the one or more rescue spaces may include maintaining a second mapping table.
  • the one or more rescue spaces may have a rescue space percentage ratio of a storage device capacity.
  • the rescue space percentage ratio may be greater than or equal to a number of failed storage devices accommodated by the storage array, divided by the total number of storage devices in the storage array.
  • the one or more rescue spaces may be allocated statically.
  • the one or more rescue spaces may be allocated dynamically.
  • a system may include a storage array including a first fault resilient storage device, a second fault resilient storage device, one or more additional fault resilient storage devices, and a volume manger configured to: determine a first fault resilient operating mode of the first fault resilient storage device, determine a second fault resilient operating mode of the second fault resilient storage device, allocate one or more rescue spaces of one or more additional fault resilient storage devices of the storage array, map user data from the first fault resilient storage device to the one or more rescue spaces, and map user data from the second fault resilient storage device to the one or more rescue spaces.
  • the volume manger may be further configured to reassign at least one device identifier (ID) of the one or more additional fault resilient storage devices to a device ID of the first fault resilient storage device.
  • ID device identifier
  • the volume manger may be further configured to redirect one or more inputs and/or outputs (IOs) from the first fault resilient storage device to the one or more additional fault resilient storage devices.
  • the user data may include a strip of data, and the volume manger may be further configured to redirect the strip of data to a target storage device of the one or more additional fault resilient storage devices based on a stripe ID of the user data.
  • the one or more rescue spaces have a rescue space percentage ratio of a storage device capacity, and the rescue space percentage ratio may be based on a number of failed storage devices accommodated by the storage array, divided by a total number of storage devices in the storage array.
  • An apparatus may include a volume manager for a storage array, the volume manager may include logic configured to: determine a first fault resilient operating mode of a first fault resilient storage device of the storage array, determine a second fault resilient operating mode of a second fault resilient storage device of the storage array, allocate one or more rescue spaces of one or more additional fault resilient storage devices of the storage array, map user data from the first fault resilient storage device to the one or more rescue spaces, and map user data from the second fault resilient storage device to the one or more rescue spaces.
  • the user data may include a strip of data, and the strip of data may be redirected to a target storage device of the one or more additional fault resilient storage devices based on a stripe identifier (ID) of the user data.
  • the one or more rescue spaces have a rescue space percentage ratio of a storage device capacity, and the rescue space percentage ratio may be based on a number of failed storage devices accommodated by the storage array, divided by a total number of storage devices in the storage array.
  • a method of operating a storage array may include allocating a first rescue space of a first fault resilient storage device of the storage array, allocating a second rescue space of a second fault resilient storage device of the storage array, determining a fault resilient operating mode of a third fault resilient storage device of the storage array, and mapping user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode.
  • a first block of the user data may be mapped to the first rescue space, and a second block of the user data may be mapped to the second rescue space.
  • the user data may include a strip of data.
  • a first portion of the strip of data may be mapped to the first rescue space, and the first portion of the strip of data may include a number of data blocks based on a size of the strip of data and a size of the data blocks.
  • the number of data blocks may be further based on a total number of storage devices in the storage array.
  • the method may further include reassigning at least one device identifier (ID) of the first fault resilient storage device to a device ID of the third fault resilient storage device.
  • the method may further include redirecting one or more inputs and/or outputs (IOs) from the third fault resilient storage device to the first rescue space and the second rescue space.
  • the first rescue space may have a capacity based on a capacity of the first fault resilient storage device and a total number of storage devices in the storage array.
  • the first rescue space may have a capacity of strips based on a size of the first rescue space and a block size.
  • a system may include a storage array including a first fault resilient storage device, a second fault resilient storage device, a third fault resilient storage device, and a volume manger configured to allocate a first rescue space of the first fault resilient storage device, allocate a second rescue space of the second fault resilient storage device, determine a fault resilient operating mode of the third fault resilient storage device, and map user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode.
  • the volume manager may be further configured to map a first block of the user data to the first rescue space, and map a second block of the user data to the second rescue space.
  • the user data may include a strip of data, and the volume manager may be further configured to map a first portion of the strip of data to the first rescue space.
  • the first portion of the strip of data may include a number of data blocks based on a size of the strip of data and a size of the data blocks.
  • the number of data blocks may be further based on a total number of storage devices in the
  • a method of operating a storage array may include determining a first parameter of a first fault resilient storage device of the storage array, determining a second parameter of a second fault resilient storage device of the storage array, and determining a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter.
  • the method may further include adjusting the first parameter based on the QoS.
  • the first parameter may be adjusted automatically based on monitoring the first parameter.
  • the first parameter may be adjusted automatically based on monitoring the second parameter.
  • the first parameter may be adjusted by configuring a component of the storage array.
  • the first parameter may be adjusted by controlling the operation of a component of the storage array.
  • the first parameter may include one of a number of storage devices in the storage array, a number of data blocks in a strip of user data for the first fault resilient storage device, a write method for redirecting data from the first fault resilient storage device to the second fault resilient storage device, a number of faulty storage devices supported by the storage array, or a storage capacity of the first fault resilient storage device.
  • a system may include a storage array including a first fault resilient storage device, a second fault resilient storage device, and a volume manger configured to determine a first parameter of a first fault resilient storage device, determine a second parameter of a second fault resilient storage device, and determine a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter.
  • the volume manger may be further configured to adjust the first parameter based on the QoS.
  • the volume manger may be further configured to adjust the first parameter automatically based on monitoring the first parameter.
  • the volume manger may be further configured to adjust the first parameter automatically based on monitoring the second parameter.
  • FIG. 1 illustrates an embodiment of a storage system in accordance with example embodiments of the disclosure.
  • FIG. 2A illustrates a table of some possible fault conditions that may be encountered by an embodiment of a fault resilient storage device in accordance with example embodiments of the disclosure.
  • FIG. 2B illustrates a table of some example embodiments of fault resilient modes and associated space types that may be implemented by a storage device in accordance with example embodiments of the disclosure.
  • FIG. 2C illustrates a table of some example embodiments of commands and subcommands that may be implemented by a storage device in accordance with example embodiments of the disclosure.
  • FIG. 2D illustrates a table of commands that a storage device in accordance with example embodiments of the disclosure may implement through an API.
  • FIG. 3A illustrates a flow chart of an embodiment of a method for operating in a fault resilient mode in accordance with example embodiments of the disclosure.
  • FIG. 3B illustrates a flow chart of an embodiment of a method of operating a storage device in accordance with example embodiments of the disclosure.
  • FIG. 4A illustrates a schematic data layout diagram of a RAID-0 system performing a write operation in accordance with example embodiments of the disclosure.
  • FIG. 4B illustrates a schematic data layout diagram of a RAID-0 system performing a read operation in accordance with example embodiments of the disclosure.
  • FIG. 4C illustrates a schematic data layout diagram of a RAID-0 system performing a remapping and write operation in accordance with example embodiments of the disclosure.
  • FIG. 5A illustrates a flowchart for a method for operating a RAID-0 system in accordance with example embodiments of the disclosure.
  • FIG. 5B illustrates a flow chart showing details of a method for operating a RAID-0 storage system in accordance with example embodiments of the disclosure.
  • FIG. 6 illustrates a schematic diagram of an embodiment of a RAID-0 system in accordance with example embodiments of the disclosure.
  • FIG. 7 illustrates a schematic diagram of an embodiment of a RAID-0 system that may implement rescue space management with data block writing in accordance with example embodiments of the disclosure.
  • FIG. 8 illustrates an example embodiment of a system for implementing quality-of service (QoS) management in a storage system in accordance with example embodiments of the disclosure.
  • QoS quality-of service
  • FIG. 9 illustrates an embodiment of a method of operating a storage array in accordance with example embodiments of the disclosure.
  • FIG. 10 illustrates an embodiment of another method of operating a storage array in accordance with example embodiments of the disclosure.
  • FIG. 11 illustrates an embodiment of a further method of operating a storage array in accordance with example embodiments of the disclosure.
  • Some of the principles in accordance with example embodiments of the disclosure relate to storage devices that may continue to operate in one or more fault resilient modes in case of a fault of the storage device.
  • a storage device may continue to operate in a limited manner that may enable a storage system to recover quickly and/or efficiently from the fault of the storage drive.
  • a storage device may implement any number of the following fault resilient (FR) modes:
  • Some embodiments may implement a power cycle mode which may involve self-healing based on power cycling the storage device.
  • Some embodiments may implement a reformat mode which may involve self-healing based on formatting all or a portion of the storage device.
  • Some embodiments may implement a reduced capacity read-only mode in which a first portion of the storage space of the storage device may operate normally, and a second portion may operate as read-only storage space.
  • Some embodiments may implement a reduced capacity mode in which a first portion of the storage space of the storage device may operate normally, and a second portion may not be available for input and/or output (I/O) operations.
  • Some embodiments may implement a reduced performance mode in which one or more aspects of the performance of the storage device may be reduced.
  • Some embodiments may implement a read-only mode in which data may be read from, but not written to, the storage device.
  • Some embodiments may implement a partial read-only mode in which a first portion of the storage space of the storage device may operate as read-only storage space, and a second portion may not be available for normal input and/or output (I/O) operations.
  • I/O input and/or output
  • Some embodiments may implement a temporary read-only mode in which data may be read from, but not written to, the storage space of the storage device, which may be temporarily valid, and may become invalid.
  • Some embodiments may implement a temporary partial read-only mode in which data may be read from, but not written to, a first portion of the storage space of the storage device, which may be temporarily valid, and may become invalid. A second portion may not be available for input and/or output (I/O) operations.
  • I/O input and/or output
  • Some embodiments may implement a vulnerable mode in which the storage device may not be available for input and/or output (I/O) operations.
  • Some embodiments may implement a normal mode in which the storage device may operate normally.
  • a storage device may implement one or more commands which may be used by a host to determine and/or manage one or more features of the storage device.
  • a storage device may implement a namespace capacity management command which may include a resize and/or zero-size subcommand.
  • FIG. 1 illustrates an embodiment of a storage system in accordance with example embodiments of the disclosure.
  • the embodiment illustrated in FIG. 1 may include a host 105 and one or more storage devices 110 . Some or all of the one or more storage devices 110 may be connected directly to the host 105 , and some or all of the one or more storage devices 110 may be connected to the host 105 through a volume manager 115 as shown in FIG. 1 .
  • Each storage device 110 may include a storage controller 120 (or “control circuit”) and a storage media 125 .
  • a storage device 110 may experience an internal fault condition, and the storage device may exhibit various fault resilient behaviors, as discussed in further detail below, to mitigate the system-level impact of the fault condition.
  • the one or more storage devices 110 may be implemented with any type of storage apparatus and associated storage media including solid state drives (SSDs), hard disk drives (HDDs), optical drives, drives based on any type of persistent memory such as cross-gridded nonvolatile memory with bulk resistance change, and/or the like, and/or any combination thereof.
  • SSDs solid state drives
  • HDDs hard disk drives
  • optical drives drives based on any type of persistent memory such as cross-gridded nonvolatile memory with bulk resistance change, and/or the like, and/or any combination thereof.
  • Data in each storage device may be arranged as blocks, key-value structures, and/or the like, and/or any combination thereof.
  • Each storage device 110 may have any form factor such as 3.5 inch, 2.5 inch, 1.8 inch, M.2, MO-297, MO-300, Enterprise and Data Center SSD Form Factor (EDSFF) and/or the like, using any connector configuration such as Serial ATA (SATA), Small Computer System Interface (SCSI), Serial Attached SCSI (SAS), U.2, and/or the like, and using any storage interface and/or protocol such as Peripheral Component Interconnect (PCI), PCI express (PCIe), Nonvolatile Memory Express (NVMe), NVMe-over-Fabrics (NVMe-oF), Ethernet, InfiniBand, Fibre Channel, and/or the like.
  • PCI Peripheral Component Interconnect
  • PCIe PCI express
  • NVMe Nonvolatile Memory Express
  • NVMe-oF NVMe-over-Fabrics
  • Ethernet InfiniBand, Fibre Channel, and/or the like.
  • Any or all of the host 105 , volume manager 115 , storage controller 120 , and/or any other components disclosed herein may be implemented with hardware, software, or any combination thereof, including combinational logic, sequential logic, one or more timers, counters, registers, state machines, complex programmable logic devices (CPLDs), field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), complex instruction set computer (CISC) processors and/or reduced instruction set computer (RISC) processors, and/or the like executing instructions stored in volatile memories such as dynamic random access memory (DRAM) and/or static random access memory (SRAM), nonvolatile memory such as flash memory and/or the like, as well as graphics processing units (GPUs), neural processing units (NPUs), and/or the like.
  • CPLDs complex programmable logic devices
  • FPGAs field programmable gate arrays
  • ASICs application specific integrated circuits
  • CISC complex instruction set computer
  • RISC reduced instruction set computer
  • the storage media 125 may have a retention period (which may depend on the usage history of the storage media 125 , and, as such, may vary within the storage media 125 ); data that has been stored longer than the retention period (i.e., data having an age exceeding the retention period) may become unreliable and may be said to have expired.
  • Data may be stored in the storage media 125 using an error correcting code, which may be, e.g., a block code.
  • an error correcting code which may be, e.g., a block code.
  • a portion, e.g., a block, of the storage media 125 may degrade to a point that the retention period becomes unacceptably short, and the block may be classified as a “bad block”.
  • reserve space referred to as “bad block management reserve space” may be present (e.g., included in each flash memory die or in each flash memory plane), and the controller 120 , or another controller internal to the flash memory die or to the flash memory plane may begin to use a block in the reserve and cease to use the bad block.
  • the operations and/or components described with respect to the embodiment illustrated in FIG. 1 are example operations and/or components. In some embodiments, some operations and/or components may be omitted and/or other operations and/or components may be included. Moreover, in some embodiments, the temporal and/or spatial order of the operations and/or components may be varied. Although some components may be illustrated as individual components, in some embodiments, some components shown separately may be integrated into single components, and/or some components shown as single components may be implemented with multiple components.
  • FIG. 2A illustrates a table of some possible fault conditions that may be encountered by an embodiment of a fault resilient storage device in accordance with example embodiments of the disclosure.
  • Each fault condition (or “fault state”) may be labeled with a case identifier (“Case ID”) in the first column.
  • the second column may indicate an operation status of the storage in the fault state.
  • the third column of the table may indicate, for each case, whether valid user data remain available.
  • the fourth column of the table may indicate whether the storage device 110 may eventually be returned to full functionality, for example, by reformatting the storage media 125 .
  • Such a state, or fault condition may be caused, for example, by a portion of the controller's firmware becoming corrupted (in which case it may be possible for the controller to restart into a safe mode, in which the corrupted instructions may not be executed) or by a failure of a processing circuit in the storage device 110 (e.g., the failure of a processing circuit that manages interactions with the storage media but is not responsible for communications with the host 105 ).
  • a fault condition of this type occurs, the storage device 110 may respond to a read or write command from the host 105 with an error message.
  • Case 2 may include a fault condition (i) in which the storage device 110 may no longer be capable of performing read or write operations and (ii) from which recovery may be possible by cycling the power of the storage device 110 , by reformatting the storage media (e.g., nonvolatile memory (NVM)), by re-loading firmware, and/or the like.
  • a fault condition may be caused, for example, by a program execution error of the controller 120 of the storage device 110 (e.g., a pointer that is out of range as a result of a bit flip in random-access memory (RAM) of the controller 120 , or an instruction that is incorrect, as a result of a bit flip).
  • a program execution error of the controller 120 of the storage device 110 e.g., a pointer that is out of range as a result of a bit flip in random-access memory (RAM) of the controller 120 , or an instruction that is incorrect, as a result of a bit flip.
  • the program execution error has not caused the controller 120 to write incorrect data to the storage media 125 (e.g., if the program execution error occurred since the most recent write to storage media by the controller), then power cycling the storage device may be sufficient to restore the storage device 110 to normal operation. If the program execution error has caused the controller 120 to write erroneous data to the storage media 125 , then reformatting the storage media 125 , and/or re-loading firmware may be sufficient to restore the storage device 110 to normal operation.
  • Case 3 may include a fault condition that may be mitigated by operating the storage device 110 in a read-only mode, and for which reformatting the storage media 125 may not restore full functionality.
  • faults may include (i) a temperature sensor failure, and (ii) a portion of the storage media 125 having transitioned to a read-only mode.
  • the failure may be detected by determining that a temperature sensor reading is out of range (e.g., has exceeded a threshold temperature), and in such a case the risk of overheating of the storage device 110 may be reduced by avoiding write operations, which may dissipate more power than read operations.
  • the transitioning to a read-only mode of a portion of the storage media 125 may occur, for example, for flash memory storage media 125 , if a flash memory plane or die exhausts a bad block management reserve space used for run time bad block management.
  • the storage device 110 may, while attempting to performing a read operation, make an unsuccessful attempt to decode a data item, determine that the block storing the data is a bad block and upon moving the data from the bad block to the bad block management reserve space, determine that the remaining bad block management reserve space is less than a threshold size and therefore insufficient to ensure the reliability of the plane or die. The storage device 110 may then determine that bad block management is no longer being performed, and transition to a read-only mode.
  • data item may refer to any quantity of data being processed in one operation, e.g., the data resulting from decoding a code block may be a data item.
  • Case 4 may include fault condition that may be mitigated by operating the storage device 110 in a write-through mode. For example, if a power supply backup capacitor in the storage device 110 fails, the device may, in response to a write command received from the host, complete the write to the storage media 125 before sending a command completion to the host 105 , so that if power fails before the write to the storage media 125 has been completed, the host is not incorrectly informed that the write was completed successfully. Operating in the write-through mode may result in a reduction of performance (e.g., in terms of throughput and/or latency).
  • Case 5 may include a fault condition that may be mitigated by operating the storage device 110 in a manner that reduces power dissipation.
  • the storage device 110 may operate in a read-only mode as mentioned above, or it may reduce the rate at which operations (e.g., write operations, which may dissipate more power than read operations) may be performed, to reduce power dissipation in the storage device 110 .
  • the storage device 110 may perform a first write to the storage media, then wait, during an interval corresponding to the reduced performance (the waiting resulting in a decrease in the rate at which write operations are performed), and then perform another (e.g., a second) write to the storage media.
  • Case 6 may include a fault condition that may be mitigated by operating the storage device 110 in a read-only mode, and for which reformatting the storage media 125 may restore full functionality.
  • various levels of fault resiliency may be implemented by a storage device 110 in accordance with example embodiments of the disclosure.
  • some embodiments may implement a fully resilient mode, a partially resilient mode, and a vulnerable mode.
  • the storage device 110 may operate with self-healing features, and the storage device 110 may be capable of recovering full functionality (although the user data in the device may be lost) by resetting operations such as power cycling, re-loading firmware, or formatting of the storage media 125 .
  • the storage device 110 may operate with lower performance, reduced capacity, or reduced capability, when a fault condition exists. For example, as mentioned above, if a power supply backup capacitor fails, writes may be completed (e.g., command completions may be sent to the host 105 ) only after data is written to the storage media 125 (i.e., synchronous writes may be performed), slowing the operation of the storage device 110 , and reducing its performance. The user data may be preserved in this circumstance. As another example, storage device 110 may operate with reduced capacity if the reserve space for bad block management run time bad block (RTBB) is exhausted. In this circumstance, the affected dies in the storage device 110 may be excluded from the disk space and the overall disk capacity may be reduced.
  • RTBB bad block management run time bad block
  • the user data on the lost space may be lost. For example, if a set in IO determinism or a zone in a zoned namespace is no longer capable of accepting new data writes, the set or the zone may be excluded from disk space but the remaining disk space may remain available for read and write operations. The user data on the zone or set may be lost.
  • the storage device 110 may operate with reduced capability, for example, if a storage device 110 does not allow write operations, and switches to a read-only mode.
  • the storage device 110 may be capable of operating in two types of read-only mode: a sustainable read-only mode (which may be referred to as a “first read-only mode”), and an unsustainable read-only mode (which may be referred to as a “second read-only mode”).
  • a sustainable read-only mode which may be referred to as a “first read-only mode”
  • an unsustainable read-only mode which may be referred to as a “second read-only mode”.
  • the unsustainable read-only mode may be employed, for example, when it may not be feasible to operate in the sustainable read-only mode, e.g., when there is insufficient unused storage space to set up a rescue space.
  • a storage device having one or more fault resilient features in accordance with example embodiments of the disclosure may extend and/or organize the features so that a host may utilize them systematically, and the device may continue to operate in some capacity despite a fault condition.
  • the storage device is used for a RAID (Redundant Array of Independent (or Inexpensive) Drives) or RAIN (Redundant Array of Independent Nodes), and a node fails, the system may recover the data by copying the data from the accessible space of the storage device without calculating the stripe parity.
  • LBA space types may be implemented by a storage device having fault resiliency features in accordance with example embodiments of the disclosure. These LBA space types may be used, for example, by a storage device such as that illustrated in FIG. 1 .
  • Some examples of LBA space types may include performing space (P), underperforming space (UP), read-only space (RO), volatile read-only space (VRO), and inaccessible space (IA).
  • an LBA space may also refer to any unit of storage space such as a page, a partition, a set, a zone, and/or the like.
  • Performing (P) space may include LBA space containing valid data, which may be capable of being read and written in a normal manner without sacrificing performance. Data in performing space may be valid.
  • Underperforming (UP) space may include LBA space containing valid data, which may be capable of being read and written in a normal manner, but with degraded performance (e.g., degraded write performance).
  • Read-only (RO) space may include LBA space containing valid data, which may be read-only.
  • LBA space containing valid data
  • a storage device may refuse to write data received from a host and/or may respond with error messages to write commands from the host directed to this type of LBA space.
  • the data in read-only space may remain valid for a period of time exceeding the retention period.
  • Volatile read-only (VRO) space may include read-only space, and the storage device may respond with error messages to write commands from a host directed to this type of LBA space.
  • Data in this type of LBA space may be temporarily valid, and may become invalid when it expires, i.e., when the age of the data in its storage media reaches the retention period of the storage media.
  • IA space may include LBA space containing invalid data, which may not be accessible from the host.
  • LBA space types may be used, for example, to implement some embodiments of fault resilient modes.
  • FIG. 2B illustrates a table of some example embodiments of fault resilient modes and associated LBA space types that may be implemented by a storage device in accordance with example embodiments of the disclosure.
  • the Mode column of the table illustrated in FIG. 2B may include a fault resilient mode number and a name which may be used to identify the mode, for example, in an application programming interface (API) through which one or more features of a storage device may be accessed in accordance with example embodiments of the disclosure.
  • the columns labeled as P, UP, RO, VRO, and IA in the table illustrated in FIG. 2B may indicate an amount of performing (P), underperforming (UP), read-only (RO), volatile read-only (VRO), and inaccessible (IA) LBA space, respectively, that may be used in the corresponding mode.
  • the modes illustrated in FIG. 2B may be invoked, for example, by a host through an API.
  • a host may query the storage device using a get feature command as described below.
  • modes identified with an asterisk (*) may provide a host with detailed information about each type of LBA space used by the mode in response to a get feature command.
  • information about the LBA space used by the other commands may be implicit.
  • the power cycle mode Mode 1
  • all memory may be of the performing (P) type.
  • P performing
  • a storage device may implement any number of the following fault resilient modes.
  • a device manufacturer may implement different combinations of these and other fault resilient modes in different products.
  • a power cycle mode may involve self-healing based on power cycling the storage device.
  • a storage device may experience a fault condition based on one or more flipped bits in memory such as SRAM or DRAM.
  • a flipped bit may be caused, for example, by aging, heating, and/or radiation due to an antenna or high elevations above sea level which may interfere with memory cells.
  • a storage device with a fault resilient power cycle mode may have self-healing capabilities such that power cycling the storage device (e.g., removing then reapplying power) may reset the current state and restore the failed SSD to a normal state. In this case, one or more inflight commands in a submission queue may be lost.
  • Whether the user data of the storage device remains valid may depend on implementation details such as the partitioning of the device, the extent to which different circuits of the storage controller are reset, and/or the like.
  • the entire storage space of the storage device in a power cycle mode, may operate normally (e.g., as performing (P) space).
  • a first portion of the storage space (e.g., X percent) of the storage device may operate normally (e.g., as performing (P) space), and a second portion (e.g., (100 ⁇ X) percent) may operate as read-only (RO) storage space.
  • the size of the performance (P) space in the storage device may be reduced, and the storage device may behave like a normal drive with respect to that space, but the read-only (RO) type of space may not be writable.
  • the storage device may provide a list of LBA ranges for the performance (P) and/or read-only (RO) spaces to a host, for example, in response to a get feature command.
  • the LBA range may represent a set. If the storage device supports Zoned Namespaces (ZNS), the LBA range may represent a zone. In some embodiments, the storage device may also provide information about address ranges for sets and/or ZNS in response to a get feature command.
  • ZNS Zoned Namespaces
  • a first portion of the storage space (e.g., X percent) of the storage device may operate normally (e.g., as performing (P) space), and a second portion (e.g., (100 ⁇ X) percent) may be inaccessible (IA).
  • the size of the performance (P) space in the storage device may be reduced, and the storage device may behave like a normal drive with respect to that space, but inaccessible (IA) space may not be available for normal IOs. For example, if an RTBB is exhausted, the problematic die may be excluded from the disk space, and thus, the overall disk capacity may be reduced.
  • the storage device may provide a list of LBA ranges for the performance (P) and/or inaccessible (IA) type of space. If the storage device supports the IO determinism, the LBA range may represent a set. If the storage device supports ZNS, the LBA range may represent a zone. In some embodiments, the storage device may provide information about the LBA ranges, sets, zones, and/or the like, in response to a get feature command.
  • a storage device may perform normal operations, but at reduced throughput and/or latency.
  • a storage device may include one or more back-up capacitors that, in the event of a loss of the main power supply, may provide power to the storage device for a long enough period of time to enable the storage device to complete a write operation. If one or more of these back-up capacitors fail, the storage device may not notify a host that a write operation is complete until after the data is written to the media.
  • the storage device may provide speculative performance information to a host which may enable the host to make decisions on sending write data to the storage device in a manner that may mitigate the system-level impact of the fault condition.
  • Read-only mode 6 the storage device may only allow read operations and may block external write operations. Depending on the implementation details, data in read-only space may remain valid, for example, after the retention period. Read-only mode may operate with 100 percent read-only (RO) space.
  • RO read-only
  • a first portion of the storage space (e.g., X percent) of the storage device may operate as read-only (RO) space, and a second portion (e.g., (100 ⁇ X) percent) may be inaccessible (IA) space.
  • the storage device may only allow read operations and external write operations may be prohibited in the first portion of the storage space.
  • data in the read-only space may still valid, for example, after the retention period.
  • the storage device may provide a list of LBA ranges for the read-only (RO) and/or inaccessible (IA) types of space. If the storage device supports the IO determinism, the LBA range may represent a set. If the storage device supports ZNS, the LBA range may represent a zone. In some embodiments, the storage device may provide information about the LBA ranges, sets, zones, and/or the like, in response to a get feature command.
  • Data may be read from the storage space of the storage device, which may operate with 100 percent VRO space, but external writes may be prohibited. Data in this space may be temporarily valid but may become invalid after the retention period.
  • a temporary partial read-only mode (Mode 9) data may be read from a first portion (e.g., X percent) of the storage space of the storage device, which may operate as VRO space, while external writes may be prohibited.
  • a second portion (e.g., (100 ⁇ X) percent) may be inaccessible (IA) space. Data in the first portion may be temporarily valid but may become invalid after the retention period.
  • the LBA range may represent a set.
  • the LBA range may represent a zone.
  • the storage device may provide information about the LBA ranges, sets, zones, and/or the like, in response to a get feature command.
  • the storage device may not be available for 1/O operations. However, it may continue to receive commands from the host and return errors.
  • the storage device may operate normally.
  • a storage device in accordance with example embodiments of the disclosure may implement one or more commands which may be used, for example, by a host to query the storage device and/or manage one or more features of the storage device.
  • FIG. 2C illustrates a table of some example embodiments of commands and subcommands that may be implemented by a storage device in accordance with example embodiments of the disclosure.
  • the subcommand column of the table illustrated in FIG. 2C may indicate a name which may be used to identify the subcommand, for example, in an API through which the commands, and responses thereto, may be passed.
  • a get feature command which may include a subcommand as shown in the table illustrated in FIG. 2C , may be passed from a host to a storage device, which may return a response thereto.
  • a storage device may respond as follows to a get feature command based on the subcommand.
  • a resiliency type subcommand may return a type of fault resiliency in case of a failure.
  • the storage device may indicate which of the fault resilient modes illustrated in FIG. 2B the device has selected to operate in based on the fault condition it has encountered.
  • a retention period subcommand may return an average retention period of the data without reprogramming the storage media. In some embodiments, this may be the upper-bound of retention time for data in the storage media from the time of the failure.
  • This subcommand may be used, for example, with temporary read-only mode (Mode 8) and/or temporary partial read-only mode (Mode 9).
  • An earliest expiry subcommand may return a maximum time remaining for data integrity. In some embodiments, this may be the lower-bound of retention time for data in the storage media from the time of the failure. The unit of time may be determined, for example, based on a patrol period. This subcommand may be used, for example, with temporary read-only mode (Mode 8) and/or temporary partial read-only mode (Mode 9).
  • An IOPS subcommand may return a percentage of the maximum available IOPS the storage device may be able to handle based on the fault condition. This subcommand may be used, for example, with reduced performance mode (Mode 5).
  • a bandwidth subcommand may return a percentage of the maximum available bandwidth the storage device may be able to handle based on the fault condition. This subcommand may be used, for example, with reduced performance mode (Mode 5).
  • a space subcommand may return an amount of storage space that may be available in the storage device based on the fault condition. This subcommand may be used, for example, with reduced capacity read-only mode (Mode 3) and/or reduced capacity mode (Mode 4).
  • a namespace capacity management command which may include a subcommand as shown in the table illustrated in FIG. 2C , may be passed from a host to a storage device, which may respond by performing the action indicated by the subcommand.
  • a storage device may respond as follows to a namespace capacity management (NCM) command based on the subcommand.
  • NCM namespace capacity management
  • a namespace may be implemented as a quantity of non-volatile memory (NVM) that may be formatted into logical blocks.
  • a resize command may cause the storage device to resize a namespace based on one or more parameters that may be included with the command.
  • this subcommand may apply to storage device that may support two or more namespaces.
  • the namespaces may support NVMe resizing.
  • a zero-sized namespace command (FR_NAMESPACE_ZERO_SIZE) may cause the storage device to reduce the size of a rescue space to zero.
  • a storage device in accordance with example embodiments of the disclosure may implement an API to enable a host to query the storage device and/or manage one or more features of the storage device.
  • FIG. 2D illustrates a table of commands that a storage device in accordance with example embodiments of the disclosure may implement through an API.
  • Some embodiments may include a hierarchy of enumerated constants, within the category of fault resilient features, that the storage device may employ to respond. As illustrated in FIG. 2D , the hierarchy may include a first level, including a fully resilient status, a partially resilient status, and a vulnerable status. Sub-statuses, and sub-sub-statuses may also be defined. For example, as illustrated in FIG.
  • the partially resilient status may include a first sub-status indicating a loss of capability, and the first sub-status may include a first sub-sub-status, indicating operation in the sustainable read-only mode, and a second sub-sub-status, indicating operation in the unsustainable read-only mode.
  • an API may be implemented, for example, using NVMe commands.
  • a feature command (FAULT_RESILIENT_FEATURE) may return the fault resilient classes and features in each class that the storage device may support.
  • a status command may return the status of the storage device after a fault resilient recovery is performed.
  • a volatile blocks command (FAULT_RESILIENT_VOLATILE_BLOCKS (H)) may return a list of LBA ranges that reach to the retention period in the next H hours. In some embodiments, this may be used to determine the blocks that need to be relocated for unsustainable read-only.
  • An invalid data blocks command (FAULT_RESILIENT_INVALID_DATA_BLOCKS) may return a list of LBA ranges that may become invalid after switching to a fault resilient mode.
  • FIG. 3A illustrates a flow chart of an embodiment of a method for operating in a fault resilient mode in accordance with example embodiments of the disclosure.
  • the method illustrated in FIG. 3A may be implemented, for example, by the systems and/or components illustrated in FIG. 1 .
  • the method may begin at operation 300 .
  • the host 105 may send or receives data from the storage device 110 .
  • the host 105 may determine whether an error has occurred in the storage device 110
  • the storage device 110 may perform an internal diagnosis and determine its fault resilient status (e.g., fully resilient, partially resilient, or vulnerable).
  • the storage device 110 may modify its performance, capacity, and/or capability (e.g., transitioning to a read-only mode) based on the diagnosis.
  • the storage device 110 may post the status upon request from the host 105 based on an application programming interface (API).
  • API application programming interface
  • the host 105 may route data of a given type to the storage device 110 or to a different storage 110 device at a given bandwidth based on the status. The method may end at operation 335 .
  • FIG. 3B illustrates a flow chart of an embodiment of a method of operating a storage device in accordance with example embodiments of the disclosure.
  • the method may begin at operation 350 .
  • the method may determine a fault condition of the storage device.
  • the method may select a fault resilient mode based on the fault condition of the storage device.
  • the method may operate the storage device in the selected fault resilient mode.
  • the method may end at operation 370 .
  • the operations and/or components described with respect to the embodiment illustrated in FIGS. 3A and 3B , as well as all of the other embodiments described herein, are example operations and/or components. In some embodiments, some operations and/or components may be omitted and/or other operations and/or components may be included. Moreover, in some embodiments, the temporal and/or spatial order of the operations and/or components may be varied. Although some components may be illustrated as individual components, in some embodiments, some components shown separately may be integrated into single components, and/or some components shown as single components may be implemented with multiple components.
  • a storage controller such as a field programmable gate array (FPGA) or embedded processor may perform internal block checks and send asynchronous updates to the host 105 on the status of the storage device 110 . Events may occur and be transmitted to the host 105 (e.g., temperature, or other parameters internal to the device). The host 105 may poll the storage devices 110 on a predetermined schedule, for example, if there is no device driver feature for providing notification.
  • a storage controller may monitor the historical performance of the storage device 110 and use machine learning to provide predictive analytics (e.g., a likelihood of the storage device being in a given fault resilient state). Commands (e.g., NVMe commands) may be implemented and/or expanded, for example, to report the state of the storage device 110 ).
  • the host may: (i) send different data types (e.g., file types such as image, video, text, or high-priority or low-priority data), based on the status of the storage device 110 (for instance, high priority data or real-time data may not be written to a device that is considered in the partially vulnerable mode); (ii) reduce the transmission rate if the storage device 110 is in a partially vulnerable state and in a lower performance state; (iii) send a reduced total amount of data if the storage device 110 is in a partially vulnerable and lower capacity state; (iv) read data at the greatest rate possible, and/or store the data elsewhere, if the storage device 110 is in a partially vulnerable unsustainable read-only mode, so as to avoid exceeding the retention period (in such a circumstance, the host may calculate the needed data rate based on the amount of data to be copied and on the retention period); (v) ignore data “read” from a vulnerable storage device 110 since it may be erroneous, and delete the data as it is received
  • a storage controller on a partially vulnerable storage device that has had a capacity decrease may filter incoming data writes and only write a portion of that data to the storage device 110 .
  • the filtering may include compression.
  • Such a storage controller may receive various types of data (e.g., file types such as image, video, text, or high-priority or low-priority data) from a host 105 and filter based on the status of the storage device 110 . For instance, the storage controller may determine that high priority data may not be written to a storage device 110 that is in the partially vulnerable mode. The storage controller may send a rejection message to the host 105 and give a reason for the rejection.
  • the RAID-0 system may transition to a second operating mode (which may also be referred to as an emergency mode), and the volume manager 115 for the array of storage devices 110 may (i) allocate a rescue space on each of the remaining, unaffected storage devices 110 B (e.g., those that remain in a read-write state) for metadata and rescued user data from faulty storage device 110 A, and/or (ii) create and/or maintain a mapping table (which may also be referred to as an emergency mapping table).
  • Rescue space may be pre-allocated statically prior to system operation, dynamically during operation, or in any combination thereof.
  • the volume manager 115 may be implemented as an independent component, or may be partially or fully integrated into the host, a RAID controller of the RAID-0 system (which may, for example, be housed in a separate enclosure from the host), or in any other configuration.
  • the volume manager 115 may be implemented, for example, with an FPGA.
  • the RAID-0 system may be self-contained and may virtualize the array of storage devices 110 so that from the perspective of the host the RAID-0 system may appear as a single storage device.
  • the volume manager may be implemented as a processing circuit (discussed in further detail below) configured (e.g., by suitable software or firmware) to perform the operations described herein as being performed by the volume manager.
  • the volume manager 115 may check the emergency mapping table to determine whether the stripe is registered e.g., whether an entry has already been made for the stripe. If no entry has been made yet (e.g., the stripe is not registered, which may also be referred to as open-mapped), the volume manager 115 may create an entry in the emergency mapping table to indicate where a strip, that ordinarily would have been written to the faulty storage device 110 A (the storage device that has transitioned to read-only mode), is to be written.
  • the emergency mapping table already contains an entry for the stripe, then the entry may be used to determine where to write the strip that ordinarily would have been written to the faulty storage device 110 A. In either case, the volume manager 115 may then write each strip, as illustrated in FIG. 4A , to the array of storage devices 110 , writing the strip 405 that ordinarily would have been written to the faulty (e.g., read-only) storage device 110 A to rescue space in another storage device 110 B.
  • the faulty (e.g., read-only) storage device 110 A to rescue space in another storage device 110 B.
  • the volume manager 115 may check the emergency mapping table to determine whether an entry has been made for the stripe. If no entry has been made, then, as illustrated in FIG. 4B , the volume manager 115 may read the stripe as it would have, in ordinary operation, reading a strip from each of the storage devices 110 , including the faulty storage device 110 A. If the emergency mapping table contains an entry for the stripe, then the entry may be used to determine where to read the strip that ordinarily would have been read from the faulty storage device 110 A.
  • Each storage device 110 of the array of storage devices 110 may have a drive identification number (or “drive ID”), which may be a number between zero and M ⁇ 1, where M may be the number of storage devices 110 in the array of storage devices 110 .
  • the volume manager 115 may reassign the drive identification numbers, e.g., assign to each unaffected storage device 110 B of the array of storage devices 110 an alternate drive identification number to be used for performing read or write operations for registered stripes (read operations for unregistered stripes may continue to use the original drive identification numbers).
  • the following formula (Formula A) may be used to generate the alternate drive identification numbers:
  • the effect of Formula A may be (i) to assign, to each storage device having an identification number less than the original drive identification number of the faulty storage device, the respective original drive identification number, and/or (ii) to assign, to each storage device having an identification number greater than the original drive identification number of the faulty storage device, the respective original drive identification number minus one.
  • FIG. 4C is a schematic diagram of an embodiment of a RAID-0 system in accordance with example embodiments of the disclosure.
  • Drive 0 may remain mapped to new Drive ID 0 (e.g., 3 mod 3)
  • Drive 2 may be mapped to new Drive ID 1 (e.g., 4 mod 3)
  • Drive 3 may be mapped to new Drive ID 2 (e.g., 5 mod 3).
  • the emergency mapping table may contain an entry for each strip having the format (Stripe ID, R-Strip ID) in which the first element may be the Stripe ID, and the second element may be the R-strip ID on the target drive.
  • an entry of (1,0) in the emergency mapping table may indicate that Strip (1,1) is mapped to R-Strip (1,0) as shown in FIG. 4C .
  • FIG. 5A illustrates a flowchart for a method for operating a RAID-0 system in accordance with example embodiments of the disclosure.
  • a storage device 110 in a RAID-0 system has a fault and transitions to a read-only mode; at 510 , the affected storage device 110 performs an internal diagnosis and determines that its fault resilient status is partially resilient and read-only; at 515 , the volume manager 115 determines that the affected storage device 110 is in a read-only mode and reassigns the IDs of (“live”) unaffected storage devices; at 520 , the volume manager 115 receives a write operation, adds an entry to an emergency mapping table to indicate that the strip of the affected device is redirected to a target (unaffected) storage device 110 , and the entire strip is written to a rescue space of the target (unaffected) storage device based on the new drive IDs of the unaffected storage devices; and, at 525 , the volume manager 115 receives a read command from the host
  • FIG. 5B illustrates a flow chart showing details of a method for operating a RAID-0 storage system in accordance with example embodiments of the disclosure.
  • the method includes, at 530 , determining that the first storage device is in a read-only state and that the second storage device is in a read-write state; at 535 , performing a write operation, of a first stripe, to the storage system, by writing a portion of the first stripe to the second storage device, and making an entry in a mapping table for the first stripe; at 540 , performing a first read operation, of a second stripe, from the storage system, by reading a portion of the second stripe from the first storage device and the second storage device; and at 545 , performing a second read operation, of the first stripe, from the storage system, by determining that the mapping table includes an entry for the first stripe, and reading a portion of the first stripe from the second storage device.
  • a RAID-0 system may be constructed to accommodate the failure of multiple (e.g., N) fault resilient storage devices 110 .
  • An example embodiment of such a system may in some ways be constructed and operate in a manner similar to the embodiment described above with respect to FIGS. 4A-4C , but the size of the rescue space R for each storage device may be determined by considering the number N of faulty storage devices 110 A the system may accommodate.
  • a rescue space percentage ratio e.g., b percent, which may also be referred to as a reservation ratio
  • R (b/100)*C.
  • the drive IDs of the unaffected storage devices 110 B may be reassigned according to the following formula (Formula B):
  • new drive ID ((current drive ID ⁇ 1)+( M′ ⁇ 1))mod( M′ ⁇ 1)
  • FIG. 6 illustrates a schematic diagram of an embodiment of a RAID-0 system in accordance with example embodiments of the disclosure.
  • Drive 0 may remain mapped to new Drive ID 0 (e.g., 1 mod 1), and Drive 3 may be mapped to new Drive ID 1 (e.g., 3 mod 1).
  • Target Drive ID sid % (M′ ⁇ 1)
  • Stripe ID 1
  • the RAID-0 system may transition to an emergency mode in which the volume manager 115 may (i) allocate a rescue space on each of the remaining, unaffected storage devices 110 B (if adequate rescue space has not been allocated already, or if insufficient space has been allocated) for metadata and rescued user data from a faulty storage device 110 A, and/or (ii) create and/or maintain a first mapping table for the first faulty storage device 110 A.
  • the RAID-0 system may then operate in a manner similar to the single device failure embodiment described above.
  • the RAID-0 system may once again allocate a rescue space on each of the remaining, unaffected storage devices 110 B (if adequate rescue space has not been allocated already, or if insufficient space has been allocated) for metadata and rescued user data from a faulty storage device 110 A.
  • the RAID-0 system may then create and/or maintain a second mapping table for the second faulty storage device 110 A.
  • a RAID-0 system may create and/or modify a single mapping table to map data stripes and/or strips of all of the faulty storage devices 110 A to the unaffected storage devices 110 B.
  • one or more mapping tables may be stored in a reserved rescue space, for example, before a Disk Data Format (DDF) structure for a RAID configuration.
  • DDF Disk Data Format
  • the RAID-0 system may then reassign drive IDs of the unaffected storage devices 110 B, for example, based on Formula B, and proceed to operate with the two faulty storage devices 110 A operating in read-only mode.
  • the volume manager 115 may check the one or more emergency mapping tables to determine whether an entry has been made for the stripe. If no entry has been made yet (e.g., the stripe is not registered) the volume manager 115 may create an entry in the one or more emergency mapping tables to indicate where the strips that ordinarily would have been written to the faulty storage devices 110 A (the storage devices that have transitioned to read-only mode), are to be written. If the one or more emergency mapping tables already contain an entry for the stripe, then the entry may be used to determine where to write the strips that ordinarily would have been written to the faulty storage devices 110 A. In either case, the volume manager 115 may then write the strips to the array of storage devices 110 , writing the strips that ordinarily would have been written to the faulty (e.g., read-only) storage devices 110 A to rescue space in the other storage devices 110 B.
  • the volume manager 115 may then write the strips to the array of storage devices 110 , writing the strips that ordinarily would have been written to the faulty (e.
  • FIG. 7 illustrates a schematic diagram of an embodiment of a RAID-0 system that may implement rescue space management with data block writing in accordance with example embodiments of the disclosure.
  • the embodiment illustrated in FIG. 7 may in some ways be constructed and operate in a manner similar to the embodiment described above with respect to FIGS.
  • a strip may be spit into rescue blocks (which may also be referred to as R-blocks) that may be distributed across the rescue spaces of some or all of the remaining unaffected storage devices 110 B.
  • the volume manager 115 may maintain an emergency mapping table in which each entry may simply be a stripe ID to indicate that the stripe has been mapped to the rescue space in the unaffected storage devices 110 B.
  • Strip (1,1) may include 10 blocks
  • the chunk of Strip (1,1) stored in new Drive ID 1 may include 3 blocks
  • the chunk stored in new Drive ID 2 may include 3 blocks
  • the chunk stored in new Drive ID 0 may include 4 blocks because new Drive ID 0 is less than 10 mod 3 (which equals 1).
  • the volume manager 115 may check the emergency mapping table to determine whether an entry has been made for the stripe of the strip to be written. If no entry has been made yet (e.g., the stripe is not registered) the volume manager 115 may create an entry in the emergency mapping table to indicate that chunks of the strip that ordinarily would have been written to the faulty storage device 110 A (the storage device that has transitioned to read-only mode), are to be written to the unaffected storage devices 110 B.
  • the emergency mapping table already contains an entry for the stripe, then the entry may be used to determine that chunks of the strip that ordinarily would have been written to the faulty storage device 110 A (the storage device that has transitioned to read-only mode), are to be written to the unaffected storage devices 110 B. In either case, the volume manager 115 may then write the chunks of the strip originally intended for Drive 1, to the rescue spaces of the unaffected storage devices 110 B as illustrated in FIG. 7 .
  • a fault resilient (FR) storage system such as an FR-RAID-0 system may implement one or more quality-of-service (QoS) management features in accordance with example embodiments of the disclosure.
  • QoS quality-of-service
  • a user and/or volume manager may adjust the size of strips in a RAID striping configuration, and/or the writing technique used to write data to in a rescue space on one or more storage devices in the RAID configuration, to provide a specific QoS level.
  • FIG. 8 illustrates an example embodiment of a system for implementing QoS management in a storage system in accordance with example embodiments of the disclosure.
  • the embodiment illustrated in FIG. 8 may include a QoS manager 802 configured to implement one or more QoS features for a storage array 804 , for example, through one or more control and/or configuration inputs 806 .
  • the storage array 804 which may be arranged, for example, as a RAID-0 array, may include a volume manager 815 and any number of storage devices 810 .
  • the storage array 804 may be implemented at least in part with any of the fault resilient storage devices, systems, and/or methods disclosed herein.
  • the QoS manager 802 may include QoS logic 808 that may receive, utilize, control, configure, direct, notify, and/or the like, any number of parameters relating to QoS such as the number of storage devices 811 A in the storage array 804 , the number of data blocks in a strip 811 B, one or more write methods 811 C used in a rescue space, the number of faulty storage devices 811 D that may be accommodated by the storage array 804 , the capacity or capacities 811 E of storage devices used in the storage array 804 , and/or the like.
  • QoS logic 808 may receive, utilize, control, configure, direct, notify, and/or the like, any number of parameters relating to QoS such as the number of storage devices 811 A in the storage array 804 , the number of data blocks in a strip 811 B, one or more write methods 811 C used in a rescue space, the number of faulty storage devices 811 D that may be accommodated by the storage array 804 , the capacity or capacities 811 E of storage
  • the write method 811 C may also be tuned, for example, because writing data blocks to rescue spaces on multiple storage devices may take less time than writing a strip to the rescue space of a single storage device.
  • Increasing the number of faulty storage devices 811 D that may be accommodated by the storage array 804 may reduce performance, for example, because accommodating more faulty devices may involve more allocating a greater percentage of storage device capacity
  • the QoS manager 802 may operate automatically, manually, or in any combination thereof.
  • the QoS manager 802 may operate automatically, for example, in response to monitoring one or more parameters 812 from the storage array 804 .
  • the QoS manager 802 may operate manually, for example, in response to one or more parameters 814 received through a user interface 816 .
  • the QoS manager 802 may provide one or more outputs 818 through the user interface 816 that may instruct a user to take one or more specific actions, for example, to add and/or remove one or more storage devices 810 .
  • Storage device capacity 1 TB; 4K write IOPS per storage device: 400K; 32K write IOPS per storage device: 200K; and RAID strip size: 32K.
  • a component or element may refer to only a portion of the component or element.
  • a reference to an integrated circuit may refer to all or only a portion of the integrated circuit, and a reference to a block may refer to the entire block or one or more subblocks.

Abstract

A method of operating a storage device may include determining a fault condition of the storage device, selecting a fault resilient mode based on the fault condition of the storage device, and operating the storage device in the selected fault resilient mode. The selected fault resilient mode may include one of a power cycle mode, a reformat mode, a reduced capacity read-only mode, a reduced capacity mode, a reduced performance mode, a read-only mode, a partial read-only mode, a temporary read-only mode, a temporary partial read-only mode, or a vulnerable mode. The storage device may be configured to perform a namespace capacity management command received from the host. The namespace capacity management command may include a resize subcommand and/or a zero-size namespace subcommand. The storage device may report the selected fault resilient mode to a host.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 17/232,144, filed Apr. 15, 2021, which claims priority to, and the benefit of, U.S. Provisional Patent Application Ser. No. 63/023,243 filed May 11, 2020 which is incorporated by reference; U.S. Provisional Patent Application Ser. No. 63/128,001 filed Dec. 18, 2020 which is incorporated by reference; U.S. Provisional Patent Application Ser. No. 63/051,158 filed Jul. 13, 2020 which is incorporated by reference; U.S. Provisional Patent Application Ser. No. 63/052,854 filed Jul. 16, 2020 which is incorporated by reference; and U.S. Provisional Patent Application Ser. No. 63/057,744 titled filed Jul. 28, 2020 which is incorporated by reference.
  • TECHNICAL FIELD
  • This disclosure relates generally to storage, and more specifically to systems, methods, and devices for fault resilient storage.
  • BACKGROUND
  • A storage device may encounter a fault condition that may affect the ability of the storage device to operate in a storage system.
  • The above information disclosed in this Background section is only for enhancement of understanding of the background of the invention and therefore it may contain information that does not constitute prior art.
  • SUMMARY
  • A method of operating a storage device may include determining a fault condition of the storage device, selecting a fault resilient mode based on the fault condition of the storage device, and operating the storage device in the selected fault resilient mode. The selected fault resilient mode may include a power cycle mode. The selected fault resilient mode may include a reformat mode. The selected fault resilient mode may include a reduced capacity read-only mode. The selected fault resilient mode may include a reduced capacity mode. The selected fault resilient mode may include a reduced performance mode. The selected fault resilient mode may include a read-only mode. The selected fault resilient mode may include a partial read-only mode. The selected fault resilient mode may include a temporary read-only mode. The selected fault resilient mode may include a temporary partial read-only mode. The selected fault resilient mode may include a vulnerable mode. The selected fault resilient mode may include a normal mode. The storage device may be configured to perform a command received from a host. The command may include a namespace capacity management command. The namespace capacity management command may include a resize subcommand. The namespace capacity management command may include a zero-size namespace command.
  • A storage device may include a storage medium, and a storage controller, wherein the storage controller is configured to determine a fault condition of the storage device, select a fault resilient mode based on the fault condition of the storage device, and operate the storage device in the selected fault resilient mode. The selected resilient mode may include one of a power cycle mode, a reformat mode, a reduced capacity read-only mode, a reduced capacity mode, a reduced performance mode, a read-only mode, a partial read-only mode, a temporary read-only mode, a temporary partial read-only mode, or a vulnerable mode. The storage device may be to perform a namespace capacity management command received from a host.
  • A system may include a host, and at least one storage device coupled to the host, wherein the storage device is configured to determine a fault condition of the storage device, select a fault resilient mode based on the fault condition of the storage device, operate in the selected fault resilient mode, and report the selected fault resilient mode to the host.
  • A method of operating a storage array may include determining a first fault resilient operating mode of a first fault resilient storage device of the storage array, determining a second fault resilient operating mode of a second fault resilient storage device of the storage array, allocating one or more rescue spaces of one or more additional fault resilient storage devices of the storage array, mapping user data from the first fault resilient storage device to the one or more rescue spaces, and mapping user data from the second fault resilient storage device to the one or more rescue spaces. The method may further include reassigning at least one device identifier (ID) of the one or more additional fault resilient storage devices to a device ID of the first fault resilient storage device. The at least one device ID of the one or more additional fault resilient storage devices may be reassigned based on a current unaffected device ID and a current faulty device ID. The method may further include redirecting one or more inputs and/or outputs (IOs) from the first fault resilient storage device to the one or more additional fault resilient storage devices. The user data may include a strip of data. The strip of data may be redirected to a target storage device of the one or more additional fault resilient storage devices based on a stripe ID of the user data. Mapping user data from the first fault resilient storage device to the one or more rescue spaces may include maintaining a first mapping table. Mapping user data from the second fault resilient storage device to the one or more rescue spaces may include maintaining a second mapping table. The one or more rescue spaces may have a rescue space percentage ratio of a storage device capacity. The rescue space percentage ratio may be greater than or equal to a number of failed storage devices accommodated by the storage array, divided by the total number of storage devices in the storage array. The one or more rescue spaces may be allocated statically. The one or more rescue spaces may be allocated dynamically.
  • A system may include a storage array including a first fault resilient storage device, a second fault resilient storage device, one or more additional fault resilient storage devices, and a volume manger configured to: determine a first fault resilient operating mode of the first fault resilient storage device, determine a second fault resilient operating mode of the second fault resilient storage device, allocate one or more rescue spaces of one or more additional fault resilient storage devices of the storage array, map user data from the first fault resilient storage device to the one or more rescue spaces, and map user data from the second fault resilient storage device to the one or more rescue spaces. The volume manger may be further configured to reassign at least one device identifier (ID) of the one or more additional fault resilient storage devices to a device ID of the first fault resilient storage device. The volume manger may be further configured to redirect one or more inputs and/or outputs (IOs) from the first fault resilient storage device to the one or more additional fault resilient storage devices. The user data may include a strip of data, and the volume manger may be further configured to redirect the strip of data to a target storage device of the one or more additional fault resilient storage devices based on a stripe ID of the user data. The one or more rescue spaces have a rescue space percentage ratio of a storage device capacity, and the rescue space percentage ratio may be based on a number of failed storage devices accommodated by the storage array, divided by a total number of storage devices in the storage array.
  • An apparatus may include a volume manager for a storage array, the volume manager may include logic configured to: determine a first fault resilient operating mode of a first fault resilient storage device of the storage array, determine a second fault resilient operating mode of a second fault resilient storage device of the storage array, allocate one or more rescue spaces of one or more additional fault resilient storage devices of the storage array, map user data from the first fault resilient storage device to the one or more rescue spaces, and map user data from the second fault resilient storage device to the one or more rescue spaces. The user data may include a strip of data, and the strip of data may be redirected to a target storage device of the one or more additional fault resilient storage devices based on a stripe identifier (ID) of the user data. The one or more rescue spaces have a rescue space percentage ratio of a storage device capacity, and the rescue space percentage ratio may be based on a number of failed storage devices accommodated by the storage array, divided by a total number of storage devices in the storage array.
  • A method of operating a storage array may include allocating a first rescue space of a first fault resilient storage device of the storage array, allocating a second rescue space of a second fault resilient storage device of the storage array, determining a fault resilient operating mode of a third fault resilient storage device of the storage array, and mapping user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode. A first block of the user data may be mapped to the first rescue space, and a second block of the user data may be mapped to the second rescue space. The user data may include a strip of data. A first portion of the strip of data may be mapped to the first rescue space, and the first portion of the strip of data may include a number of data blocks based on a size of the strip of data and a size of the data blocks. The number of data blocks may be further based on a total number of storage devices in the storage array. The method may further include reassigning at least one device identifier (ID) of the first fault resilient storage device to a device ID of the third fault resilient storage device. The method may further include redirecting one or more inputs and/or outputs (IOs) from the third fault resilient storage device to the first rescue space and the second rescue space. The first rescue space may have a capacity based on a capacity of the first fault resilient storage device and a total number of storage devices in the storage array. The first rescue space may have a capacity of strips based on a size of the first rescue space and a block size.
  • A system may include a storage array including a first fault resilient storage device, a second fault resilient storage device, a third fault resilient storage device, and a volume manger configured to allocate a first rescue space of the first fault resilient storage device, allocate a second rescue space of the second fault resilient storage device, determine a fault resilient operating mode of the third fault resilient storage device, and map user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode. The volume manager may be further configured to map a first block of the user data to the first rescue space, and map a second block of the user data to the second rescue space. The user data may include a strip of data, and the volume manager may be further configured to map a first portion of the strip of data to the first rescue space. The first portion of the strip of data may include a number of data blocks based on a size of the strip of data and a size of the data blocks. The number of data blocks may be further based on a total number of storage devices in the storage array.
  • A method of operating a storage array may include determining a first parameter of a first fault resilient storage device of the storage array, determining a second parameter of a second fault resilient storage device of the storage array, and determining a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter. The method may further include adjusting the first parameter based on the QoS. The first parameter may be adjusted automatically based on monitoring the first parameter. The first parameter may be adjusted automatically based on monitoring the second parameter. The first parameter may be adjusted by configuring a component of the storage array. The first parameter may be adjusted by controlling the operation of a component of the storage array. The first parameter may include one of a number of storage devices in the storage array, a number of data blocks in a strip of user data for the first fault resilient storage device, a write method for redirecting data from the first fault resilient storage device to the second fault resilient storage device, a number of faulty storage devices supported by the storage array, or a storage capacity of the first fault resilient storage device.
  • A system may include a storage array including a first fault resilient storage device, a second fault resilient storage device, and a volume manger configured to determine a first parameter of a first fault resilient storage device, determine a second parameter of a second fault resilient storage device, and determine a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter. The volume manger may be further configured to adjust the first parameter based on the QoS. The volume manger may be further configured to adjust the first parameter automatically based on monitoring the first parameter. The volume manger may be further configured to adjust the first parameter automatically based on monitoring the second parameter.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The figures are not necessarily drawn to scale and elements of similar structures or functions may generally be represented by like reference numerals or portions thereof for illustrative purposes throughout the figures. The figures are only intended to facilitate the description of the various embodiments described herein. The figures do not describe every aspect of the teachings disclosed herein and do not limit the scope of the claims. To prevent the drawing from becoming obscured, not all of the components, connections, and the like may be shown, and not all of the components may have reference numbers. However, patterns of component configurations may be readily apparent from the drawings. The accompanying drawings, together with the specification, illustrate example embodiments in accordance with the disclosure, and, together with the description, serve to explain the principles of the present disclosure.
  • FIG. 1 illustrates an embodiment of a storage system in accordance with example embodiments of the disclosure.
  • FIG. 2A illustrates a table of some possible fault conditions that may be encountered by an embodiment of a fault resilient storage device in accordance with example embodiments of the disclosure.
  • FIG. 2B illustrates a table of some example embodiments of fault resilient modes and associated space types that may be implemented by a storage device in accordance with example embodiments of the disclosure.
  • FIG. 2C illustrates a table of some example embodiments of commands and subcommands that may be implemented by a storage device in accordance with example embodiments of the disclosure.
  • FIG. 2D illustrates a table of commands that a storage device in accordance with example embodiments of the disclosure may implement through an API.
  • FIG. 3A illustrates a flow chart of an embodiment of a method for operating in a fault resilient mode in accordance with example embodiments of the disclosure.
  • FIG. 3B illustrates a flow chart of an embodiment of a method of operating a storage device in accordance with example embodiments of the disclosure.
  • FIG. 4A illustrates a schematic data layout diagram of a RAID-0 system performing a write operation in accordance with example embodiments of the disclosure.
  • FIG. 4B illustrates a schematic data layout diagram of a RAID-0 system performing a read operation in accordance with example embodiments of the disclosure.
  • FIG. 4C illustrates a schematic data layout diagram of a RAID-0 system performing a remapping and write operation in accordance with example embodiments of the disclosure.
  • FIG. 5A illustrates a flowchart for a method for operating a RAID-0 system in accordance with example embodiments of the disclosure.
  • FIG. 5B illustrates a flow chart showing details of a method for operating a RAID-0 storage system in accordance with example embodiments of the disclosure.
  • FIG. 6 illustrates a schematic diagram of an embodiment of a RAID-0 system in accordance with example embodiments of the disclosure.
  • FIG. 7 illustrates a schematic diagram of an embodiment of a RAID-0 system that may implement rescue space management with data block writing in accordance with example embodiments of the disclosure.
  • FIG. 8 illustrates an example embodiment of a system for implementing quality-of service (QoS) management in a storage system in accordance with example embodiments of the disclosure.
  • FIG. 9 illustrates an embodiment of a method of operating a storage array in accordance with example embodiments of the disclosure.
  • FIG. 10 illustrates an embodiment of another method of operating a storage array in accordance with example embodiments of the disclosure.
  • FIG. 11 illustrates an embodiment of a further method of operating a storage array in accordance with example embodiments of the disclosure.
  • DETAILED DESCRIPTION Overview
  • Some of the principles in accordance with example embodiments of the disclosure relate to storage devices that may continue to operate in one or more fault resilient modes in case of a fault of the storage device. For example, a storage device may continue to operate in a limited manner that may enable a storage system to recover quickly and/or efficiently from the fault of the storage drive.
  • In some embodiments, a storage device may implement any number of the following fault resilient (FR) modes:
  • Some embodiments may implement a power cycle mode which may involve self-healing based on power cycling the storage device.
  • Some embodiments may implement a reformat mode which may involve self-healing based on formatting all or a portion of the storage device.
  • Some embodiments may implement a reduced capacity read-only mode in which a first portion of the storage space of the storage device may operate normally, and a second portion may operate as read-only storage space.
  • Some embodiments may implement a reduced capacity mode in which a first portion of the storage space of the storage device may operate normally, and a second portion may not be available for input and/or output (I/O) operations.
  • Some embodiments may implement a reduced performance mode in which one or more aspects of the performance of the storage device may be reduced.
  • Some embodiments may implement a read-only mode in which data may be read from, but not written to, the storage device.
  • Some embodiments may implement a partial read-only mode in which a first portion of the storage space of the storage device may operate as read-only storage space, and a second portion may not be available for normal input and/or output (I/O) operations.
  • Some embodiments may implement a temporary read-only mode in which data may be read from, but not written to, the storage space of the storage device, which may be temporarily valid, and may become invalid.
  • Some embodiments may implement a temporary partial read-only mode in which data may be read from, but not written to, a first portion of the storage space of the storage device, which may be temporarily valid, and may become invalid. A second portion may not be available for input and/or output (I/O) operations.
  • Some embodiments may implement a vulnerable mode in which the storage device may not be available for input and/or output (I/O) operations.
  • Some embodiments may implement a normal mode in which the storage device may operate normally.
  • In some embodiments, a storage device may implement one or more commands which may be used by a host to determine and/or manage one or more features of the storage device. For example, in some embodiments, a storage device may implement a namespace capacity management command which may include a resize and/or zero-size subcommand.
  • The principles disclosed herein have independent utility and may be embodied individually, and not every embodiment may utilize every principle. However, the principles may also be embodied in various combinations, some of which may amplify the benefits of the individual principles in a synergistic manner.
  • Storage Systems
  • FIG. 1 illustrates an embodiment of a storage system in accordance with example embodiments of the disclosure. The embodiment illustrated in FIG. 1 may include a host 105 and one or more storage devices 110. Some or all of the one or more storage devices 110 may be connected directly to the host 105, and some or all of the one or more storage devices 110 may be connected to the host 105 through a volume manager 115 as shown in FIG. 1. Each storage device 110 may include a storage controller 120 (or “control circuit”) and a storage media 125. In some embodiments, a storage device 110 may experience an internal fault condition, and the storage device may exhibit various fault resilient behaviors, as discussed in further detail below, to mitigate the system-level impact of the fault condition.
  • The one or more storage devices 110 may be implemented with any type of storage apparatus and associated storage media including solid state drives (SSDs), hard disk drives (HDDs), optical drives, drives based on any type of persistent memory such as cross-gridded nonvolatile memory with bulk resistance change, and/or the like, and/or any combination thereof. Data in each storage device may be arranged as blocks, key-value structures, and/or the like, and/or any combination thereof. Each storage device 110 may have any form factor such as 3.5 inch, 2.5 inch, 1.8 inch, M.2, MO-297, MO-300, Enterprise and Data Center SSD Form Factor (EDSFF) and/or the like, using any connector configuration such as Serial ATA (SATA), Small Computer System Interface (SCSI), Serial Attached SCSI (SAS), U.2, and/or the like, and using any storage interface and/or protocol such as Peripheral Component Interconnect (PCI), PCI express (PCIe), Nonvolatile Memory Express (NVMe), NVMe-over-Fabrics (NVMe-oF), Ethernet, InfiniBand, Fibre Channel, and/or the like. Some embodiments may be implemented entirely or partially with, and/or used in connection with, a server chassis, server rack, dataroom, datacenter, edge datacenter, mobile edge datacenter, and/or any combinations thereof, and/or the like.
  • Any or all of the host 105, volume manager 115, storage controller 120, and/or any other components disclosed herein may be implemented with hardware, software, or any combination thereof, including combinational logic, sequential logic, one or more timers, counters, registers, state machines, complex programmable logic devices (CPLDs), field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), complex instruction set computer (CISC) processors and/or reduced instruction set computer (RISC) processors, and/or the like executing instructions stored in volatile memories such as dynamic random access memory (DRAM) and/or static random access memory (SRAM), nonvolatile memory such as flash memory and/or the like, as well as graphics processing units (GPUs), neural processing units (NPUs), and/or the like.
  • Although the inventive principles are not limited to any particular implementation details, for purposes of illustration, in some embodiments, each storage device 110 may be implemented as an SSD in which the storage media may be implemented, for example, with not AND (NAND) flash memory, and each storage controller 120 may implement any functionality associated with operating the SSD including a flash translation layer (FTL), a storage interface, and any functionality associated with implementing the fault resilient features disclosed herein. The smallest erasable unit in the storage device 110 may be referred to as a “block” and the smallest writeable unit in the storage device 110 may be referred to as a “page”.
  • The storage media 125 may have a retention period (which may depend on the usage history of the storage media 125, and, as such, may vary within the storage media 125); data that has been stored longer than the retention period (i.e., data having an age exceeding the retention period) may become unreliable and may be said to have expired. Data may be stored in the storage media 125 using an error correcting code, which may be, e.g., a block code. When data is read from the storage media 125, a quantity of raw data, referred to as a code block, may be read from the storage media 125, and an attempt to decode it may be made. If the attempt fails, additional attempts (e.g., read retrials) may be made. With use, a portion, e.g., a block, of the storage media 125 may degrade to a point that the retention period becomes unacceptably short, and the block may be classified as a “bad block”. To avoid allowing this circumstance to render the entire storage media 125 inoperable, reserve space, referred to as “bad block management reserve space” may be present (e.g., included in each flash memory die or in each flash memory plane), and the controller 120, or another controller internal to the flash memory die or to the flash memory plane may begin to use a block in the reserve and cease to use the bad block.
  • The operations and/or components described with respect to the embodiment illustrated in FIG. 1, as well as all of the other embodiments described herein, are example operations and/or components. In some embodiments, some operations and/or components may be omitted and/or other operations and/or components may be included. Moreover, in some embodiments, the temporal and/or spatial order of the operations and/or components may be varied. Although some components may be illustrated as individual components, in some embodiments, some components shown separately may be integrated into single components, and/or some components shown as single components may be implemented with multiple components.
  • Fault Conditions
  • FIG. 2A illustrates a table of some possible fault conditions that may be encountered by an embodiment of a fault resilient storage device in accordance with example embodiments of the disclosure. Each fault condition (or “fault state”) may be labeled with a case identifier (“Case ID”) in the first column. The second column may indicate an operation status of the storage in the fault state. The third column of the table may indicate, for each case, whether valid user data remain available. The fourth column of the table may indicate whether the storage device 110 may eventually be returned to full functionality, for example, by reformatting the storage media 125.
  • Case 1 may include a fault condition in which the storage device 110 may no longer be capable of performing read or write operations, and that may not be resolved by cycling power and/or reformatting the storage media. A state in which the storage device 110 behaves in this manner may have various sub-states, with, e.g., each sub-state corresponding to a different failure mechanism. Such a state, or fault condition (in which the storage device 110 is no longer capable of performing read or write operations, and that may not be resolved by cycling power or reformatting the storage media) may be caused, for example, by a portion of the controller's firmware becoming corrupted (in which case it may be possible for the controller to restart into a safe mode, in which the corrupted instructions may not be executed) or by a failure of a processing circuit in the storage device 110 (e.g., the failure of a processing circuit that manages interactions with the storage media but is not responsible for communications with the host 105). When a fault condition of this type occurs, the storage device 110 may respond to a read or write command from the host 105 with an error message.
  • Case 2 may include a fault condition (i) in which the storage device 110 may no longer be capable of performing read or write operations and (ii) from which recovery may be possible by cycling the power of the storage device 110, by reformatting the storage media (e.g., nonvolatile memory (NVM)), by re-loading firmware, and/or the like. Such a fault condition may be caused, for example, by a program execution error of the controller 120 of the storage device 110 (e.g., a pointer that is out of range as a result of a bit flip in random-access memory (RAM) of the controller 120, or an instruction that is incorrect, as a result of a bit flip). If the program execution error has not caused the controller 120 to write incorrect data to the storage media 125 (e.g., if the program execution error occurred since the most recent write to storage media by the controller), then power cycling the storage device may be sufficient to restore the storage device 110 to normal operation. If the program execution error has caused the controller 120 to write erroneous data to the storage media 125, then reformatting the storage media 125, and/or re-loading firmware may be sufficient to restore the storage device 110 to normal operation.
  • Case 3 may include a fault condition that may be mitigated by operating the storage device 110 in a read-only mode, and for which reformatting the storage media 125 may not restore full functionality. Examples of such faults may include (i) a temperature sensor failure, and (ii) a portion of the storage media 125 having transitioned to a read-only mode. In the case of the temperature sensor failure, the failure may be detected by determining that a temperature sensor reading is out of range (e.g., has exceeded a threshold temperature), and in such a case the risk of overheating of the storage device 110 may be reduced by avoiding write operations, which may dissipate more power than read operations. The transitioning to a read-only mode of a portion of the storage media 125 may occur, for example, for flash memory storage media 125, if a flash memory plane or die exhausts a bad block management reserve space used for run time bad block management. For example, the storage device 110 may, while attempting to performing a read operation, make an unsuccessful attempt to decode a data item, determine that the block storing the data is a bad block and upon moving the data from the bad block to the bad block management reserve space, determine that the remaining bad block management reserve space is less than a threshold size and therefore insufficient to ensure the reliability of the plane or die. The storage device 110 may then determine that bad block management is no longer being performed, and transition to a read-only mode. In some embodiments, data item may refer to any quantity of data being processed in one operation, e.g., the data resulting from decoding a code block may be a data item.
  • Case 4 may include fault condition that may be mitigated by operating the storage device 110 in a write-through mode. For example, if a power supply backup capacitor in the storage device 110 fails, the device may, in response to a write command received from the host, complete the write to the storage media 125 before sending a command completion to the host 105, so that if power fails before the write to the storage media 125 has been completed, the host is not incorrectly informed that the write was completed successfully. Operating in the write-through mode may result in a reduction of performance (e.g., in terms of throughput and/or latency).
  • Case 5 may include a fault condition that may be mitigated by operating the storage device 110 in a manner that reduces power dissipation. For example, in the case of a temperature sensor failure, the storage device 110 may operate in a read-only mode as mentioned above, or it may reduce the rate at which operations (e.g., write operations, which may dissipate more power than read operations) may be performed, to reduce power dissipation in the storage device 110. For example, the storage device 110 may perform a first write to the storage media, then wait, during an interval corresponding to the reduced performance (the waiting resulting in a decrease in the rate at which write operations are performed), and then perform another (e.g., a second) write to the storage media.
  • Case 6 may include a fault condition that may be mitigated by operating the storage device 110 in a read-only mode, and for which reformatting the storage media 125 may restore full functionality.
  • Fault Resiliency
  • Based on one or more fault conditions such as those exemplified by the cases listed in FIG. 2A, in some embodiments, various levels of fault resiliency may be implemented by a storage device 110 in accordance with example embodiments of the disclosure. For example, some embodiments may implement a fully resilient mode, a partially resilient mode, and a vulnerable mode. In the fully resilient mode, the storage device 110 may operate with self-healing features, and the storage device 110 may be capable of recovering full functionality (although the user data in the device may be lost) by resetting operations such as power cycling, re-loading firmware, or formatting of the storage media 125.
  • In the partially resilient mode, the storage device 110 may operate with lower performance, reduced capacity, or reduced capability, when a fault condition exists. For example, as mentioned above, if a power supply backup capacitor fails, writes may be completed (e.g., command completions may be sent to the host 105) only after data is written to the storage media 125 (i.e., synchronous writes may be performed), slowing the operation of the storage device 110, and reducing its performance. The user data may be preserved in this circumstance. As another example, storage device 110 may operate with reduced capacity if the reserve space for bad block management run time bad block (RTBB) is exhausted. In this circumstance, the affected dies in the storage device 110 may be excluded from the disk space and the overall disk capacity may be reduced. The user data on the lost space may be lost. For example, if a set in IO determinism or a zone in a zoned namespace is no longer capable of accepting new data writes, the set or the zone may be excluded from disk space but the remaining disk space may remain available for read and write operations. The user data on the zone or set may be lost.
  • The storage device 110 may operate with reduced capability, for example, if a storage device 110 does not allow write operations, and switches to a read-only mode. In some embodiments, the storage device 110 may be capable of operating in two types of read-only mode: a sustainable read-only mode (which may be referred to as a “first read-only mode”), and an unsustainable read-only mode (which may be referred to as a “second read-only mode”). In the sustainable read-only mode, the storage device 110 may continue to serve read requests beyond the retention period of the storage media 125. The unsustainable read-only mode may be employed, for example, when it may not be feasible to operate in the sustainable read-only mode, e.g., when there is insufficient unused storage space to set up a rescue space. When transitioning to the unsustainable read-only mode, the storage device 110 may send to the host 105 a notification that the storage device 110 is operating in the second (unsustainable) read-only mode, and that data items stored in the storage device 110 may be allowed to expire (e.g., at the end of their respective retention periods). In the unsustainable read-only mode, the storage device 110 may continue to serve read requests during the retention period of the storage media 125, and, if the storage device 110 encounters data integrity issues (as detected, for example, by one or more unsuccessful attempts to decode data during read operations), the storage device 110 may report the invalid data region.
  • A storage device 110 operating in the vulnerable mode may be incapable of performing normal read and/or write operations, and may perform a graceful exit, for example, by continuing to receive commands from the host and returning errors.
  • Thus, in some embodiments, a storage device having one or more fault resilient features in accordance with example embodiments of the disclosure may extend and/or organize the features so that a host may utilize them systematically, and the device may continue to operate in some capacity despite a fault condition. In some embodiments, for example, if the storage device is used for a RAID (Redundant Array of Independent (or Inexpensive) Drives) or RAIN (Redundant Array of Independent Nodes), and a node fails, the system may recover the data by copying the data from the accessible space of the storage device without calculating the stripe parity.
  • Logical Block Address Space Types
  • In some embodiments, various logical block address (LBA) space types may be implemented by a storage device having fault resiliency features in accordance with example embodiments of the disclosure. These LBA space types may be used, for example, by a storage device such as that illustrated in FIG. 1. Some examples of LBA space types may include performing space (P), underperforming space (UP), read-only space (RO), volatile read-only space (VRO), and inaccessible space (IA). In some embodiments, an LBA space may also refer to any unit of storage space such as a page, a partition, a set, a zone, and/or the like.
  • Performing (P) space may include LBA space containing valid data, which may be capable of being read and written in a normal manner without sacrificing performance. Data in performing space may be valid.
  • Underperforming (UP) space may include LBA space containing valid data, which may be capable of being read and written in a normal manner, but with degraded performance (e.g., degraded write performance).
  • Read-only (RO) space may include LBA space containing valid data, which may be read-only. For example, a storage device may refuse to write data received from a host and/or may respond with error messages to write commands from the host directed to this type of LBA space. The data in read-only space may remain valid for a period of time exceeding the retention period.
  • Volatile read-only (VRO) space may include read-only space, and the storage device may respond with error messages to write commands from a host directed to this type of LBA space. Data in this type of LBA space may be temporarily valid, and may become invalid when it expires, i.e., when the age of the data in its storage media reaches the retention period of the storage media.
  • Inaccessible (IA) space may include LBA space containing invalid data, which may not be accessible from the host.
  • Fault Resilient Modes
  • In some embodiments, LBA space types may be used, for example, to implement some embodiments of fault resilient modes. FIG. 2B illustrates a table of some example embodiments of fault resilient modes and associated LBA space types that may be implemented by a storage device in accordance with example embodiments of the disclosure. The Mode column of the table illustrated in FIG. 2B may include a fault resilient mode number and a name which may be used to identify the mode, for example, in an application programming interface (API) through which one or more features of a storage device may be accessed in accordance with example embodiments of the disclosure. The columns labeled as P, UP, RO, VRO, and IA in the table illustrated in FIG. 2B may indicate an amount of performing (P), underperforming (UP), read-only (RO), volatile read-only (VRO), and inaccessible (IA) LBA space, respectively, that may be used in the corresponding mode.
  • In some embodiments, the modes illustrated in FIG. 2B may be invoked, for example, by a host through an API. In some embodiments, a host may query the storage device using a get feature command as described below. In some implementations, modes identified with an asterisk (*) may provide a host with detailed information about each type of LBA space used by the mode in response to a get feature command. In some implementations, information about the LBA space used by the other commands may be implicit. For example, in the power cycle mode (Mode 1), all memory may be of the performing (P) type. In some embodiments, however, other combinations of LBA space types, and/or portions thereof, may be used.
  • In some embodiments, a storage device may implement any number of the following fault resilient modes. For example, a device manufacturer may implement different combinations of these and other fault resilient modes in different products.
  • A power cycle mode (Mode 1) may involve self-healing based on power cycling the storage device. For example, a storage device may experience a fault condition based on one or more flipped bits in memory such as SRAM or DRAM. A flipped bit may be caused, for example, by aging, heating, and/or radiation due to an antenna or high elevations above sea level which may interfere with memory cells. A storage device with a fault resilient power cycle mode may have self-healing capabilities such that power cycling the storage device (e.g., removing then reapplying power) may reset the current state and restore the failed SSD to a normal state. In this case, one or more inflight commands in a submission queue may be lost. Whether the user data of the storage device remains valid may depend on implementation details such as the partitioning of the device, the extent to which different circuits of the storage controller are reset, and/or the like. In some embodiments, in a power cycle mode, the entire storage space of the storage device (100 percent) may operate normally (e.g., as performing (P) space).
  • A reformat mode (Mode 2) may involve self-healing based on formatting all or a portion of the storage device. In some embodiments, formatting the storage device may reset its current state and restore the failed storage device to its normal state. However, depending on the implementation details (e.g., quick format, full format, partitioning details, and/or the like) all data on the disk may be lost. In some embodiments, in a reformat mode, the entire storage space of the storage device (100 percent) may operate normally (e.g., as performing (P) space).
  • In a reduced capacity read-only mode (Mode 3), a first portion of the storage space (e.g., X percent) of the storage device may operate normally (e.g., as performing (P) space), and a second portion (e.g., (100−X) percent) may operate as read-only (RO) storage space. Thus, the size of the performance (P) space in the storage device may be reduced, and the storage device may behave like a normal drive with respect to that space, but the read-only (RO) type of space may not be writable. In some embodiments, the storage device may provide a list of LBA ranges for the performance (P) and/or read-only (RO) spaces to a host, for example, in response to a get feature command. If the storage device supports the IO determinism, the LBA range may represent a set. If the storage device supports Zoned Namespaces (ZNS), the LBA range may represent a zone. In some embodiments, the storage device may also provide information about address ranges for sets and/or ZNS in response to a get feature command.
  • In a reduced capacity mode (Mode 4), a first portion of the storage space (e.g., X percent) of the storage device may operate normally (e.g., as performing (P) space), and a second portion (e.g., (100−X) percent) may be inaccessible (IA). Thus, the size of the performance (P) space in the storage device may be reduced, and the storage device may behave like a normal drive with respect to that space, but inaccessible (IA) space may not be available for normal IOs. For example, if an RTBB is exhausted, the problematic die may be excluded from the disk space, and thus, the overall disk capacity may be reduced. The storage device may provide a list of LBA ranges for the performance (P) and/or inaccessible (IA) type of space. If the storage device supports the IO determinism, the LBA range may represent a set. If the storage device supports ZNS, the LBA range may represent a zone. In some embodiments, the storage device may provide information about the LBA ranges, sets, zones, and/or the like, in response to a get feature command.
  • In a reduced performance mode (Mode 5) one or more aspects of the performance of the storage device may be reduced. For example, the storage device may perform normal operations, but at reduced throughput and/or latency. In some embodiments, a storage device may include one or more back-up capacitors that, in the event of a loss of the main power supply, may provide power to the storage device for a long enough period of time to enable the storage device to complete a write operation. If one or more of these back-up capacitors fail, the storage device may not notify a host that a write operation is complete until after the data is written to the media. (This may be referred to as a synchronous write operation.) This may reduce the input and/or output operations per second (IOPS) and/or increase latency, thereby reducing the performance of the storage device. Thus, in some embodiments, reduced performance mode may operate with 100 percent underperforming (UP) space. Depending on the implementation details, some or all of the user data may remain valid. In some embodiments, the storage device may provide speculative performance information to a host which may enable the host to make decisions on sending write data to the storage device in a manner that may mitigate the system-level impact of the fault condition.
  • In a read-only mode (Mode 6), the storage device may only allow read operations and may block external write operations. Depending on the implementation details, data in read-only space may remain valid, for example, after the retention period. Read-only mode may operate with 100 percent read-only (RO) space.
  • In a partial read-only mode (Mode 7), a first portion of the storage space (e.g., X percent) of the storage device may operate as read-only (RO) space, and a second portion (e.g., (100−X) percent) may be inaccessible (IA) space. Thus, the storage device may only allow read operations and external write operations may be prohibited in the first portion of the storage space. Depending on the implementation details, data in the read-only space may still valid, for example, after the retention period. The storage device may provide a list of LBA ranges for the read-only (RO) and/or inaccessible (IA) types of space. If the storage device supports the IO determinism, the LBA range may represent a set. If the storage device supports ZNS, the LBA range may represent a zone. In some embodiments, the storage device may provide information about the LBA ranges, sets, zones, and/or the like, in response to a get feature command.
  • In a temporary read-only mode (Mode 8), data may be read from the storage space of the storage device, which may operate with 100 percent VRO space, but external writes may be prohibited. Data in this space may be temporarily valid but may become invalid after the retention period.
  • In a temporary partial read-only mode (Mode 9), data may be read from a first portion (e.g., X percent) of the storage space of the storage device, which may operate as VRO space, while external writes may be prohibited. A second portion (e.g., (100−X) percent) may be inaccessible (IA) space. Data in the first portion may be temporarily valid but may become invalid after the retention period. If the storage device supports the IO determinism, the LBA range may represent a set. If the storage device supports ZNS, the LBA range may represent a zone. In some embodiments, the storage device may provide information about the LBA ranges, sets, zones, and/or the like, in response to a get feature command.
  • In a vulnerable mode (Mode 10), the storage device may not be available for 1/O operations. However, it may continue to receive commands from the host and return errors.
  • In a normal mode (Mode 11), the storage device may operate normally.
  • Commands
  • In some embodiments, a storage device in accordance with example embodiments of the disclosure may implement one or more commands which may be used, for example, by a host to query the storage device and/or manage one or more features of the storage device. FIG. 2C illustrates a table of some example embodiments of commands and subcommands that may be implemented by a storage device in accordance with example embodiments of the disclosure. The subcommand column of the table illustrated in FIG. 2C may indicate a name which may be used to identify the subcommand, for example, in an API through which the commands, and responses thereto, may be passed.
  • A get feature command, which may include a subcommand as shown in the table illustrated in FIG. 2C, may be passed from a host to a storage device, which may return a response thereto. In some embodiments, a storage device may respond as follows to a get feature command based on the subcommand.
  • A resiliency type subcommand (FR_INFO_RESILIENCY_TYPE) may return a type of fault resiliency in case of a failure. For example, the storage device may indicate which of the fault resilient modes illustrated in FIG. 2B the device has selected to operate in based on the fault condition it has encountered.
  • A retention period subcommand (FR_INFO_RETENTION_PERIOD) may return an average retention period of the data without reprogramming the storage media. In some embodiments, this may be the upper-bound of retention time for data in the storage media from the time of the failure. This subcommand may be used, for example, with temporary read-only mode (Mode 8) and/or temporary partial read-only mode (Mode 9).
  • An earliest expiry subcommand (FR_INFO_EARLIEST_EXPIRY) may return a maximum time remaining for data integrity. In some embodiments, this may be the lower-bound of retention time for data in the storage media from the time of the failure. The unit of time may be determined, for example, based on a patrol period. This subcommand may be used, for example, with temporary read-only mode (Mode 8) and/or temporary partial read-only mode (Mode 9).
  • An IOPS subcommand (FR_INFO_IOPS) may return a percentage of the maximum available IOPS the storage device may be able to handle based on the fault condition. This subcommand may be used, for example, with reduced performance mode (Mode 5).
  • A bandwidth subcommand (FR_INFO_BW) may return a percentage of the maximum available bandwidth the storage device may be able to handle based on the fault condition. This subcommand may be used, for example, with reduced performance mode (Mode 5).
  • A space subcommand (FR_INFO_SPACE) may return an amount of storage space that may be available in the storage device based on the fault condition. This subcommand may be used, for example, with reduced capacity read-only mode (Mode 3) and/or reduced capacity mode (Mode 4).
  • A namespace capacity management command, which may include a subcommand as shown in the table illustrated in FIG. 2C, may be passed from a host to a storage device, which may respond by performing the action indicated by the subcommand. In some embodiments, a storage device may respond as follows to a namespace capacity management (NCM) command based on the subcommand. In some embodiments, a namespace may be implemented as a quantity of non-volatile memory (NVM) that may be formatted into logical blocks.
  • A resize command (FR_NAMESPACE_RESIZE) may cause the storage device to resize a namespace based on one or more parameters that may be included with the command. In some embodiments, this subcommand may apply to storage device that may support two or more namespaces. In some embodiments, the namespaces may support NVMe resizing.
  • A zero-sized namespace command (FR_NAMESPACE_ZERO_SIZE) may cause the storage device to reduce the size of a rescue space to zero.
  • Application Programming Interface
  • In some embodiments, as mentioned above, a storage device in accordance with example embodiments of the disclosure may implement an API to enable a host to query the storage device and/or manage one or more features of the storage device. FIG. 2D illustrates a table of commands that a storage device in accordance with example embodiments of the disclosure may implement through an API. Some embodiments may include a hierarchy of enumerated constants, within the category of fault resilient features, that the storage device may employ to respond. As illustrated in FIG. 2D, the hierarchy may include a first level, including a fully resilient status, a partially resilient status, and a vulnerable status. Sub-statuses, and sub-sub-statuses may also be defined. For example, as illustrated in FIG. 2D, the partially resilient status may include a first sub-status indicating a loss of capability, and the first sub-status may include a first sub-sub-status, indicating operation in the sustainable read-only mode, and a second sub-sub-status, indicating operation in the unsustainable read-only mode. In some embodiments, an API may be implemented, for example, using NVMe commands.
  • A feature command (FAULT_RESILIENT_FEATURE) may return the fault resilient classes and features in each class that the storage device may support.
  • A status command (FAULT_RESILIENT_STATUS) may return the status of the storage device after a fault resilient recovery is performed.
  • A volatile blocks command (FAULT_RESILIENT_VOLATILE_BLOCKS (H)) may return a list of LBA ranges that reach to the retention period in the next H hours. In some embodiments, this may be used to determine the blocks that need to be relocated for unsustainable read-only.
  • An invalid data blocks command (FAULT_RESILIENT_INVALID_DATA_BLOCKS) may return a list of LBA ranges that may become invalid after switching to a fault resilient mode.
  • Additional Embodiments
  • FIG. 3A illustrates a flow chart of an embodiment of a method for operating in a fault resilient mode in accordance with example embodiments of the disclosure. The method illustrated in FIG. 3A may be implemented, for example, by the systems and/or components illustrated in FIG. 1. The method may begin at operation 300. At operation 305, the host 105 may send or receives data from the storage device 110. at operation 310, the host 105 may determine whether an error has occurred in the storage device 110 At operation 315, the storage device 110 may perform an internal diagnosis and determine its fault resilient status (e.g., fully resilient, partially resilient, or vulnerable). At operation 320, the storage device 110 may modify its performance, capacity, and/or capability (e.g., transitioning to a read-only mode) based on the diagnosis. At operation 325, the storage device 110 may post the status upon request from the host 105 based on an application programming interface (API). At operation 330, the host 105 may route data of a given type to the storage device 110 or to a different storage 110 device at a given bandwidth based on the status. The method may end at operation 335.
  • FIG. 3B illustrates a flow chart of an embodiment of a method of operating a storage device in accordance with example embodiments of the disclosure. The method may begin at operation 350. At operation 355, the method may determine a fault condition of the storage device. At operation 360, the method may select a fault resilient mode based on the fault condition of the storage device. At operation 365, the method may operate the storage device in the selected fault resilient mode. The method may end at operation 370.
  • The operations and/or components described with respect to the embodiment illustrated in FIGS. 3A and 3B, as well as all of the other embodiments described herein, are example operations and/or components. In some embodiments, some operations and/or components may be omitted and/or other operations and/or components may be included. Moreover, in some embodiments, the temporal and/or spatial order of the operations and/or components may be varied. Although some components may be illustrated as individual components, in some embodiments, some components shown separately may be integrated into single components, and/or some components shown as single components may be implemented with multiple components.
  • Any number of embodiments and/or variations on the embodiments disclosed herein may also be constructed. A storage controller such as a field programmable gate array (FPGA) or embedded processor may perform internal block checks and send asynchronous updates to the host 105 on the status of the storage device 110. Events may occur and be transmitted to the host 105 (e.g., temperature, or other parameters internal to the device). The host 105 may poll the storage devices 110 on a predetermined schedule, for example, if there is no device driver feature for providing notification. A storage controller may monitor the historical performance of the storage device 110 and use machine learning to provide predictive analytics (e.g., a likelihood of the storage device being in a given fault resilient state). Commands (e.g., NVMe commands) may be implemented and/or expanded, for example, to report the state of the storage device 110).
  • In some embodiments, the host may: (i) send different data types (e.g., file types such as image, video, text, or high-priority or low-priority data), based on the status of the storage device 110 (for instance, high priority data or real-time data may not be written to a device that is considered in the partially vulnerable mode); (ii) reduce the transmission rate if the storage device 110 is in a partially vulnerable state and in a lower performance state; (iii) send a reduced total amount of data if the storage device 110 is in a partially vulnerable and lower capacity state; (iv) read data at the greatest rate possible, and/or store the data elsewhere, if the storage device 110 is in a partially vulnerable unsustainable read-only mode, so as to avoid exceeding the retention period (in such a circumstance, the host may calculate the needed data rate based on the amount of data to be copied and on the retention period); (v) ignore data “read” from a vulnerable storage device 110 since it may be erroneous, and delete the data as it is received by the host 105; (vi) temporarily reroute read/write input and output to a cache in a fully resilient storage device 110 that is being power cycled and/or formatted, based on messages that control the timing of such events between the host and the storage devices 110. A storage controller on a partially vulnerable storage device that has had a capacity decrease may filter incoming data writes and only write a portion of that data to the storage device 110. In some cases, the filtering may include compression. Such a storage controller may receive various types of data (e.g., file types such as image, video, text, or high-priority or low-priority data) from a host 105 and filter based on the status of the storage device 110. For instance, the storage controller may determine that high priority data may not be written to a storage device 110 that is in the partially vulnerable mode. The storage controller may send a rejection message to the host 105 and give a reason for the rejection. Alternatively, the storage controller may filter out a certain type of data (e.g., image data) for writing to a partially resilient lower-capacity state storage device 110. For example, if a storage device 110 loses performance (e.g., operates at a reduced write rate), latency-sensitive reads and writes may be rejected.
  • Fault Resilient System with Fault Resilient Storage Devices
  • In some embodiments, a RAID-0 system including an array of storage devices 110 and a volume manager 115 may be constructed to accommodate a transition of any of the fault resilient storage devices 110 of the RAID-0 system to a read-only mode. In normal operation, the volume manager 115 may be responsible for striping data across the array of storage devices 110, e.g., writing one strip of each stripe to a respective storage device 110 of the array of storage devices 110. In such a system, when any of the array of storage devices 110 transitions to a read-only mode (indicated as 110A), the RAID-0 system may transition to a second operating mode (which may also be referred to as an emergency mode), and the volume manager 115 for the array of storage devices 110 may (i) allocate a rescue space on each of the remaining, unaffected storage devices 110B (e.g., those that remain in a read-write state) for metadata and rescued user data from faulty storage device 110A, and/or (ii) create and/or maintain a mapping table (which may also be referred to as an emergency mapping table). Rescue space may be pre-allocated statically prior to system operation, dynamically during operation, or in any combination thereof.
  • The rescue space (which may be indicated as R) on each storage device 110A may be capable of storing n strips, where n=R/(strip size), R=C/M, C may be the capacity of each of the storage devices of the array of storage devices 110, and M may be the total number of storage devices. In some embodiments, the volume manager 115 may be implemented as an independent component, or may be partially or fully integrated into the host, a RAID controller of the RAID-0 system (which may, for example, be housed in a separate enclosure from the host), or in any other configuration. In some embodiments, the volume manager 115 may be implemented, for example, with an FPGA. The RAID-0 system may be self-contained and may virtualize the array of storage devices 110 so that from the perspective of the host the RAID-0 system may appear as a single storage device. In some embodiments, the volume manager may be implemented as a processing circuit (discussed in further detail below) configured (e.g., by suitable software or firmware) to perform the operations described herein as being performed by the volume manager.
  • When the RAID-0 system is operating in an emergency mode, and a write command is received from the host 105 requesting that data be written to a stripe of the array of storage devices 110, the volume manager 115 may check the emergency mapping table to determine whether the stripe is registered e.g., whether an entry has already been made for the stripe. If no entry has been made yet (e.g., the stripe is not registered, which may also be referred to as open-mapped), the volume manager 115 may create an entry in the emergency mapping table to indicate where a strip, that ordinarily would have been written to the faulty storage device 110A (the storage device that has transitioned to read-only mode), is to be written. If the emergency mapping table already contains an entry for the stripe, then the entry may be used to determine where to write the strip that ordinarily would have been written to the faulty storage device 110A. In either case, the volume manager 115 may then write each strip, as illustrated in FIG. 4A, to the array of storage devices 110, writing the strip 405 that ordinarily would have been written to the faulty (e.g., read-only) storage device 110A to rescue space in another storage device 110B.
  • When a read command is received from the host 105 requesting that data of a stripe be read from the array of storage devices 110, the volume manager 115 may check the emergency mapping table to determine whether an entry has been made for the stripe. If no entry has been made, then, as illustrated in FIG. 4B, the volume manager 115 may read the stripe as it would have, in ordinary operation, reading a strip from each of the storage devices 110, including the faulty storage device 110A. If the emergency mapping table contains an entry for the stripe, then the entry may be used to determine where to read the strip that ordinarily would have been read from the faulty storage device 110A.
  • The remapping of strips that ordinarily would have been written to the faulty storage device 110A may be accomplished, for example, as follows. Each storage device 110 of the array of storage devices 110 may have a drive identification number (or “drive ID”), which may be a number between zero and M−1, where M may be the number of storage devices 110 in the array of storage devices 110. The volume manager 115 may reassign the drive identification numbers, e.g., assign to each unaffected storage device 110B of the array of storage devices 110 an alternate drive identification number to be used for performing read or write operations for registered stripes (read operations for unregistered stripes may continue to use the original drive identification numbers). The following formula (Formula A) may be used to generate the alternate drive identification numbers:

  • If drive ID<faulty drive ID,

  • new drive ID=((drive ID−1)+M)mod(M−1)

  • Otherwise,

  • new drive ID=((drive ID−1)+(M−1))mod(M−1).  (Formula A)
  • The effect of Formula A may be (i) to assign, to each storage device having an identification number less than the original drive identification number of the faulty storage device, the respective original drive identification number, and/or (ii) to assign, to each storage device having an identification number greater than the original drive identification number of the faulty storage device, the respective original drive identification number minus one.
  • Using the alternate drive numbers, a target drive, to which a strip that ordinarily would have been written to the faulty storage device 110A may be written, may be identified (e.g., on a per stripe basis) using the formula Target Drive ID=sid % (M−1) where Target Drive ID may be the alternate drive identification number of the target drive, sid may be the stripe identifier of the strip that ordinarily may have been written to the faulty storage device 110, and “%” may be the modulo (mod) operator.
  • FIG. 4C is a schematic diagram of an embodiment of a RAID-0 system in accordance with example embodiments of the disclosure. The embodiment illustrated in FIG. 4C may include four fault resilient storage devices 110 (i.e., M=4) in which the storage device identified as Drive 1 has transitioned to a read-only mode. Using Formula A described above, Drive 0 may remain mapped to new Drive ID 0 (e.g., 3 mod 3), Drive 2 may be mapped to new Drive ID 1 (e.g., 4 mod 3), and Drive 3 may be mapped to new Drive ID 2 (e.g., 5 mod 3).
  • The target drive ID (e.g., for a read or write operation) may be implicitly determined by the equation Target Drive ID=Stripe ID % (M−1). For example, if M=4 and Stripe 1 is written, Stripe ID=1, and thus, Target Drive ID=1% 3=1. That is, the target drive may be the storage device 110B with alternate (New) drive identification number 1 (i.e., previous Drive 2). Within the storage device, the rescue space may be split into strips (which may be referred to as rescue strips, or R-Strips) the size of which may be the same as the strip size. In some embodiments, the emergency mapping table may contain an entry for each strip having the format (Stripe ID, R-Strip ID) in which the first element may be the Stripe ID, and the second element may be the R-strip ID on the target drive. For example, an entry of (1,0) in the emergency mapping table may indicate that Strip (1,1) is mapped to R-Strip (1,0) as shown in FIG. 4C.
  • FIG. 5A illustrates a flowchart for a method for operating a RAID-0 system in accordance with example embodiments of the disclosure. At 505, A storage device 110 in a RAID-0 system has a fault and transitions to a read-only mode; at 510, the affected storage device 110 performs an internal diagnosis and determines that its fault resilient status is partially resilient and read-only; at 515, the volume manager 115 determines that the affected storage device 110 is in a read-only mode and reassigns the IDs of (“live”) unaffected storage devices; at 520, the volume manager 115 receives a write operation, adds an entry to an emergency mapping table to indicate that the strip of the affected device is redirected to a target (unaffected) storage device 110, and the entire strip is written to a rescue space of the target (unaffected) storage device based on the new drive IDs of the unaffected storage devices; and, at 525, the volume manager 115 receives a read command from the host 105, and reads all strips of a stripe from the live unaffected storage devices 110 of the RAID system while the strip of the affected storage device is read from the rescue space of the target (unaffected) storage device.
  • FIG. 5B illustrates a flow chart showing details of a method for operating a RAID-0 storage system in accordance with example embodiments of the disclosure. The method includes, at 530, determining that the first storage device is in a read-only state and that the second storage device is in a read-write state; at 535, performing a write operation, of a first stripe, to the storage system, by writing a portion of the first stripe to the second storage device, and making an entry in a mapping table for the first stripe; at 540, performing a first read operation, of a second stripe, from the storage system, by reading a portion of the second stripe from the first storage device and the second storage device; and at 545, performing a second read operation, of the first stripe, from the storage system, by determining that the mapping table includes an entry for the first stripe, and reading a portion of the first stripe from the second storage device.
  • System with Resilience to N-Device Failures
  • In some embodiments, a RAID-0 system may be constructed to accommodate the failure of multiple (e.g., N) fault resilient storage devices 110. An example embodiment of such a system may in some ways be constructed and operate in a manner similar to the embodiment described above with respect to FIGS. 4A-4C, but the size of the rescue space R for each storage device may be determined by considering the number N of faulty storage devices 110A the system may accommodate. For example, in some embodiments, a rescue space percentage ratio (e.g., b percent, which may also be referred to as a reservation ratio) may be used to determine the size of the rescue space R for each storage device, where R=(b/100)*C. In some embodiments in which a system has M storage devices and may accommodate N faulty storage devices that may transition to read-only mode, setting b such that N/M<=b/100 may ensure that all data from the N faulty storage devices 110A may be written to rescue space in the remaining unaffected storage devices (which may be referred to as live storage devices). For example, in a system that may have five storage devices and that may accommodate two faulty storage devices that may transition to read-only mode, b may be set to 2/5=40 percent. Thus, the size of the rescue space R for each storage device may be set to R=40/100*C. The rescue space R on each unaffected storage device 110A may be capable of storing n strips, where n=R/(strip size), but in this embodiment, R may be set to R=(b/100)*C. Rescue space may be pre-allocated statically prior to system operation, dynamically during operation, or in any combination thereof.
  • In a system that may accommodate N fault resilient storage device failures, M′ may represent the number of unaffected (e.g., live) storage devices such that M′<=M. In some embodiments, the drive IDs of the unaffected storage devices 110B may be reassigned according to the following formula (Formula B):

  • If current drive ID>current faulty drive ID

  • new drive ID=((current drive ID−1)+(M′−1))mod(M′−1)

  • Otherwise

  • new drive ID=((current drive ID−1)+M′)mod(M′−1).  (Formula B)
  • Using the alternate drive numbers, a target storage device for a write operation may be implicitly identified (e.g., on a per stripe basis) using the formula Target Drive ID=sid % (M′−1) where Target Drive ID may be the alternate (new) drive identification number of the target storage device, and sid may be the stripe identifier of the strip that ordinarily may have been written to the faulty storage device 110A, and which may now be written to the target storage device having the Target Drive ID.
  • FIG. 6 illustrates a schematic diagram of an embodiment of a RAID-0 system in accordance with example embodiments of the disclosure. The embodiment illustrated in FIG. 6 may include four fault resilient storage devices 110 (i.e., M=4) in which the storage devices identified as Drive 1 and Drive 2 have transitioned to a read-only mode. Therefore, M′=2. Using Formula B described above, Drive 0 may remain mapped to new Drive ID 0 (e.g., 1 mod 1), and Drive 3 may be mapped to new Drive ID 1 (e.g., 3 mod 1).
  • Also, using the formula Target Drive ID=sid % (M′−1), if stripe 1 is written, Stripe ID=1, and thus, Target Drive ID=1% 2=1. That is, the target drive may be the storage device 110B with alternate (New) drive identification number 1 (i.e., previous Drive 2).
  • In some embodiments, when a first faulty storage device 110A transitions to a read-only mode, the RAID-0 system may transition to an emergency mode in which the volume manager 115 may (i) allocate a rescue space on each of the remaining, unaffected storage devices 110B (if adequate rescue space has not been allocated already, or if insufficient space has been allocated) for metadata and rescued user data from a faulty storage device 110A, and/or (ii) create and/or maintain a first mapping table for the first faulty storage device 110A. The RAID-0 system may then operate in a manner similar to the single device failure embodiment described above.
  • In some embodiments, if a second faulty storage device 110A transitions to a read-only mode, the RAID-0 system may once again allocate a rescue space on each of the remaining, unaffected storage devices 110B (if adequate rescue space has not been allocated already, or if insufficient space has been allocated) for metadata and rescued user data from a faulty storage device 110A. In some embodiments, the RAID-0 system may then create and/or maintain a second mapping table for the second faulty storage device 110A. Each of the mapping tables may be designated as the Lth mapping table, where L=1 . . . M′, and the Lth mapping table corresponds to the Lth faulty storage device. In other embodiments, a RAID-0 system may create and/or modify a single mapping table to map data stripes and/or strips of all of the faulty storage devices 110A to the unaffected storage devices 110B. In some embodiments, one or more mapping tables may be stored in a reserved rescue space, for example, before a Disk Data Format (DDF) structure for a RAID configuration.
  • The RAID-0 system may then reassign drive IDs of the unaffected storage devices 110B, for example, based on Formula B, and proceed to operate with the two faulty storage devices 110A operating in read-only mode.
  • When a read command is received from the host, the volume manager 115 may check the one or more emergency mapping tables to determine whether an entry has been made for the stripe to be read. If no entry has been made, then the volume manager 115 may read the stripe as it would have, in ordinary operation, reading a strip from each of the storage devices 110, including the two faulty storage devices 110A. If the one or more emergency mapping tables contain an entry for the stripe, then the entry may be used to determine where to read the strip that ordinarily would have been read from one or both of the faulty storage devices 110A.
  • When a write command is received from the host, the volume manager 115 may check the one or more emergency mapping tables to determine whether an entry has been made for the stripe. If no entry has been made yet (e.g., the stripe is not registered) the volume manager 115 may create an entry in the one or more emergency mapping tables to indicate where the strips that ordinarily would have been written to the faulty storage devices 110A (the storage devices that have transitioned to read-only mode), are to be written. If the one or more emergency mapping tables already contain an entry for the stripe, then the entry may be used to determine where to write the strips that ordinarily would have been written to the faulty storage devices 110A. In either case, the volume manager 115 may then write the strips to the array of storage devices 110, writing the strips that ordinarily would have been written to the faulty (e.g., read-only) storage devices 110A to rescue space in the other storage devices 110B.
  • Rescue Space Management with Data Block Write
  • FIG. 7 illustrates a schematic diagram of an embodiment of a RAID-0 system that may implement rescue space management with data block writing in accordance with example embodiments of the disclosure. The embodiment illustrated in FIG. 7 may include four fault resilient storage devices 110 (i.e., M=4) in which the storage device identified as Drive 1 has transitioned to a read-only mode. The embodiment illustrated in FIG. 7 may in some ways be constructed and operate in a manner similar to the embodiment described above with respect to FIGS. 4A-4C, but rather than redirecting an entire strip from a faulty storage device 110A to the rescue space of a single unaffected storage device 110B, a strip may be spit into rescue blocks (which may also be referred to as R-blocks) that may be distributed across the rescue spaces of some or all of the remaining unaffected storage devices 110B.
  • Within each storage device 110, some or all of the rescue space may be split into rescue blocks (which may be referred to as R-blocks). The size of R-blocks may be set, for example, to the same size as a data block size used generally by the storage device.
  • In some embodiments, the volume manager 115 may maintain an emergency mapping table in which each entry may simply be a stripe ID to indicate that the stripe has been mapped to the rescue space in the unaffected storage devices 110B. For example, in the embodiment illustrated in FIG. 7 in which the storage device 110A designated as Drive ID 1 has entered read-only mode, an entry in the emergency mapping table of (stripe ID)=(1) may indicate that Strip (1,1) is split into 3 chunks and mapped to all unaffected (e.g., live) storage devices 110B.
  • In some embodiments, the portion of the strip from the faulty storage device that may be stored in the rescue space of each storage device (which may be referred to as a chunk) may be equal to the strip size/block size/(M−1). To accommodate a strip size and block size that may not be evenly divided into the number of unaffected storage devices 110B, the chunk stored in the rescue space of a target storage device 110B that satisfies the formula Target Drive ID<(strip size/block size) mod (M−1) may include an extra block. Thus, for the example illustrated in FIG. 7 in which Strip (1,1) may include 10 blocks, the chunk of Strip (1,1) stored in new Drive ID 1 may include 3 blocks, the chunk stored in new Drive ID 2 may include 3 blocks, but the chunk stored in new Drive ID 0 may include 4 blocks because new Drive ID 0 is less than 10 mod 3 (which equals 1).
  • In the embodiment illustrated in FIG. 7, the target drive ID (e.g., for a read or write operation) may be implicitly determined, for example, by the equation Target Drive ID=Stripe ID % (M−1) where M may indicate the number of storage devices 110. For example, if M=4 and Stripe 1 is written, Stripe ID=1, and thus, Target Drive ID=1%3=1.
  • In the embodiment illustrated in FIG. 7, the volume manager 115 may reassign the drive identification numbers for each unaffected storage device 110B using, for example, Formula C as follows:

  • If drive ID<faulty drive ID,

  • new drive ID=((drive ID−1)+M)mod(M−1)

  • Otherwise,

  • new drive ID=((drive ID−1)+(M−1))mod(M−1).  (Formula C)
  • In some embodiments, the size of rescue space R in each storage device 110 may be set, for example, to R=C/M, where C may be the capacity of each of the storage devices of the array of storage devices 110, and M may be the total number of storage devices. The rescue space R in each storage device 110 may be capable of storing n blocks, where n=R/(block size).
  • When a read command is received from the host, the volume manager 115 may check the emergency mapping table to determine whether an entry has been made for the stripe of the strip to be read. If no entry has been made, then the volume manager 115 may read the stripe as it would have in ordinary operation, reading a strip from each of the storage devices 110, including the faulty storage device 110A. If the emergency mapping table contains an entry for the stripe, the chunks of the strip corresponding to the faulty storage device 110A (in this example, Drive 1) may be read from the rescue space of the unaffected storage devices 110B (in this example, the storage devices with new Drive IDs 0, 1, and 2) and reassembled into Strip (1,1).
  • When a write command is received from the host, the volume manager 115 may check the emergency mapping table to determine whether an entry has been made for the stripe of the strip to be written. If no entry has been made yet (e.g., the stripe is not registered) the volume manager 115 may create an entry in the emergency mapping table to indicate that chunks of the strip that ordinarily would have been written to the faulty storage device 110A (the storage device that has transitioned to read-only mode), are to be written to the unaffected storage devices 110B. If the emergency mapping table already contains an entry for the stripe, then the entry may be used to determine that chunks of the strip that ordinarily would have been written to the faulty storage device 110A (the storage device that has transitioned to read-only mode), are to be written to the unaffected storage devices 110B. In either case, the volume manager 115 may then write the chunks of the strip originally intended for Drive 1, to the rescue spaces of the unaffected storage devices 110B as illustrated in FIG. 7.
  • Fault Resilient System with Quality-of-Service Management
  • In some embodiments, a fault resilient (FR) storage system such as an FR-RAID-0 system may implement one or more quality-of-service (QoS) management features in accordance with example embodiments of the disclosure. For example, a user and/or volume manager may adjust the size of strips in a RAID striping configuration, and/or the writing technique used to write data to in a rescue space on one or more storage devices in the RAID configuration, to provide a specific QoS level.
  • FIG. 8 illustrates an example embodiment of a system for implementing QoS management in a storage system in accordance with example embodiments of the disclosure. The embodiment illustrated in FIG. 8 may include a QoS manager 802 configured to implement one or more QoS features for a storage array 804, for example, through one or more control and/or configuration inputs 806. The storage array 804, which may be arranged, for example, as a RAID-0 array, may include a volume manager 815 and any number of storage devices 810. In some embodiments, the storage array 804 may be implemented at least in part with any of the fault resilient storage devices, systems, and/or methods disclosed herein.
  • The QoS manager 802 may include QoS logic 808 that may receive, utilize, control, configure, direct, notify, and/or the like, any number of parameters relating to QoS such as the number of storage devices 811A in the storage array 804, the number of data blocks in a strip 811B, one or more write methods 811C used in a rescue space, the number of faulty storage devices 811D that may be accommodated by the storage array 804, the capacity or capacities 811E of storage devices used in the storage array 804, and/or the like.
  • For example, in some embodiments, a QoS metric such as performance may be influenced by the parameters 811A-811E in any number of the following manners. Increasing the number of storage devices 811A in the storage array 804 may increase performance, for example, in terms of storage capacity, latency, throughput, and/or the like. The number of data blocks 811B in a strip may be tuned based on the type of anticipated storage transactions. For example, using larger data blocks may provide greater throughput with larger, less frequent transactions, whereas smaller data blocks may provide a greater number of input and/or output operations per second (IOPS) with small, more frequent transactions. The write method 811C may also be tuned, for example, because writing data blocks to rescue spaces on multiple storage devices may take less time than writing a strip to the rescue space of a single storage device. Increasing the number of faulty storage devices 811D that may be accommodated by the storage array 804 may reduce performance, for example, because accommodating more faulty devices may involve more allocating a greater percentage of storage device capacity
  • The QoS manager 802 may operate automatically, manually, or in any combination thereof. The QoS manager 802 may operate automatically, for example, in response to monitoring one or more parameters 812 from the storage array 804. The QoS manager 802 may operate manually, for example, in response to one or more parameters 814 received through a user interface 816. Additionally, the QoS manager 802 may provide one or more outputs 818 through the user interface 816 that may instruct a user to take one or more specific actions, for example, to add and/or remove one or more storage devices 810.
  • In some embodiments, given system requirements from a user, the QoS manager 802 may determine one or more parameters based on storage performance information. For example, a user may specify that the storage array 804 may operate as an RF-RAID-0 configuration that may accommodate one storage device failure with 500K IOPS for 32K blocks and a total storage capacity of 8 TB. Based on these inputs, the QoS manager 802 may determine the following parameters to arrive at a number of storage devices that may be used to provide the specified performance:
  • Storage device capacity: 1 TB;
    4K write IOPS per storage device: 400K;
    32K write IOPS per storage device: 200K; and
    RAID strip size: 32K.
    Solving for capacity: (1−1/M)*2*(M−1)>=8, M{circumflex over ( )}2−6M+1>0, (M−3){circumflex over ( )}2>8, and thus, M=6.
    Solving for performance: 200 K*(M−1)/2>=500 K, and thus, M=6.
    Therefore, six storage devices may be used to provide the specified performance.
  • In some embodiments, the QoS manager 802 and/or QoS logic 808 may be implemented with hardware, software, or any combination thereof, including combinational logic, sequential logic, one or more timers, counters, registers, state machines, CPLDs, FPGAs, ASICs, CISC processors and/or RISC processors, and/or the like executing instructions stored in volatile memories such as DRAM and/or SRAM, nonvolatile memory such as flash memory and/or the like, as well as GPUs, NPUs, and/or the like. The QoS manager 802 and/or QoS logic 808 may be implemented as one or more separate components, integrated with one or more other components such as the volume manager 815, a host, and/or any combination thereof.
  • FIG. 9 illustrates an embodiment of a method of operating a storage array in accordance with example embodiments of the disclosure. The method may begin at operation 902. At operation 904, the method may determine a first fault resilient operating mode of a first fault resilient storage device of the storage array. At operation 906, the method may determine a second fault resilient operating mode of a second fault resilient storage device of the storage array. At operation 908, the method may allocate one or more rescue spaces of one or more additional fault resilient storage devices of the storage array. At operation 910, the method may map user data from the first fault resilient storage device to the one or more rescue spaces. At operation 912, the method may map user data from the second fault resilient storage device to the one or more rescue spaces. The method may end at operation 914.
  • FIG. 10 illustrates an embodiment of another method of operating a storage array in accordance with example embodiments of the disclosure. The method may begin at operation 1002. At operation 1004, the method may allocate a first rescue space of a first fault resilient storage device of the storage array. At operation 1006, the method may allocate a second rescue space of a second fault resilient storage device of the storage array. At operation 1008, the method may determine a fault resilient operating mode of a third fault resilient storage device of the storage array. At operation 1010, the method may map user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode. The method may end at operation 1012.
  • FIG. 11 illustrates an embodiment of a further method of operating a storage array in accordance with example embodiments of the disclosure. The method may begin at operation 1102. At operation 1104, the method may determine a first parameter of a first fault resilient storage device of the storage array. At operation 1106, the method may determine a second parameter of a second fault resilient storage device of the storage array. At operation 1108, the method may determine a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter. The method may end at operation 1110.
  • The operations and/or components described with respect to the embodiments illustrated in FIGS. 9-11, as well as all of the other embodiments described herein, are example operations and/or components. In some embodiments, some operations and/or components may be omitted and/or other operations and/or components may be included. Moreover, in some embodiments, the temporal and/or spatial order of the operations and/or components may be varied.
  • The embodiments described above have been described in the context of various implementation details, but the principles of this disclosure are not limited to these or any other specific details. For example, some storage arrays have been described in the context of systems in which the capacity and/or size of storage devices and/or rescue spaces may be the same for each storage devices, but different capacity and/or size of storage devices and/or rescue spaces may be used. As another example, some embodiments have been described in the context of RAID system such as RAID-0, but the principles may also be applied to any other type of storage arrays.
  • As another example, some functionality has been described as being implemented by certain components, but in other embodiments, the functionality may be distributed between different systems and components in different locations and having various user interfaces. Certain embodiments have been described as having specific processes, operations, etc., but these terms also encompass embodiments in which a specific process, step, etc. may be implemented with multiple processes, operations, etc., or in which multiple processes, operations, etc. may be integrated into a single process, step, etc. A reference to a component or element may refer to only a portion of the component or element. For example, a reference to an integrated circuit may refer to all or only a portion of the integrated circuit, and a reference to a block may refer to the entire block or one or more subblocks. The use of terms such as “first” and “second” in this disclosure and the claims may only be for purposes of distinguishing the things they modify and may not indicate any spatial or temporal order unless apparent otherwise from context. In some embodiments, based on” may refer to “based at least in part on.” In some embodiments, “disabled” may refer to “disabled at least in part.” A reference to a first element may not imply the existence of a second element. Various organizational aids such as section headings and the like may be provided as a convenience, but the subject matter arranged according to these aids and the principles of this disclosure are not defined or limited by these organizational aids.
  • The various details and embodiments described above may be combined to produce additional embodiments according to the inventive principles of this patent disclosure. Since the inventive principles of this patent disclosure may be modified in arrangement and detail without departing from the inventive concepts, such changes and modifications are considered to fall within the scope of the following claims.

Claims (25)

1. A method of operating a storage array, the method comprising:
allocating a first rescue space of a first fault resilient storage device of the storage array;
allocating a second rescue space of a second fault resilient storage device of the storage array;
determining a fault resilient operating mode of a third fault resilient storage device of the storage array; and
mapping user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode.
2. The method of claim 1, wherein:
a first block of the user data is mapped to the first rescue space; and
a second block of the user data is mapped to the second rescue space.
3. The method of claim 1, wherein the user data comprises a strip of data.
4. The method of claim 3, wherein:
a first portion of the strip of data is mapped to the first rescue space; and
the first portion of the strip of data comprises a number of data blocks based on a size of the strip of data and a size of the data blocks.
5. The method of claim 4, wherein the number of data blocks is further based on a total number of storage devices in the storage array.
6. The method of claim 1, further comprising reassigning at least one device identifier (ID) of the first fault resilient storage device to a device ID of the third fault resilient storage device.
7. The method of claim 1, further comprising redirecting one or more inputs and/or outputs (IOs) from the third fault resilient storage device to the first rescue space and the second rescue space.
8. The method of claim 1, wherein the first rescue space has a capacity based on a capacity of the first fault resilient storage device and a total number of storage devices in the storage array.
9. The method of claim 3, wherein the first rescue space has a capacity of strips based on a size of the first rescue space and a block size.
10. A system comprising a storage array comprising:
a first fault resilient storage device;
a second fault resilient storage device;
a third fault resilient storage device; and
a volume manger configured to:
allocate a first rescue space of the first fault resilient storage device;
allocate a second rescue space of the second fault resilient storage device;
determine a fault resilient operating mode of the third fault resilient storage device; and
map user data from the third fault resilient storage device to the first rescue space and the second rescue space based on determining the fault resilient operating mode.
11. The system of claim 10, wherein the volume manager is further configured to:
map a first block of the user data to the first rescue space; and
map a second block of the user data to the second rescue space.
12. The system of claim 10, wherein:
the user data comprises a strip of data; and
the volume manager is further configured to map a first portion of the strip of data to the first rescue space.
13. The system of claim 12, wherein the first portion of the strip of data comprises a number of data blocks based on a size of the strip of data and a size of the data blocks.
14. The system of claim 13, wherein the number of data blocks is further based on a total number of storage devices in the storage array.
15. A method of operating a storage array, the method comprising:
determining a first parameter of a first fault resilient storage device of the storage array;
determining a second parameter of a second fault resilient storage device of the storage array; and
determining a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter.
16. The method of claim 15, further comprising adjusting the first parameter based on the QoS.
17. The method of claim 16, wherein the first parameter is adjusted automatically based on monitoring the first parameter.
18. The method of claim 16, wherein the first parameter is adjusted automatically based on monitoring the second parameter.
19. The method of claim 16, wherein the first parameter is adjusted by configuring a component of the storage array.
20. The method of claim 16, wherein the first parameter is adjusted by controlling the operation of a component of the storage array.
21. The method of claim 15, wherein the first parameter comprises one of a number of storage devices in the storage array, a number of data blocks in a strip of user data for the first fault resilient storage device, a write method for redirecting data from the first fault resilient storage device to the second fault resilient storage device, a number of faulty storage devices supported by the storage array, or a storage capacity of the first fault resilient storage device.
22. A system comprising a storage array comprising:
a first fault resilient storage device;
a second fault resilient storage device; and
a volume manger configured to:
determine a first parameter of a first fault resilient storage device;
determine a second parameter of a second fault resilient storage device; and
determine a quality-of-service (QoS) of the storage array based on the first parameter and the second parameter.
23. The system of claim 22, wherein the volume manger is further configured to adjust the first parameter based on the QoS.
24. The system of claim 23, wherein the volume manger is further configured to adjust the first parameter automatically based on monitoring the first parameter.
25. The system of claim 23, wherein the volume manger is further configured to adjust the first parameter automatically based on monitoring the second parameter.
US17/827,657 2020-05-11 2022-05-27 Systems, methods, and devices for fault resilient storage Pending US20220291996A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/827,657 US20220291996A1 (en) 2020-05-11 2022-05-27 Systems, methods, and devices for fault resilient storage

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US202063023243P 2020-05-11 2020-05-11
US202063051158P 2020-07-13 2020-07-13
US202063052854P 2020-07-16 2020-07-16
US202063057744P 2020-07-28 2020-07-28
US202063128001P 2020-12-18 2020-12-18
US17/232,144 US20210349782A1 (en) 2020-05-11 2021-04-15 Systems, methods, and devices for fault resilient storage
US17/827,657 US20220291996A1 (en) 2020-05-11 2022-05-27 Systems, methods, and devices for fault resilient storage

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US17/232,144 Continuation US20210349782A1 (en) 2020-05-11 2021-04-15 Systems, methods, and devices for fault resilient storage

Publications (1)

Publication Number Publication Date
US20220291996A1 true US20220291996A1 (en) 2022-09-15

Family

ID=78412659

Family Applications (3)

Application Number Title Priority Date Filing Date
US17/232,144 Pending US20210349782A1 (en) 2020-05-11 2021-04-15 Systems, methods, and devices for fault resilient storage
US17/741,440 Pending US20220269566A1 (en) 2020-05-11 2022-05-10 Systems, methods, and devices for fault resilient storage
US17/827,657 Pending US20220291996A1 (en) 2020-05-11 2022-05-27 Systems, methods, and devices for fault resilient storage

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US17/232,144 Pending US20210349782A1 (en) 2020-05-11 2021-04-15 Systems, methods, and devices for fault resilient storage
US17/741,440 Pending US20220269566A1 (en) 2020-05-11 2022-05-10 Systems, methods, and devices for fault resilient storage

Country Status (4)

Country Link
US (3) US20210349782A1 (en)
KR (1) KR20210138502A (en)
CN (1) CN113641303A (en)
TW (1) TW202147092A (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11803446B2 (en) 2020-07-13 2023-10-31 Samsung Electronics Co., Ltd. Fault resilient storage device
US20210349780A1 (en) * 2020-05-11 2021-11-11 Samsung Electronics Co., Ltd. Systems, methods, and devices for data recovery with spare storage device and fault resilient storage device
US11782791B2 (en) * 2021-08-25 2023-10-10 Western Digital Technologies, Inc. System and method for memory hang recovery using a distributed power management system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8099623B1 (en) * 2008-10-08 2012-01-17 Netapp, Inc. Efficient distributed hot sparing scheme in a parity declustered RAID organization
US20130173955A1 (en) * 2012-01-04 2013-07-04 Xtremlo Ltd Data protection in a random access disk array
US10884648B2 (en) * 2019-06-03 2021-01-05 International Business Machines Corporation Temporary relocation of data within local storage of a dispersed storage network

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6886108B2 (en) * 2001-04-30 2005-04-26 Sun Microsystems, Inc. Threshold adjustment following forced failure of storage device
US7395452B2 (en) * 2004-09-24 2008-07-01 Microsoft Corporation Method and system for improved reliability in storage devices
JP2007293448A (en) * 2006-04-21 2007-11-08 Hitachi Ltd Storage system and its power supply control method
JP4884885B2 (en) * 2006-08-25 2012-02-29 株式会社日立製作所 Storage controller and storage controller failure recovery method
US9542272B2 (en) * 2014-03-24 2017-01-10 Avago Technologies General Ip (Singapore) Pte. Ltd. Write redirection in redundant array of independent disks systems
US9477549B2 (en) * 2014-09-15 2016-10-25 Sandisk Technologies Llc Methods, systems, and computer readable media for address and data integrity checking in flash memory operations
US9519545B2 (en) * 2014-11-11 2016-12-13 Facebook, Inc. Storage drive remediation in a raid system
US10402092B2 (en) * 2016-06-01 2019-09-03 Western Digital Technologies, Inc. Resizing namespaces for storage devices
US10210060B2 (en) * 2016-08-22 2019-02-19 Hewlett Packard Enterprise Development Lp Online NVM format upgrade in a data storage system operating with active and standby memory controllers
US10936453B2 (en) * 2017-09-19 2021-03-02 Seagate Technology Llc Data storage systems using elastic spares
US11869586B2 (en) * 2018-07-11 2024-01-09 Pure Storage, Inc. Increased data protection by recovering data from partially-failed solid-state devices

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8099623B1 (en) * 2008-10-08 2012-01-17 Netapp, Inc. Efficient distributed hot sparing scheme in a parity declustered RAID organization
US20130173955A1 (en) * 2012-01-04 2013-07-04 Xtremlo Ltd Data protection in a random access disk array
US10884648B2 (en) * 2019-06-03 2021-01-05 International Business Machines Corporation Temporary relocation of data within local storage of a dispersed storage network

Also Published As

Publication number Publication date
TW202147092A (en) 2021-12-16
US20220269566A1 (en) 2022-08-25
US20210349782A1 (en) 2021-11-11
KR20210138502A (en) 2021-11-19
CN113641303A (en) 2021-11-12

Similar Documents

Publication Publication Date Title
US20220291996A1 (en) Systems, methods, and devices for fault resilient storage
US10459639B2 (en) Storage unit and storage system that suppress performance degradation of the storage unit
US20090077302A1 (en) Storage apparatus and control method thereof
US20230244570A1 (en) Fault resilient storage device
US9898201B2 (en) Non-volatile memory device, and storage apparatus to reduce a read retry occurrence frequency and prevent read performance from lowering
US10338844B2 (en) Storage control apparatus, control method, and non-transitory computer-readable storage medium
US8762771B2 (en) Method for completing write operations to a RAID drive pool with an abnormally slow drive in a timely fashion
US10802958B2 (en) Storage device, its controlling method, and storage system having the storage device
US20210349781A1 (en) Systems, methods, and devices for data recovery using parity space as recovery space
US20210349780A1 (en) Systems, methods, and devices for data recovery with spare storage device and fault resilient storage device
US7330955B2 (en) Recovery record for updating a system configuration
US20230297249A1 (en) Storage system and method of controlling storage system

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KI, YANG SEOK;RYU, SUNGWOOK;TRAN, ALAIN;AND OTHERS;SIGNING DATES FROM 20220412 TO 20220503;REEL/FRAME:064681/0583

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED