US20150248348A1 - Physically-addressable solid state disk (ssd) and a method of addressing the same - Google Patents

Physically-addressable solid state disk (ssd) and a method of addressing the same Download PDF

Info

Publication number
US20150248348A1
US20150248348A1 US14/697,544 US201514697544A US2015248348A1 US 20150248348 A1 US20150248348 A1 US 20150248348A1 US 201514697544 A US201514697544 A US 201514697544A US 2015248348 A1 US2015248348 A1 US 2015248348A1
Authority
US
United States
Prior art keywords
storage system
recited
physically
flash
bus
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/697,544
Inventor
Siamack Nemazie
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Avalanche Technology Inc
Original Assignee
Avalanche Technology Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/570,202 external-priority patent/US20130080687A1/en
Priority claimed from US13/673,866 external-priority patent/US20140047161A1/en
Priority claimed from US13/745,686 external-priority patent/US9009396B2/en
Application filed by Avalanche Technology Inc filed Critical Avalanche Technology Inc
Priority to US14/697,544 priority Critical patent/US20150248348A1/en
Publication of US20150248348A1 publication Critical patent/US20150248348A1/en
Assigned to ABEDIFARD, EBRAHIM reassignment ABEDIFARD, EBRAHIM ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ABEDIFARD, EBRAHIM
Assigned to NEMAZIE, SIAMACK reassignment NEMAZIE, SIAMACK ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NEMAZIE, SIAMACK
Assigned to Avalanche Technology reassignment Avalanche Technology ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NEMAZIE, SIAMACK
Assigned to SILICON VALLEY BANK reassignment SILICON VALLEY BANK SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AVALANCHE TECHNOLOGY, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/06Addressing a physical block of locations, e.g. base addressing, module addressing, memory dedication
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/023Free address space management
    • G06F12/0238Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory
    • G06F12/0246Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory in block erasable memory, e.g. flash memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/023Free address space management
    • G06F12/0253Garbage collection, i.e. reclamation of unreferenced memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0875Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches with dedicated cache, e.g. instruction or stack
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F13/38Information transfer, e.g. on bus
    • G06F13/42Bus transfer protocol, e.g. handshake; Synchronisation
    • G06F13/4204Bus transfer protocol, e.g. handshake; Synchronisation on a parallel bus
    • G06F13/4221Bus transfer protocol, e.g. handshake; Synchronisation on a parallel bus being an input/output bus, e.g. ISA bus, EISA bus, PCI bus, SCSI bus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • G06F3/0619Improving the reliability of storage systems in relation to data integrity, e.g. data losses, bit errors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0638Organizing or formatting or addressing of data
    • G06F3/064Management of blocks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0646Horizontal data movement in storage systems, i.e. moving data in between storage devices or systems
    • G06F3/0652Erasing, e.g. deleting, data cleaning, moving of data to a wastebasket
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • G06F3/0688Non-volatile semiconductor memory arrays
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C11/00Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor
    • G11C11/02Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using magnetic elements
    • G11C11/16Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using magnetic elements using elements in which the storage effect is based on magnetic spin effect
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C14/00Digital stores characterised by arrangements of cells having volatile and non-volatile storage properties for back-up when the power is down
    • G11C14/0009Digital stores characterised by arrangements of cells having volatile and non-volatile storage properties for back-up when the power is down in which the volatile element is a DRAM cell
    • G11C14/0036Digital stores characterised by arrangements of cells having volatile and non-volatile storage properties for back-up when the power is down in which the volatile element is a DRAM cell and the nonvolatile element is a magnetic RAM [MRAM] element or ferromagnetic cell
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C14/00Digital stores characterised by arrangements of cells having volatile and non-volatile storage properties for back-up when the power is down
    • G11C14/0054Digital stores characterised by arrangements of cells having volatile and non-volatile storage properties for back-up when the power is down in which the volatile element is a SRAM cell
    • G11C14/0081Digital stores characterised by arrangements of cells having volatile and non-volatile storage properties for back-up when the power is down in which the volatile element is a SRAM cell and the nonvolatile element is a magnetic RAM [MRAM] element or ferromagnetic cell
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C7/00Arrangements for writing information into, or reading information out from, a digital store
    • G11C7/10Input/output [I/O] data interface arrangements, e.g. I/O data control circuits, I/O data buffers
    • G11C7/1072Input/output [I/O] data interface arrangements, e.g. I/O data control circuits, I/O data buffers for memories with random access ports synchronised on clock signal pulse trains, e.g. synchronous memories, self timed memories
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1032Reliability improvement, data loss prevention, degraded operation etc
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/20Employing a main memory using a specific memory technology
    • G06F2212/202Non-volatile memory
    • G06F2212/2022Flash memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/21Employing a record carrier using a specific recording technology
    • G06F2212/214Solid state disk
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/45Caching of specific data in cache memory
    • G06F2212/69
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/70Details relating to dynamic memory management
    • G06F2212/702Conservative garbage collection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/72Details relating to flash memory management
    • G06F2212/7201Logical to physical mapping or translation of blocks or pages
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/72Details relating to flash memory management
    • G06F2212/7205Cleaning, compaction, garbage collection, erase control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/72Details relating to flash memory management
    • G06F2212/7207Details relating to flash memory management management of metadata or control data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/72Details relating to flash memory management
    • G06F2212/7208Multiple device management, e.g. distributing data over multiple flash devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2213/00Indexing scheme relating to interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F2213/0026PCI express

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Computer Hardware Design (AREA)
  • Computer Security & Cryptography (AREA)
  • Techniques For Improving Reliability Of Storages (AREA)

Abstract

A storage system includes a Central Processing Unit (CPU) that has a physically-addressed solid state disk (SSD) and is addressable using physical addresses associated with user data that are provided by a host. The user data is to be stored in or retrieved from the physically-addressed SSD in blocks. Further, a non-volatile memory module is coupled to the CPU and includes flash tables used to manage blocks in the physically addressed SSD. The flash tables have tables that are used to map logical to physical blocks for identifying the location of the user data in the physically-addressed SSD. The flash tables are maintained in the non-volatile memory modules.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 14/542,516, filed on Nov. 14, 2014, by Siamack Nemazie and entitled “A Computer System With Physically-Addressable Solid State Disk (SSD) and A Method of Addressing the Same”, which is a continuation of U.S. patent application Ser. No. 13/769,710, filed on Feb. 18, 2013, by Siamack Nemazie and entitled “Storage System Employing MRAM and Physically Addressed Solid State Disk”, which is a continuation-in-part of U.S. patent application Ser. No. 13/745,686, filed on Jan. 18, 2013, by Siamack Nemazie and entitled “Physically Addressed Solid State Disk Employing MRAM”, which is a continuation-in-part of U.S. patent application Ser. No. 13/673,866, filed on Nov. 9, 2012, by Siamack Nemazie and entitled “SYSTEM EMPLOYING MARAM AND PHYSICALLY ADDRESSED SOLID STATE DISK”, which is a continuation-in-part of U.S. patent application Ser. No. 13/570,202, filed on Aug. 8, 2012, by Siamack Nemazie and Ngon Van Le, and entitled “SOLID STATE DISK EMPLOYING FLASH AND MAGNETIC RANDOM ACCESS MEMORY (MRAM)”, which claims priority U.S. Provisional Application No. 61/538,697, filed on Sep. 23, 2011, entitled “Solid State Disk Employing Flash and MRAM”, by Siamack Nemazie, incorporated herein by reference as though set forth in full.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • This invention relates generally to storage systems, and particularly to storage systems utilizing physically-addressed solid state disk (SSD).
  • 2. Background
  • Solid State Drives (SSDs) using flash memories have become a viable alternative to Hard Disc Drives (HDDs) in many applications. Such applications include storage for notebook, tablets, servers and network-attached storage appliances. In notebook and tablet applications, storage capacity is not too high, and power and or weight and form factor are key metric. In server applications, power and performance (sustained read/write, random read/write) are key metrics. In network-attached storage appliances, capacity, power, and performance are key metrics with large capacity being achieved by employing a number of SSDs in the appliance. SSD may be directly attached to the system via a bus such as SATA, SAS or PCIe.
  • Flash memory is a block-based non-volatile memory with each block organized into and made of various pages. After a block is programmed into the flash memory, it must be erased prior to being programmed again. Most flash memory require sequential programming of pages within a block. Another limitation of flash memory is that blocks can only be erased for a limited number of times, thus, frequent erase operations reduce the life time of the flash memory. A Flash memory does not allow in-place updates. That is, it cannot overwrite existing data with new data. The new data are written to erased areas (out-of-place updates), and the old data are invalidated for reclamation in the future. This out-of-place update causes the coexistence of invalid (i.e. outdated) and valid data in the same block.
  • Garbage Collection is the process to reclaim the space occupied by the invalid data, by moving valid data to a new block and erasing the old block. But garbage collection results in significant performance overhead as well as unpredictable operational latency. As mentioned, flash memory blocks can be erased for a limited number of times. Wear leveling is the process to improve flash memory lifetime by evenly distributing erases over the entire flash memory (within a band).
  • The management of blocks within flash-based memory systems, including SSDs, is referred to as flash block management and includes: Logical to Physical Mapping; Defect management for managing defective blocks (blocks that were identified to be defective at manufacturing and grown defective blocks thereafter); Wear leveling to keep program/erase cycle of blocks within a band; Keeping track of free available blocks; and Garbage collection for collecting valid pages from a plurality of blocks (with a mix of valid and invalid page) into one block and in the process creating free blocks. The flash block management requires maintaining various tables referred to as flash block management tables (or “flash tables”). These tables are generally proportional to the capacity of SSD.
  • Generally, the flash block management tables can be constructed from metadata maintained on flash pages. Metadata is non-user information written on a page. Such reconstruction is time consuming and generally performed very infrequently upon recovery during power-up from a failure (such as power fail). In one prior art technique, the flash block management tables are maintained in a volatile memory, and as mentioned, the flash block management tables is constructed from metadata maintained in flash pages during power-up. In another prior art technique, the flash block management tables are maintained in a battery-backed volatile memory, utilized to maintain the contents of volatile memory for an extended period of time until power is back and tables can be saved in flash memory. In yet another prior art technique, the flash block management tables are maintained in a volatile RAM, the flash block management tables are periodically and/or based on some events (such as a Sleep Command) saved (copied) back to flash, and to avoid the time consuming reconstruction upon power-up from a power failure additionally a power back-up means provides enough power to save the flash block management tables in the flash in the event of a power failure. Such power back-up may comprise of a battery, a rechargeable battery, or a dynamically charged super capacitor.
  • The flash block management is generally performed in the SSD and the tables reside in the SSD. Alternatively, the flash block management may be performed in the system by a software or hardware, commands additionally include commands for flash management commands and the commands use physical addresses rather than logical addresses. An SSD with commands using physical addresses is referred to as Physically-Addressed SSD. The flash block management tables are maintained in the (volatile) system memory.
  • A storage system (also referred to as “storage array”, or “storage appliance”) is a special purpose computer system attached to a network, dedicated to data storage and management. The storage system may be connected to Internet Protocol (IP) Network running Network File System (NFS) protocol or Common Internet File System (CIFS) protocol or Internet Small Computer System (iSCSI) protocol or to a Storage Area Network (SAN) such as Fiber Channel (FC) or Serial Attached SCSI (SAS) for block storage.
  • In a storage system employing physically-addressed SSD which maintains the flash block management tables on the system memory that has no power back-up means for the system and no power back-up means for the system memory, the flash block management tables that reside in the system memory are lost and if copies are maintained in the flash onboard the SSD, the copies may not be updated and/or may be corrupted if power failure occurs during the time a table is being saved (or updated) in the flash memory.
  • Hence, during a subsequent power up, during initialization, the tables have to be inspected for corruption due to power fail and, if necessary, recovered. The recovery requires reconstruction of the tables to be completed by reading metadata from flash pages and results in further increase in delay for system to complete initialization. The process of complete reconstruction of all tables is time consuming, as it requires metadata on all pages of SSD to be read and processed to reconstruct the tables. Metadata is non-user information written on a page. This flash block management table recovery, during power-up, further delays the system initialization, the time to initialize the system is a key metric in many applications.
  • Yet another similar problem of data corruption and power fail recovery arises in SSDs and also HDDs when write data for write commands (or queued write commands when command queuing is supported) is cached in a volatile system memory and command completion issued prior to writing to media (flash or Hard Disc Drive). It is well known in the art that caching write data for write commands (or queued write commands when command queuing is supported) and issuing command completion prior to writing to media significantly improves performance.
  • Additionally, file systems and storage systems employ journaling or logging for error recovery, the journal or log associated with a command or commands is saved in a persistent storage. In the event of a power fail or system crash or failure the journal or log is played back to restore the system to a known state.
  • As mentioned before, in some prior art techniques, a battery-backed volatile memory is utilized to maintain the contents of volatile memory for an extended period of time until power is back and tables can be saved in flash memory.
  • Battery backup solutions for saving system management data or cached user data during unplanned shutdowns are long-established but have certain disadvantage including up-front costs, replacement costs, service calls, disposal costs, system space limitations, reliability and “green” content requirements.
  • What is needed is a system employing physically-addressed SSD to reliably and efficiently preserve flash block management tables in the event of a power interruption.
  • SUMMARY OF THE INVENTION
  • Briefly, in accordance with one embodiment of the invention, a storage system includes a Central Processing Unit (CPU) that has a physically-addressed solid state disk (SSD) and is addressable using physical addresses associated with user data that are provided by a host. The user data is to be stored in or retrieved from the physically-addressed SSD in blocks. Further, a non-volatile memory module is coupled to the CPU and includes flash tables used to manage blocks in the physically addressed SSD. The flash tables have tables that are used to map logical to physical blocks for identifying the location of the user data in the physically-addressed SSD. The flash tables are maintained in the non-volatile memory modules.
  • Further, a non-volatile memory module is coupled to the CPU and used for storing flash block management tables (also referred to herein as “flash tables”), caching data, and maintaining a journal (the terms “journal” and “log” are used interchangeably herein), in addition to flash tables used to manage blocks in the physically-addressed SSD. The flash tables have tables that are used to map logical to physical blocks for identifying the location of stored data in the physically-addressed SSD. The flash tables are maintained in the non-volatile memory modules thereby avoiding reconstruction of the flash tables upon power interruption.
  • These and other features of the invention will no doubt become apparent to those skilled in the art after having read the following detailed description of the various embodiments illustrated in the several figures of the drawing.
  • IN THE DRAWINGS
  • FIG. 1 shows a storage system 700, in accordance with an embodiment of the invention.
  • FIG. 1 a shows exemplary contents of the system memory 762, the NV module 762, and the flash subsystem 110, in accordance with an embodiment of the invention.
  • FIG. 1 b shows exemplary contents of the system memory 746, the NV module 762, and the flash subsystem 110, in accordance with another embodiment of the invention.
  • FIG. 2 shows a storage system 810, in accordance with another embodiment of the invention.
  • FIG. 3 shows a storage system 866, in accordance with another embodiment of the invention.
  • DETAILED DESCRIPTION OF THE VARIOUS EMBODIMENTS
  • In an embodiment of the invention, a storage system includes a Central Processing Unit (CPU) a system memory, a network interface controller for connection to a network and one or more physically addressed SSD (paSSD), coupled to the CPU through a system bus such as Peripheral Component Interconnect Express (PCIe) bus and addressable using physical addresses for storing data provided via the network. The data is to be stored in or retrieved from the physically-addressed SSD in blocks.
  • Further, a non-volatile memory module is coupled to the CPU and used for storing flash block management tables (also referred to herein as “flash tables”), caching data, and maintaining a journal. The flash tables have tables that are used to map logical to physical blocks for identifying the location of stored data in the physically addressed SSD. The flash tables are maintained in the non-volatile memory modules thereby avoiding reconstruction of the flash tables upon power interruption.
  • In one embodiment, all flash block management tables are in one or more non-volatile memory module comprised of MRAM coupled to processor though memory channels.
  • In an alternate embodiment, tables are maintained in system memory and are near periodically saved in flash onboard the physically-addressed SSD and the parts of the tables that are updated since last saved are additionally maintained in a non-volatile memory module comprised of MRAM that is coupled to the processor though memory channels, the current version of the block management table that is in flash along with the updates that are saved in MRAM being used to reconstruct the flash block management tables in system memory upon system power-up.
  • In yet another alternate embodiment, in order to reduce the size of updates in MRAM and frequency of flash table copy back to flash, one or more of the updates (along with revision numbers), are also copied to flash. The current version of the block management table that is saved in flash, along with past updates, that are saved in flash and recent updates that are saved in MRAM are used to reconstruct the flash block management tables in system memory upon system power-up.
  • In yet another embodiment, the contents of MRAM are coupled to the processor through a system bus such as Serial Peripheral Interface (SPI) bus or PCIe with analogous methods of the invention used to reconstruct the flash block management tables in system memory upon system power-up, such as by either using the current version of the block management table in flash along with recent updates that are saved in MRAM or using the current version of the block management table in flash along with the past updates that are saved in flash. The recent updates that are saved in the MRAM are used to reconstruct the flash block management tables in the system memory upon power-up.
  • In yet another embodiment, the physically-addressable SSD includes a flash subsystem and a non-volatile memory comprised of MRAM. In some embodiment of the invention, flash block management tables are maintained in the system memory and are nearly periodically saved in the flash subsystem onboard the paSSD and the parts of the tables that are updated since the last save are additionally maintained in the MRAM that is onboard the paSSD via methods that physically address the MRAM rather than the flash.
  • In all the above embodiments, cached data and/or journals can optionally be stored in the MRAM.
  • Referring now to FIG. 1, a storage system 800 is shown, in accordance with an embodiment of the invention. The system 800 is shown to include a Central Processor Unit (CPU) 710 (also known herein as “processor 710”), a system memory 746, a non-volatile (NV) memory module 762, a basic input and output system (BIOS) 740, and a bank of physically-addressed solid state disks (SSD) 750-1 to 750-p, ‘p’ being an integer value, in accordance with an embodiment of the invention. A “bank”, as used herein, refers to one or more.
  • The CPU 710 of system 800 is shown to include a bank of CPU cores 712-1 through 712-n, ‘n’ being an integer value, a bank of memory controllers 724-1 through 724-m, ‘m’ being an integer value, shown coupled to a bank of memory channels 726-1 through 726-m, a PCIe controller 730. The CPU 710 is further shown to include an NV module controller 760, and a SPI controller 732. The network interface controller 802 is shown coupled through a PCIe bus 804 to PCIe controller 730 and CPU 710.
  • The bank of paSSDs 750-1 to 750-p is shown coupled to the CPU 710 through a respective one of a bank of sockets 737-1 to 737-p. The NV module 762 is shown coupled to the CPU 710 through a NV memory channel 764. The system memory 746 is shown coupled to include a bank of volatile RAM (DRAM) modules 747-1 through 747-m that are coupled to the memory controllers 724-1 through 724-m through a respective one of a bank of memory channels 726-1 to 726-m.
  • The PCIe controller 730 is shown coupled to a bank of PCIe busses 731-1 through 731-p that couple the CPU 710 to a bank of paSSD 750-1 through 750-p, Further shown in FIG. 1 is the CPU 710 including a (SPI) controller 732, which is shown coupled to the BIOS 740, the BIOS 740 shown residing externally to the CPU 710.
  • The NV module 762, which is also shown to reside externally to the CU 710, includes a bank of MRAMs 763-1 through 763-k that are shown coupled to the NV module controller 760 via the NV memory channel 764. In an embodiment of the invention, the NV memory channel 764 is analogous to the memory channels 726/728 and the NV module controller 760 is analogous to the memory controller 724.
  • The NV memory channel 764 couples the NV module 762 to the NV module controller 760 of the CPU 710. In an embodiment of the invention, the NV memory channel 764 is a DRAM memory channel.
  • In some embodiments, in addition to using the NV memory module 762 as flash tables, the NV memory module 762 is used by the system 800 as non-volatile cache for storing in-coming data (data that is input to the system 800) and or storing a journal.
  • Although a memory module is employed in the embodiments wherein one or memory devices are on a removable unit, other embodiments of employing memory devices in the system such as non-removable memory devices all within the scope of invention.
  • Network Interface Controller 802 is shown coupled to the processor 710 via the PCIe bus 804 and coupled to network interface 806 for connection to a network via network interface 806. The network interface controller 802 implements the circuitry required to communicate with a specific physical layer and the data link layer for receiving and transmitting information packets including command/status and data. The network interface controller 802 implements the circuitry required for upper layer protocols (the layer above the data link layer, such as transport layer, application layer, and the like).
  • In some embodiments, the network interface 806 is a Gigabit or ten Gigabit Ethernet running Internet Small Computer System Interface (iSCSI) and in other embodiments, it is a Serial Attached SCSI (SAS) or Fiber Channel (FC), which are generally used with block storage protocols. In yet other embodiments, the network interface 806 is Gigabit or ten Gigabit Ethernet running network file storage (NFS) protocol. All of the foregoing interfaces are known in the art. In particular, the Ethernet capabilities are either integrated into the CPU or implemented via a low-cost dedicated NIC 802, connected through the PCIe bus 804 as shown in FIG. 1.
  • In some embodiments, the flash subsystem 110 is made of flash NAND memory. In some embodiment, the flash subsystem 110 is made of flash NOR memory.
  • The CPU 710 of storage system 800 is shown to include one or more physically-addressed solid state disk 750-1 through 750-p, wherein the blocks are addressed with a physical rather than a logical address. The paSSD 750 includes flash subsystem 110. For example, the physically-addressed solid state disk 750-1 is shown to include the flash subsystem 110-1. In the storage system 800 of FIG. 1, flash block management is performed by a software driver (also known herein as the “driver”) 702 that is loaded during initialization of the system 800, after power-up. In addition to user commands, commands sent to the paSSD 750 include commands for flash management (including garbage collection, wear leveling, saving flash tables) and these commands use physical address rather than logical address.
  • In one embodiment of the invention, as shown in FIG. 1 a, the flash table 201 is saved in the non-volatile memory module 762 that is made of the MRAMs 763-1 thru 763-k of the embodiment of FIG. 1.
  • FIG. 1 a shows exemplary contents of the system memory 746, the NV module 762, and the flash subsystem 110, in accordance with an embodiment of the invention. The system memory 746 is shown to include a driver 702, the NV module 762 is shown to include the flash tables 201, journal 251, cache 261, and the flash subsystem 110 is shown to include the user data 366. The driver 702, shown saved in system memory 746, in FIG. 1 a, performs flash block management. The flash tables 201 are tables generally used for management of the flash memory blocks within the paSSD 750 of FIG. 1 and the user data 366 is generally information received by the paSSD 750 from the CPU 710 to be saved. The flash tables 201 include tables used for managing flash memory blocks. The driver 702 generally manages the flash memory blocks. As shown in FIG. 1 a, the flash table 201 is maintained in the module 762.
  • As noted above, the flash subsystem 110 is addressed using physical and not logical addresses, provided by the CPU 710.
  • In an alternate embodiment, the flash tables 201 are maintained in the system memory 762 and are substantially periodically saved in the flash subsystem 110 of the paSSD 750, and the parts of the tables 201 that are updated (modified) since the previous save are additionally saved in the non-volatile memory module 762.
  • FIG. 1 b shows exemplary contents of the system memory 746, the NV module 762, and the flash subsystem 110, in accordance with another embodiment of the invention. In FIG. 1 b, the system memory 746 is shown to include the driver 702 in addition to the flash tables 201. The NV module 762 is shown to include the table updates 302, journal 251, cache 261 and the flash subsystem 110 is shown to include table copies 360 and the user data 366. As previously noted, the flash tables 201 are tables that are generally used for management of blocks within the SSD 750. The table updates 302 are generally updates to the flash tables 201 since the last copy of the flash tables 201 was initiated until a subsequent copy is initiated. The table copies 360 are snapshots of the flash tables 201 that are saved in the flash subsystem 110. This is further explained in U.S. patent application Ser. No. 13/673,866, filed on Nov. 9, 2012, by Siamack Nemazie, and entitled “System Employing MRAM and Physically Addressed Solid State Disk”, the contents of which are incorporated herein by reference as though set forth in full. The user data 366 is information provided by the host (CPU 710 of FIG. 1).
  • In some embodiments, the NV module 762 includes spin torque transfer MRAM (STTMRAM).
  • In some embodiments, the NV module 762 is coupled to the CPU 710 via a system bus. An exemplary system bus is Serial Protocol Interconnect (SPI) or PCIe.
  • Accordingly, storage system 800 and the flash tables 201 are used to manage blocks in the paSSD 750. The flash tables 201 include tables that are used to map logical blocks to physical blocks for identifying the location of stored data in the paSSD 750 and the flash tables are maintained in the NV module 762, which advantageously avoids reconstruction of the flash tables upon power interruption of the storage system 800.
  • Additionally, a cache or in particular a write back cache is maintained in the non-volatile memory module 762. The write cache contents are written to the paSSD 750 upon triggers, such trigger including, among others, the cached data in the cache 261 to reach a certain limit. In one embodiment, the contents of the cache 261 are written to the psSSDs in units of pages that are each aligned to a page boundary. In one embodiment, partial pages are maintained in the non-volatile memory module 762 and coalesced with other writes to write a full page to SSDs. In another embodiment, the partial page in the non-volatile memory module 762 is merged with the unmodified portion of the page in SSD, so that the page in the cache reflects the state of the page.
  • Journaling is generally a known practice for logging the changes that will be made in a journal (usually a circular log in a dedicated area) before committing them to the main storage. In the event of a system crash or power failure, such file systems are quicker to bring back online and less likely to become corrupted. A “journal” maybe optionally maintained in the non-volatile memory module 762. In the exemplary embodiments the NV module 762 is shown to include a journal 251
  • FIG. 2 shows a storage system 810, in accordance with another embodiment of the invention. The system 810 is analogous to the system 800 except that the system 810 additionally includes a PCIe-to-Flash bus Bridge controller 735 to couple a plurality of PCIe buses 731-1 through 731- to a bank of flash buses 733-I (where i is from 1 through q) each coupled to a paSSD 750-i through connector 737-I, ‘i’ being an integer value. An example of a flash bus is Open NAND Flash Interface (ONFI).
  • Except for a different interface to the paSSDs, the operation of system 810 is analogous to system 800.
  • FIG. 3 shows a storage system 865, in accordance with another embodiment of the invention. The system 865 is analogous to the system 800 except that the system 865 does not employ an NV module, and the paSSDs 850-1 through 850-p employ MRAM specifically MRAMs 140/150-1 through 140/150-p respectively. The operation of paSSDs 850-1 through 850-p and its operation in a system is further explained in U.S. patent application Ser. No. 13/745,686, filed on Jan. 18, 2013, by Siamack Nemazie, and entitled “Physically Addressed Solid State Disk Employing MRAM”, the contents of which are incorporated herein by reference as though set forth in full.
  • Except for a removing NV module 762 and employing a paSSD employing MRAM the structure of system 865 is analogous to system 800.
  • Although the invention has been described in terms of specific embodiments, it is anticipated that alterations and modifications thereof will no doubt become apparent to those skilled in the art. It is therefore intended that the following claims be interpreted as covering all such alterations and modification as fall within the true spirit and scope of the invention.

Claims (27)

What is claimed is:
1. A storage system comprising:
a Central Processing Unit (CPU):
at least one physically-addressed solid state disk (SSD) that is addressable using physical addresses associated with user data, the at least one physically-addressed SSD being provided by a host, the user data to be stored in or retrieved from the at least one physically-addressed SSD in blocks;
a non-volatile memory module coupled to the CPU and configured to store changes to tables, the tables used to map logical to physical block addresses for identifying the location of stored data in the at least one physically-addressed SSD, wherein the tables are moved to the at least one physically-addressed SSDs and the changes to the tables are maintained in the non-volatile memory module; and
a network interface controller coupled through a bus to the CPU.
2. The storage system of claim 1, wherein upon the non-volatile memory module having associated therewith an available capacity and upon the available capacity reaching a desired available capacity, the tables being moved to the memory.
3. The storage system, as recited in claim 1, wherein the at least one physically-addressed solid state disk (SSD) includes a subsystem and the subsystem being made of flash NAND memory.
4. The storage system, as recited in claim 1, further including a bank of memory controllers coupled to a bank of volatile RAM modules through a respective one of a bank of memory channels.
5. The storage system, as recited in claim 1, wherein the at least one physically-addressable SSD includes a subsystem.
6. The storage system, as recited in claim 5, wherein the subsystem is made of flash NAND memory.
7. The storage system, as recited in claim 1, wherein the at least one physically-addressed SSD is coupled through another bus to the CPU.
8. The storage system, as recited in claim 7, wherein the bus and the another bus are the same bus.
9. The storage system, as recited in claim 7, wherein the bus and the another bus are different busses.
10. The storage system, as recited in claim 7, wherein the bus is a PCIe bus.
11. The storage system, as recited in claim 7, wherein the another bus is a PCIe bus.
12. The storage system, as recited in claim 7, further including a bridge controller configured to couple the bus and the another bus.
13. The storage system, as recited in claim 7, wherein the bus and the another bus are PCIe busses.
14. The storage system, as recited in claim 1, wherein the non-volatile memory module includes magnetic random access memory (MRAM).
15. The storage system, as recited in claim 1, wherein the non-volatile memory module includes spin transfer torque magnetic random access memory (STTMRAM).
16. The storage system, as recited in claim 1, wherein the non-volatile memory module is coupled to the CPU through a memory channel.
17. The storage system, as recited in claim 16, wherein the non-volatile memory module is coupled to a CPU through a dynamic random access memory (DRAM) memory channel.
18. The storage system, as recited in claim 1, wherein the non-volatile memory module is coupled to the CPU via a system bus.
19. The storage system, as recited in claim 1, further including a system memory configured to store a software driver configured to manage the flash tables.
20. The storage system, as recited in claim 1, wherein the non-volatile memory module includes a cache configured to store the user data.
21. The storage system, as recited in claim 1, wherein the non-volatile memory module is configured to store a journal and the journal is configured to log the changes before the changes are made to the non-volatile memory module.
22. The storage system, as recited in claim 1, further including another bus coupling the at least one physically-addressed SSD to the CPU.
23. The storage system, as recited in claim 22, wherein the another bus is a PCIe bus.
24. The storage system, as recited in claim 23, wherein the CPU includes a controller.
25. The storage system, as recited in claim 24, wherein the controller is coupled to the another bus.
26. The storage system, as recited in claim 1, wherein the CPU comprises a plurality of CPU cores.
27. The storage system, as recited in claim 1, further including a bridge controller configured to couple the at least one physically-addressed SSD to the CPU.
US14/697,544 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same Abandoned US20150248348A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/697,544 US20150248348A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US201161538697P 2011-09-23 2011-09-23
US13/570,202 US20130080687A1 (en) 2011-09-23 2012-08-08 Solid state disk employing flash and magnetic random access memory (mram)
US13/673,866 US20140047161A1 (en) 2012-08-08 2012-11-09 System Employing MRAM and Physically Addressed Solid State Disk
US13/745,686 US9009396B2 (en) 2011-09-23 2013-01-18 Physically addressed solid state disk employing magnetic random access memory (MRAM)
US13/769,710 US8909855B2 (en) 2012-08-08 2013-02-18 Storage system employing MRAM and physically addressed solid state disk
US14/542,516 US9037787B2 (en) 2011-09-23 2014-11-14 Computer system with physically-addressable solid state disk (SSD) and a method of addressing the same
US14/697,544 US20150248348A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/542,516 Continuation US9037787B2 (en) 2011-09-23 2014-11-14 Computer system with physically-addressable solid state disk (SSD) and a method of addressing the same

Publications (1)

Publication Number Publication Date
US20150248348A1 true US20150248348A1 (en) 2015-09-03

Family

ID=50067077

Family Applications (5)

Application Number Title Priority Date Filing Date
US13/769,710 Active US8909855B2 (en) 2011-09-23 2013-02-18 Storage system employing MRAM and physically addressed solid state disk
US14/542,516 Active US9037787B2 (en) 2011-09-23 2014-11-14 Computer system with physically-addressable solid state disk (SSD) and a method of addressing the same
US14/697,546 Abandoned US20150248349A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same
US14/697,544 Abandoned US20150248348A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same
US14/697,538 Abandoned US20150248346A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US13/769,710 Active US8909855B2 (en) 2011-09-23 2013-02-18 Storage system employing MRAM and physically addressed solid state disk
US14/542,516 Active US9037787B2 (en) 2011-09-23 2014-11-14 Computer system with physically-addressable solid state disk (SSD) and a method of addressing the same
US14/697,546 Abandoned US20150248349A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/697,538 Abandoned US20150248346A1 (en) 2011-09-23 2015-04-27 Physically-addressable solid state disk (ssd) and a method of addressing the same

Country Status (1)

Country Link
US (5) US8909855B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108762670A (en) * 2018-05-22 2018-11-06 郑州云海信息技术有限公司 Management method, system and the device of data block in a kind of SSD firmwares

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5404798B2 (en) * 2009-09-21 2014-02-05 株式会社東芝 Virtual storage management device and storage management device
ITRM20120094A1 (en) * 2012-03-14 2013-09-14 Istituto Naz Di Fisica Nuclea Re NETWORK INTERFACE CARD FOR PARALLEL CALCULATION NETWORK KNOT ON GPU, AND RELATIVE INTERNODAL COMMUNICATION METHOD
US9443584B2 (en) * 2012-12-17 2016-09-13 Adesto Technologies Corporation Network interface with logging
JP6130800B2 (en) * 2014-03-06 2017-05-17 株式会社日立製作所 Computer equipment
US9996303B2 (en) * 2014-05-08 2018-06-12 Kabushiki Kaisha Toshiba Hybrid-HDD that gives up old NAND data at the last moment
CN105701021B (en) * 2014-12-10 2021-03-02 慧荣科技股份有限公司 Data storage device and data writing method thereof
CN104579880A (en) * 2014-12-25 2015-04-29 西安航天恒星科技实业(集团)公司 Data transmitting, processing and distributing system and method based on PCIE
US20160378352A1 (en) * 2015-06-26 2016-12-29 Intel Corporation Efficient solid state drive data compression scheme and layout
US10114550B2 (en) 2016-01-07 2018-10-30 Samsung Electronics Co., Ltd. Data storage device and data processing system including the data storage device
US10452290B2 (en) 2016-12-19 2019-10-22 Pure Storage, Inc. Block consolidation in a direct-mapped flash storage system
US11550481B2 (en) 2016-12-19 2023-01-10 Pure Storage, Inc. Efficiently writing data in a zoned drive storage system
US10635654B2 (en) 2017-06-12 2020-04-28 Samsung Electronics Co., Ltd. Data journaling for large solid state storage devices with low DRAM/SRAM
US10521343B2 (en) 2017-06-12 2019-12-31 Western Digital Technologies, Inc. Storage system with a controller having a persistent memory interface to local memory

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080189485A1 (en) * 2007-02-01 2008-08-07 Samsung Electronics Co., Ltd. Cooperative memory management

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100191896A1 (en) * 2009-01-23 2010-07-29 Magic Technologies, Inc. Solid state drive controller with fast NVRAM buffer and non-volatile tables
US20140047159A1 (en) * 2012-08-10 2014-02-13 Sandisk Technologies Inc. Enterprise server with flash storage modules

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080189485A1 (en) * 2007-02-01 2008-08-07 Samsung Electronics Co., Ltd. Cooperative memory management

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Gary Orenstein, “Optimizing I/O Operations via the Flash Translation Layer", August 8th, 2011, pages 1-38. *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108762670A (en) * 2018-05-22 2018-11-06 郑州云海信息技术有限公司 Management method, system and the device of data block in a kind of SSD firmwares

Also Published As

Publication number Publication date
US20150248349A1 (en) 2015-09-03
US9037787B2 (en) 2015-05-19
US20140047165A1 (en) 2014-02-13
US20150248346A1 (en) 2015-09-03
US20150089123A1 (en) 2015-03-26
US8909855B2 (en) 2014-12-09

Similar Documents

Publication Publication Date Title
US9037787B2 (en) Computer system with physically-addressable solid state disk (SSD) and a method of addressing the same
US10037272B2 (en) Storage system employing MRAM and array of solid state disks with integrated switch
US9009396B2 (en) Physically addressed solid state disk employing magnetic random access memory (MRAM)
KR102168838B1 (en) Mapping tables for storage devices
US9983993B2 (en) Apparatus, system, and method for conditional and atomic storage operations
US8725946B2 (en) Mass storage system and method of using hard disk, solid-state media, PCIe edge connector, and raid controller
US20180182459A1 (en) Solid state drive architectures
CN109643275B (en) Wear leveling apparatus and method for storage class memory
US9251086B2 (en) Apparatus, system, and method for managing a cache
US8756375B2 (en) Non-volatile cache
US9342256B2 (en) Epoch based storage management for a storage device
US10042758B2 (en) High availability storage appliance
US20160224588A1 (en) Data integrity and loss resistance in high performance and high capacity storage deduplication
US20130080687A1 (en) Solid state disk employing flash and magnetic random access memory (mram)
US20190324859A1 (en) Method and Apparatus for Restoring Data after Power Failure for An Open-Channel Solid State Drive
WO2012050934A2 (en) Apparatus, system, and method for a direct interface between a memory controller and non-volatile memory using a command protocol
US10459803B2 (en) Method for management tables recovery
US9798673B2 (en) Paging enablement of storage translation metadata
US20140047161A1 (en) System Employing MRAM and Physically Addressed Solid State Disk
US11640371B2 (en) Snapshot management in partitioned storage
US11966590B2 (en) Persistent memory with cache coherent interconnect interface
US20230236737A1 (en) Storage Controller Managing Different Types Of Blocks, Operating Method Thereof, And Operating Method Of Storage Device Including The Same
CN116662217A (en) Persistent memory device and method for applying same

Legal Events

Date Code Title Description
AS Assignment

Owner name: NEMAZIE, SIAMACK, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NEMAZIE, SIAMACK;REEL/FRAME:037605/0920

Effective date: 20160128

Owner name: ABEDIFARD, EBRAHIM, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ABEDIFARD, EBRAHIM;REEL/FRAME:037605/0769

Effective date: 20160128

Owner name: AVALANCHE TECHNOLOGY, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NEMAZIE, SIAMACK;REEL/FRAME:037607/0499

Effective date: 20160127

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: SILICON VALLEY BANK, CALIFORNIA

Free format text: SECURITY INTEREST;ASSIGNOR:AVALANCHE TECHNOLOGY, INC.;REEL/FRAME:053156/0223

Effective date: 20200212