US20140195551A1 - Optimizing snapshot lookups - Google Patents

Optimizing snapshot lookups Download PDF

Info

Publication number
US20140195551A1
US20140195551A1 US14/058,757 US201314058757A US2014195551A1 US 20140195551 A1 US20140195551 A1 US 20140195551A1 US 201314058757 A US201314058757 A US 201314058757A US 2014195551 A1 US2014195551 A1 US 2014195551A1
Authority
US
United States
Prior art keywords
medium
range
entries
mapping table
underlying
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/058,757
Inventor
John Colgrove
Ethan MILLER
John Hayes
Cary Sandvig
Christopher Golden
Jianting Cao
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pure Storage Inc
Original Assignee
Pure Storage Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pure Storage Inc filed Critical Pure Storage Inc
Priority to US14/058,757 priority Critical patent/US20140195551A1/en
Assigned to PURE STORAGE, INC. reassignment PURE STORAGE, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GOLDEN, CHRISTOPHER, HAYES, JOHN, CAO, JIANTING, COLGROVE, JOHN, MILLER, ETHAN, SANDVIG, Cary
Priority to PCT/US2014/010885 priority patent/WO2014110266A1/en
Publication of US20140195551A1 publication Critical patent/US20140195551A1/en
Priority to US15/291,373 priority patent/US10013317B1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0646Horizontal data movement in storage systems, i.e. moving data in between storage devices or systems
    • G06F3/065Replication mechanisms
    • G06F17/30345
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/17Details of further file system functions
    • G06F16/174Redundancy elimination performed by the file system
    • G06F16/1748De-duplication implemented within the file system, e.g. based on file segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/23Updating
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0604Improving or facilitating administration, e.g. storage management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0604Improving or facilitating administration, e.g. storage management
    • G06F3/0605Improving or facilitating administration, e.g. storage management by facilitating the interaction with a user or administrator
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0608Saving storage space on storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/061Improving I/O performance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/061Improving I/O performance
    • G06F3/0611Improving I/O performance in relation to response time
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • G06F3/0619Improving the reliability of storage systems in relation to data integrity, e.g. data losses, bit errors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0626Reducing size or complexity of storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0629Configuration or reconfiguration of storage systems
    • G06F3/0632Configuration or reconfiguration of storage systems by initialisation or re-initialisation of storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0638Organizing or formatting or addressing of data
    • G06F3/064Management of blocks
    • G06F3/0641De-duplication techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0638Organizing or formatting or addressing of data
    • G06F3/0644Management of space entities, e.g. partitions, extents, pools
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0646Horizontal data movement in storage systems, i.e. moving data in between storage devices or systems
    • G06F3/0652Erasing, e.g. deleting, data cleaning, moving of data to a wastebasket
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0655Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0662Virtualisation aspects
    • G06F3/0664Virtualisation aspects at device level, e.g. emulation of a storage device or system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0662Virtualisation aspects
    • G06F3/0665Virtualisation aspects at area level, e.g. provisioning of virtual or logical volumes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0662Virtualisation aspects
    • G06F3/0667Virtualisation aspects at data level, e.g. file, record or object virtualisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/067Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • G06F3/0685Hybrid storage combining heterogeneous device types, e.g. hierarchical storage, hybrid arrays
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • G06F3/0689Disk arrays, e.g. RAID, JBOD
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1008Correctness of operation, e.g. memory ordering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1016Performance improvement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1032Reliability improvement, data loss prevention, degraded operation etc
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1041Resource optimization
    • G06F2212/1044Space efficiency improvement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/15Use in a specific computing environment
    • G06F2212/154Networked environment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/16General purpose computing application
    • G06F2212/163Server or database system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/26Using a specific storage system architecture
    • G06F2212/261Storage comprising a plurality of storage devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/26Using a specific storage system architecture
    • G06F2212/263Network storage, e.g. SAN or NAS

Definitions

  • This invention relates to optimizing snapshot lookup operations in a storage system.
  • a datacenter which also may be referred to as a server room, is a centralized repository, either physical or virtual, for the storage, management, and dissemination of data pertaining to one or more businesses.
  • a distributed storage system may be coupled to client computers interconnected by one or more networks. If any portion of the distributed storage system has poor performance, company operations may be impaired. A distributed storage system therefore maintains high standards for data availability and high-performance functionality.
  • snapshots may be utilized to capture and store data at a particular point in time.
  • a snapshot may be taken of a logical volume, and the snapshot may be stored to preserve the contents of the volume. If the data associated with the volume is later lost or corrupted, the volume can be restored from the snapshot.
  • the storage utilization and processing overhead of the storage system likewise increases.
  • the proliferation of snapshots in a storage system results in more processing overhead for the storage system as well as increasing numbers of lookups through multiple snapshots for some data requests.
  • the storage system may be required to manage a large number of snapshots, some of which may no longer be externally visible by clients.
  • a storage system may include a storage controller and one or more storage devices.
  • the storage system may be coupled to one or more host client systems.
  • the storage controller may utilize volumes and mediums to track client data that is stored in the storage system.
  • a medium is defined as a logical grouping of data, and each medium has an identifier with which to identify the logical grouping of data.
  • the storage controller may maintain a volume to medium mapping table to map each volume to a single medium, and this medium is referred to as the volume's anchor medium.
  • Each medium may be mapped to any number of other mediums, and the storage controller may also maintain a medium mapping table to track relationships between anchor mediums and underlying mediums.
  • the storage controller may also maintain an address translation table and corresponding indexes for mapping mediums and block numbers to physical storage locations.
  • the storage controller may be configured to perform medium lookup optimizing operations on a periodic basis.
  • the storage controller may search through all of the mediums in the system to find mediums that can be reduced or collapsed to make future searches of the system more efficient.
  • the storage controller may look for active mediums in a read-write state which have one or more levels of underlying mediums beneath them, with each underlying medium representing a snapshot. These active mediums with one or more underlying mediums may be collapsed by the storage controller.
  • the storage controller may not make changes that cause the result of a lookup of a block on a volume to change, but may make changes in mediums that are invisible from the point of view of all volumes on the storage controller.
  • the storage controller may attempt to collapse the medium mapping table so as to reduce the number of mediums in which the storage controller will have to look to find data blocks for a given data request.
  • the storage controller may collapse the medium mapping table by promoting individual mappings from lower-level mediums to higher-level mediums.
  • the storage controller may also collapse the medium mapping table by pushing medium range pointers higher up in the table so as to reduce the number of mediums that need to be searched.
  • the storage controller may promote individual mappings from existing entries in the address translation table into new entries for the destination medium. For a given medium with multiple levels of underlying mediums, all of the mappings for these levels of underlying mediums may be merged, resulting in a new version of the given medium. This process does not change the result of any individual mapping but rather improves the performance of the storage system by allowing the given medium to be converted into a quiesced medium that refers to itself. After receiving a data request targeting the given medium, the storage controller will have access to all of its mappings by performing a single lookup to the address translation table indexes for only the given medium.
  • another technique may be utilized by the storage controller to reduce the number of lookups required by shortcutting the medium mapping table.
  • the system may collapse ranges of masked mediums together.
  • a range of a medium is considered unmasked if one or more address translation table entries exist in the range for that medium, which indicates that the range has been previously written to.
  • a range of a medium is considered masked if there are no entries in the range for that medium. If a first medium has multiple levels of underlying mediums below it, and its immediate underlying medium is masked, then a shortcut through the medium mapping table may be created, bypassing the medium directly below the first medium. In this case, a new entry may be created and entered into the medium mapping table, with the new entry mapping the first medium directly to a medium at least two levels below the first medium in the medium graph.
  • another technique for collapsing the medium graph may be utilized by the storage controller. This technique may involve merging one or more mediums that cannot be referenced externally into the medium directly above them in the medium graph. If a second medium underlies a first medium and the second medium is no longer externally visible, the storage controller may merge the second medium with the first medium, combining the entries of the second medium and the first medium and renumbering the entries.
  • FIG. 1 is a generalized block diagram illustrating one embodiment of a storage system.
  • FIG. 2 is a generalized block diagram of one embodiment of a directed acyclic graph (DAG) of mediums.
  • DAG directed acyclic graph
  • FIG. 3 illustrates one embodiment of a medium mapping table.
  • FIG. 4 illustrates a block diagram of one embodiment of an address translation table.
  • FIG. 5 is a generalized block diagram of one embodiment of a search optimization operation.
  • FIG. 6 illustrates one embodiment of a portion of a medium mapping table.
  • FIG. 7 is a generalized flow diagram illustrating one embodiment of a method for shortcutting a medium mapping table.
  • FIG. 8 is a generalized flow diagram illustrating one embodiment of a method for performing medium graph collapsing operations.
  • Storage system 100 may include storage controller 110 and storage device groups 130 and 140 , which are representative of any number of storage device groups (or data storage arrays).
  • storage device group 130 includes storage devices 135 A-N, which are representative of any number and type of storage devices (e.g., solid-state drives (SSDs)).
  • Storage controller 110 may be coupled directly to client computer system 125 , and storage controller 110 may be coupled remotely over network 120 to client computer system 115 .
  • Clients 115 and 125 are representative of any number of clients which may utilize storage controller 110 for storing and accessing data in system 100 . It is noted that some systems may include only a single client, connected directly or remotely to storage controller 110 .
  • Storage controller 110 may include software and/or hardware configured to provide access to storage devices 135 A-N. Although storage controller 110 is shown as being separate from storage device groups 130 and 140 , in some embodiments, storage controller 110 may be located within one or each of storage device groups 130 and 140 . Storage controller 110 may include or be coupled to a base operating system (OS), a volume manager, and additional control logic for implementing the various techniques disclosed herein.
  • OS base operating system
  • volume manager volume manager
  • Storage controller 110 may include and/or execute on any number of processors and may include and/or execute on a single host computing device or be spread across multiple host computing devices, depending on the embodiment. In some embodiments, storage controller 110 may generally include or execute on one or more file servers and/or block servers. Storage controller 110 may use any of various techniques for replicating data across devices 135 A-N to prevent loss of data due to the failure of a device or the failure of storage locations within a device. Storage controller 110 may also utilize any of various deduplication techniques for reducing the amount of data stored in devices 135 A-N by deduplicating common data.
  • Storage controller 110 may also be configured to create and manage snapshots in system 100 .
  • a set of mediums may be recorded and maintained by storage controller 110 .
  • Most of the mediums may be read-only except for one or more selected mediums such as the most recent medium in use by a particular volume.
  • Each medium logically comprises all of the blocks in the medium. However, only the blocks that were changed from the time the medium was created to the time the medium was closed are saved and mappings to these blocks may also be maintained with the medium.
  • mapping tables may be maintained by storage controller 110 .
  • These mapping tables may include a medium mapping table and a volume-to-medium mapping table. These tables may be utilized to record and maintain the mappings between mediums and underlying mediums and the mappings between volumes and mediums.
  • Storage controller 110 may also include an address translation table with a plurality of entries, wherein each entry holds a virtual-to-physical mapping for a corresponding data component.
  • This mapping table may be used to map logical read/write requests from each of the client computer systems 115 and 125 to physical locations in storage devices 135 A-N.
  • a “physical” pointer value may be read from the mappings associated with a given medium during a lookup operation corresponding to a received read/write request.
  • This physical pointer value may then be used to locate a physical location within the storage devices 135 A-N. It is noted the physical pointer value may be used to access another mapping table within a given storage device of the storage devices 135 A-N. Consequently, one or more levels of indirection may exist between the physical pointer value and a target storage location.
  • the address translation table may be accessed using a key comprising a medium ID, a logical or virtual address, a sector number, and so forth.
  • a received read/write storage access request may identify a particular volume, sector, and length.
  • the volume ID may be mapped to a medium ID using the volume to medium mapping table.
  • a sector may be a logical block of data stored in a medium. Sectors may have different sizes on different mediums.
  • the address translation table may map a medium in sector-size units. In other embodiments, other types of address translation tables may be utilized.
  • the address translation table may map mediums to physical pointer values. For a given storage access request, an access may be made to the “mappings” of the address translation table.
  • the term “mappings” is defined as the one or more entries of the address translation table which convert a given medium ID and block number into a physical pointer value.
  • a physical pointer value may be a physical address or a logical address which the storage device maps to a physical location within the device.
  • an index may be utilized to access the address translation table. The index may identify locations of mappings within the address translation table.
  • the index may be queried with a key value generated from a medium ID and sector number, and the index may be searched for one or more entries which match, or otherwise correspond to, the key value. Information from a matching entry may then be used to locate and retrieve a mapping which identifies a storage location which is the target of a received read or write request.
  • a hit in the index provides a corresponding virtual page ID identifying a page within the storage devices of the storage system, wherein the page stores both the key value and a corresponding physical pointer value. The page may then be searched with the key value to find the physical pointer value.
  • the number and type of client computers, storage controllers, networks, storage device groups, and data storage devices is not limited to those shown in FIG. 1 . At various times one or more clients may operate offline. In addition, during operation, individual client computer connection types may change as users connect, disconnect, and reconnect to system 100 . Further, the systems and methods described herein may be applied to directly attached storage systems or network attached storage systems and may include a host operating system configured to perform one or more aspects of the described methods. Numerous such alternatives are possible and are contemplated.
  • Network 120 may utilize a variety of techniques including wireless connection, direct local area network (LAN) connections, wide area network (WAN) connections such as the Internet, a router, storage area network, Ethernet, and others.
  • Network 120 may comprise one or more LANs that may also be wireless.
  • Network 120 may further include remote direct memory access (RDMA) hardware and/or software, transmission control protocol/internet protocol (TCP/IP) hardware and/or software, router, repeaters, switches, grids, and/or others. Protocols such as Fibre Channel, Fibre Channel over Ethernet (FCoE), iSCSI, and so forth may be used in network 120 .
  • the network 120 may interface with a set of communications protocols used for the Internet such as the Transmission Control Protocol (TCP) and the Internet Protocol (IP), or TCP/IP.
  • TCP Transmission Control Protocol
  • IP Internet Protocol
  • Client computer systems 115 and 125 are representative of any number of stationary or mobile computers such as desktop personal computers (PCs), servers, server farms, workstations, laptops, handheld computers, servers, personal digital assistants (PDAs), smart phones, and so forth.
  • client computer systems 115 and 125 include one or more processors comprising one or more processor cores.
  • Each processor core includes circuitry for executing instructions according to a predefined general-purpose instruction set. For example, the x86 instruction set architecture may be selected. Alternatively, the ARM®, Alpha®, PowerPC®, SPARC®, or any other general-purpose instruction set architecture may be selected.
  • the processor cores may access cache memory subsystems for data and computer program instructions.
  • the cache subsystems may be coupled to a memory hierarchy comprising random access memory (RAM) and a storage device.
  • RAM random access memory
  • FIG. 2 a block diagram illustrating a directed acyclic graph (DAG) 200 of mediums is shown. Also shown is a volume to medium mapping table 205 which shows which medium a volume maps to for each volume in use by a storage system. Volumes may be considered pointers into graph 200 .
  • DAG directed acyclic graph
  • medium is defined as a logical grouping of data.
  • a medium may have a corresponding identifier with which to identify the logical grouping of data.
  • Each medium may also include or be associated with mappings of logical block numbers to content location, deduplication entries, and other information.
  • medium identifiers may be used by the storage controller but medium identifiers may not be user-visible.
  • a user or client may send a data request accompanied by a volume ID to specify which data is targeted by the request, and the storage controller may map the volume ID to a medium ID and then use the medium ID when processing the request.
  • a storage medium is defined as an actual physical device (e.g., SSD, HDD) that is utilized to store data.
  • a computer readable storage medium (or non-transitory computer readable storage medium) is defined as a physical storage medium configured to store program instructions which are executable by a processor or other hardware device.
  • program instructions that implement the methods and/or mechanisms described herein may be conveyed or stored on a computer readable medium.
  • volume to medium mapping table may refer to multiple tables rather than just a single table.
  • intermediate mapping table may also refer to multiple tables rather than just a single table.
  • volume to medium mapping table 205 is only one example of a volume to medium mapping table. Other volume to medium mapping tables may have other numbers of entries for other numbers of volumes.
  • Each medium is depicted in graph 200 as three conjoined boxes, with the leftmost box showing the medium ID, the middle box showing the underlying medium, and the rightmost box displaying the status of the medium (RO—read-only) or (RW—read-write).
  • Read-write mediums may be referred to as active mediums, while read-only mediums may represent previously taken snapshots.
  • a medium points to its underlying medium.
  • medium 20 points to medium 12 to depict that medium 12 is the underlying medium of medium 20.
  • Medium 12 also points to medium 10, which in turn points to medium 5, which in turn points to medium 1.
  • Some mediums are the underlying medium for more than one higher-level medium.
  • three separate mediums (12, 17, 11) point to medium 10
  • two separate mediums (18, 10) point to medium 5
  • two separate mediums (6, 5) point to medium 1.
  • Each of the mediums which is an underlying medium to at least one higher-level medium has a status of read-only.
  • the set of mediums on the bottom left of graph 200 is an example of a linear set.
  • medium 3 was created first and then a snapshot was taken resulting in medium 3 becoming stable (i.e., the result of a lookup for a given block in medium 3 will always return the same value after this point).
  • Medium 7 was created with medium 3 as its underlying medium. Any blocks written after medium 3 became stable were labeled as being in medium 7. Lookups to medium 7 return the value from medium 7 if one is found, but will look in medium 3 if a block is not found in medium 7. At a later time, a snapshot of medium 7 is taken, medium 7 becomes stable, and medium 14 is created.
  • Volume to medium mapping table 205 maps user-visible volumes to mediums. Each volume may be mapped to a single medium, also known as the anchor medium. This anchor medium, as with all other mediums, may take care of its own lookups. A medium on which multiple volumes depend (such as medium 10) tracks its own blocks independently of the volumes which depend on it. Each medium may also be broken up into ranges of blocks, and each range may be treated separately in medium DAG 200 .
  • a medium mapping table 300 is shown. Any portion of or the entirety of medium mapping table 300 may be stored in storage controller 110 and/or in one or more of storage devices 135 A-N.
  • a volume identifier (ID) may be used to access volume to medium mapping table 205 to determine a medium ID corresponding to the volume ID. This medium ID may then be used to access medium mapping table 300 .
  • table 300 is merely one example of a medium mapping table, and that in other embodiments, other medium mapping tables, with other numbers of entries, may be utilized.
  • a medium mapping table may include other attributes and be organized in a different manner than that shown in FIG. 3 .
  • any suitable data structure may be used to store the mapping table information in order to provide for efficient searches (e.g., b-trees, binary trees, hash tables, etc.). All such data structures are contemplated.
  • Each medium may be identified by a medium ID, as shown in the leftmost column of table 300 .
  • a range attribute may also be included in each entry of table 300 , and the range may be in terms of data blocks. The size of a block of data (e.g., 4 KB, 8 KB) may vary depending on the embodiment.
  • a medium may be broken up into multiple ranges, and each range of a medium may be treated as if it is an independent medium with its own attributes and mappings.
  • medium ID 2 has two separate ranges. Range 0-99 of medium ID 2 has a separate entry in table 300 from the entry for range 100-999 of medium ID 2.
  • both of these ranges of medium ID 2 map to underlying medium ID 1
  • separate ranges from medium ID 35 map to separate underlying mediums.
  • range 0-299 of medium ID 35 maps to underlying medium ID 18 with an offset of 400. This indicates that blocks 0-299 of medium ID 35 map to blocks 400-699 of medium ID 18.
  • range 300-499 of medium ID 35 maps to underlying medium ID 33 with an offset of ⁇ 300
  • range 500-899 of medium ID 35 maps to underlying medium ID 5 with an offset of ⁇ 400.
  • These entries indicate that blocks 300-499 of medium ID 35 map to blocks 0-199 of medium ID 33 while blocks 500-899 of medium ID 35 map to blocks 100-499 of medium ID 5. It is noted that in other embodiments, mediums may be broken up into more than three ranges.
  • the state column of table 300 records information that allows lookups for blocks to be performed more efficiently.
  • a state of “Q” indicates the medium is quiescent, “R” indicates the medium is registered, and “U” indicates the medium is unmasked.
  • a lookup is performed on exactly one or two mediums specified in table 300 .
  • the registered state a lookup is performed recursively. If an entry in table 300 is unmasked, then this indicates that a lookup should be performed in the basis medium. If an entry is masked, then the lookup should only be performed in the underlying medium.
  • another state “X” may be used to specify that the source medium is unmapped.
  • the unmapped state indicates that the source medium contains no reachable data and can be discarded. This unmapped state may apply to a range of a source medium. If an entire medium is unmapped, then the medium ID may be entered into a sequence invalidation table and eventually discarded.
  • the medium when a medium is created, the medium is in the registered state if it has an underlying medium, or the medium is in the quiescent state if it is a brand-new volume with no pre-existing state. As the medium is written to, parts of it can become unmasked, with mappings existing both in the medium itself and the underlying medium. This may be done by splitting a single range into multiple range entries, some of which retain the original masked status, and others of which are marked as unmasked.
  • each entry in table 300 may include a basis attribute, which indicates the basis of the medium, which in this case points to the source medium itself.
  • Each entry may also include an offset field, which specifies the offset that should be applied to the block address when mapping the source medium to an underlying medium. This allows mediums to map to other locations within an underlying medium rather than only being built on top of an underlying medium from the beginning block of the underlying medium.
  • medium 8 has an offset of 500, which indicates that block 0 of medium 8 will map to block 500 of its underlying medium (medium 1). Therefore, a lookup of medium 1 via medium 8 will add an offset of 500 to the original block number of the request.
  • the offset column allows a medium to be composed of multiple mediums. For example, in one embodiment, a medium may be composed of a “gold master” operating system image and per-VM (virtual machine) scratch space. Other flexible mappings are also possible and contemplated.
  • Each entry also includes an underlying medium attribute, which indicates the underlying medium of the source medium. If the underlying medium points to the source medium (as with medium 1), then this indicates that the source medium does not have an underlying medium, and all lookups will only be performed in the source medium.
  • Each entry may also include a stable attribute, with “Y” (yes) indicating the medium is stable (or read-only), and with “N” (no) indicating the medium is read-write.
  • a stable medium the data corresponding to a given block in the medium never changes, though the mapping that produces this data may change.
  • medium 2 is stable, but block 50 in medium 2 might be recorded in medium 2 or in medium 1, which may be searched logically in that order, though the searches may be done in parallel if desired.
  • a medium will be stable if the medium is used as an underlying medium by any medium other than itself.
  • a given received read/write request received by a storage controller may identify a particular volume, sector (or block number), and length.
  • the volume may be translated into a medium ID using the volume-to-medium mapping table.
  • the medium ID and block number may then be used to access index 410 of address translation table 400 to locate an index entry corresponding to the specific medium ID and block number.
  • the index entry may store a level ID and page ID of a corresponding entry in translation table 420 .
  • the corresponding translation table entry may be located and a pointer to the storage location may be returned from this entry.
  • the pointer may be used to identify or locate data stored in the storage devices of the storage system.
  • the storage system may include storage devices (e.g., SSDs) which have internal mapping mechanisms.
  • the pointer in the mapping table entry may not be an actual physical address per se. Rather, the pointer may be a logical address which the storage device maps to a physical location within the device.
  • the key value used to access entries in index 410 is the medium ID and block number corresponding to the data request.
  • a key generator may generate a key from the medium ID, block number, and/or one or more other requester data inputs, and the key may be used to access index 410 and locate a corresponding entry.
  • index 410 When index 410 is accessed with a query key value, index 410 may be searched for one or more entries which match, or otherwise correspond to, the key value. Attributes from the matching entry may then be used to locate and retrieve a mapping in translation table 420 .
  • a hit in the index provides a corresponding level ID and page ID identifying a level and page within translation table 420 storing both the key value and a corresponding physical pointer value. The page identified by the corresponding page ID may be searched with the key value so as to retrieve the corresponding pointer.
  • Translation table 420 may comprise one or more levels.
  • table 420 may comprise 16 to 64 levels, although another number of levels supported within a mapping table is possible and contemplated. Three levels labeled Level “N”, Level “N ⁇ 1” and Level “N ⁇ 2” are shown for ease of illustration.
  • Each level within table 420 may include one or more partitions.
  • each partition is a 4 kilo-byte (KB) page.
  • a corresponding index 410 may be included in each level of translation table 420 .
  • each level and each corresponding index 410 may be physically stored in a random-access manner within the storage devices.
  • index 410 may be divided into partitions, such as partitions 412 a - 412 b .
  • the size of the partitions may range from a 4 kilobyte (KB) page to 256 KB, though other sizes are possible and are contemplated.
  • Each entry of index 410 may store a key value, and the key value may be based on the medium ID, block number, and other values.
  • the key value in each entry is represented by the medium ID and block number. This is shown merely to aid in the discussion of mapping between mediums and entries in index 410 . In other embodiments, the key values of entries in index 410 may vary in how they are generated.
  • portions of index 410 may be cached, or otherwise stored in a relatively fast access memory.
  • the entire index 410 may be cached.
  • secondary, tertiary, or other index portions may be used in the cache to reduce its size.
  • mapping pages corresponding to recent hits may be cached for at least some period of time. In this manner, processes which exhibit accesses with temporal locality can be serviced more rapidly (i.e., recently accessed locations will have their mappings cached and readily available).
  • index 410 may be a secondary index which may be used to find a key value for accessing a primary index. The primary index may then be used for locating corresponding entries in address translation table 400 . It is to be understood that any number of levels of indexes may be utilized in various embodiments. In addition, any number of levels of redirection may be utilized for performing the address translation of received data requests, depending on the embodiment. In another embodiment, index 410 may be a separate entity or entities from address translation table 400 . It is noted that in other embodiments, other types of indexes and translation tables may be utilized to map medium IDs and block numbers to physical storage locations.
  • the search optimization operation may be performed to collapse medium 54 into a medium that points to itself (i.e., has no underlying medium). This will ensure that future lookups to medium 54 are more efficient by only having to perform lookups to a single medium.
  • address translation table entries may be consolidated for medium 54, resulting in all of the blocks of medium 54 being mapped directly from medium 54 rather than from a lower level medium as was previously the case.
  • FIG. 5 shows the status of medium 54 prior to the search optimization operation being performed.
  • medium 54 has a range of 0-299 and an underlying medium of 38, and an offset of 200 is applied to blocks when going down to medium 38 from medium 54.
  • Table 505 A illustrates the entries of the medium mapping table relevant to medium 54.
  • Table 505 A illustrates the relationships between the mediums prior to the search optimization operation taking place.
  • medium 54 points to medium 38, which in turn points to medium 37.
  • Table 500 A shows only index entries associated with medium 54. These entries may actually be scattered throughout the overall address translation table but are shown as being adjacent entries in table 500 A merely for ease of illustration. It is also noted that there may be other levels of indirection in the overall address translation table that convert a medium and block number to a corresponding storage location, but these other levels are not shown to avoid cluttering the figure.
  • the index entries shown in table 500 A represent all of the index entries in the address translation table associated with medium 54. While each entry in table 500 A corresponds to a range size of 100 blocks, it is noted that in other embodiments, entries may correspond to other range sizes. In other embodiments, each entry in an address translation table may correspond to a smaller number of blocks. For example, in another embodiment, each entry in an address translation table may correspond to eight blocks.
  • the storage controller performed a search to locate all index entries associated with medium 54, and the index entries shown in table 500 A represent the result of this search.
  • the storage controller would search for index entries corresponding to the underlying medium of medium 54, which in this case is medium 38. Since medium 54 maps to medium 38 with an offset of 200, block 0 of medium 54 translates to block 200 of medium 38, and so the storage controller would perform a search for block 200 of medium 38. It is assumed that this lookup also resulted in a miss, in which case the storage controller would search the underlying medium of 38, which in this case is medium 37.
  • a search for block 100 of medium 54 would eventually locate the entry for block 300 of medium 38, and accordingly, an entry corresponding to block 300 of medium 38 is shown in table 500 A.
  • a search for block 200 of medium 54 would locate an entry assigned to medium 54, and therefore, an entry corresponding to block 200 of medium 54 is shown in table 500 A. It will be assumed for the purposes of this discussion that the three entries shown in table 500 A cover the entire address space of medium 54.
  • a storage controller may perform a search optimization operation for medium 54, and the results of this operation are shown on the right-side of FIG. 5 .
  • the first three entries of address translation table 500 B are the same as the entries of table 500 A.
  • the entries assigned to medium 38 and medium 37 may be reclaimed by the storage controller if these blocks of medium 38 and medium 37 are no longer reachable (by upper-level mediums or by user volumes). However, these entries may remain in table 500 B for a period of time prior to being reclaimed.
  • Two new entries have been added to table 500 B as part of the search optimization operation, and these two entries are assigned to medium 54. These two new entries correspond to blocks 0 and 100 of medium 54, and attributes (e.g., page, level) may be copied from the corresponding existing entries and stored in the new entries. After the search optimization operation for medium 54 is performed, a lookup of table 500 B for any block of medium 54 will result in a hit for this lookup. Therefore, future lookups of medium 54 will become more efficient. It is noted that other levels of address translation table 500 B may also be updated as part of the search optimization operation for medium 54, but these updates are not shown in FIG. 5 to avoid cluttering the figure.
  • Medium mapping table 505 B shows an updated entry for medium 54, which now points to itself as its own underlying medium. This is the case because a lookup of address translation table 500 B for medium 54 will always result in a hit for any blocks of medium 54.
  • Medium graph 510 B also illustrates the new relationships between mediums after the search optimization operation, with medium 54 being its own underlying medium, and with medium 38 still pointing to medium 37.
  • medium 54 may be the underlying medium of one or more other mediums. Any upper-level mediums that have medium 54 as their underlying medium may benefit from more efficient lookups after the search optimization operation is performed since they will have fewer underlying mediums to traverse now that medium 54 has been consolidated. In addition, mediums which have medium 54 as their underlying medium may be converted into quiesced mediums since lookups for these mediums will only have to search at most two mediums.
  • address translation tables 500 A and 500 B may have multiple levels of indirection and in these embodiments, each level of indirection which is accessed using a key generated from the medium ID and block number may be updated as part of the search optimization operation.
  • FIG. 5 includes a small number of table entries and only two levels of underlying mediums beneath medium 54, in actual scenarios encountered by operational storage systems, the number of table entries and levels of underlying mediums may be substantially larger. In these embodiments, higher performance and efficiency gains can be attained by performing search optimization operations.
  • FIG. 6 one embodiment of a portion of a medium mapping table 600 is shown.
  • the entries for medium 5 shown in medium mapping table 600 illustrate another technique for medium graph collapsing by shortcutting the medium mapping table.
  • Table 600 includes the same entries for medium 2 and medium 1 as shown in table 300 of FIG. 3 . However, the entries for medium 5 have been updated to shortcut the table and reduce the number of lookups that are needed for medium 5.
  • medium 5 has a single entry in table 300 of FIG. 3
  • medium 5 now has two entries in table 600 after the shortcut has been created.
  • the storage controller may detect that blocks 100-999 of medium 5 are only found in medium 2 (at the same offset), while these blocks in medium 2 are only found in medium 1. This condition may be detected by determining that blocks 100-999 of medium 2 are masked, which indicates that all blocks in this range are in the underlying medium (medium 1). Therefore, the storage controller may split up the single entry for medium 5 in table 300 into two range entries in table 600 . The first new entry is for blocks 0-99 of medium 5, and this entry is still in the registered, unmasked (RU) state with underlying medium 2.
  • RU unmasked
  • the second new entry is for blocks 100-999 of medium 5, and this entry is in the quiescent, unmasked (QU) state with underlying medium 1.
  • Lookups for blocks 100-999 of medium 5 may now be performed in a single underlying medium search (medium 1) rather than requiring the storage controller to traverse two underlying mediums (medium 2 and medium 1).
  • Another technique for collapsing the medium graph is to merge one or more mediums that cannot be referenced externally. These one or more mediums may be merged to the medium directly above them in the medium graph. For example, referring back to table 300 of FIG. 3 , if medium 10 were no longer externally visible, the storage controller could merge medium 10 with medium 14 by coalescing entries in medium 10 into medium 14 and renumbering these entries. While the storage controller could merge the mappings in medium 14 into medium 25, the storage controller could not then delete medium 14 because medium 18 still uses medium 14 as its underlying medium, even though medium 14 might not be externally visible.
  • FIG. 7 one embodiment of a method 700 for shortcutting a medium mapping table is shown.
  • the components embodied in system 100 described above e.g., storage controller 110 ) may generally operate in accordance with method 700 .
  • the steps in this embodiment are shown in sequential order. However, some steps may occur in a different order than shown, some steps may be performed concurrently, some steps may be combined with other steps, and some steps may be absent in another embodiment.
  • the storage controller may initiate a process for shortcutting the medium mapping table (block 705 ).
  • this process for shortcutting the medium mapping table may only be one of multiple techniques for collapsing the medium graph and optimizing lookups that may be implemented.
  • this process for shortcutting may be combined with a process for consolidating entries in the medium mapping table (as described below in FIG. 8 ).
  • Other techniques may also be combined with these methods to optimize lookups and reduce the number of mediums in the storage system.
  • the storage controller may search for mediums to be read optimized (block 710 ).
  • the storage controller may select a medium range that has a status of registered (block 715 ). If a medium range has a registered status, this indicates that there are two or more levels of underlying medium ranges below this medium range. A status of registered also indicates that a search for blocks in that medium range may proceed recursively through multiple medium ranges.
  • the medium range selected in block 715 may also be referred to as the original medium range for the purposes of this discussion. It is noted that in some cases, the selected medium may only have a single entry in the medium mapping table corresponding to a single range.
  • the storage controller may go down a level in the medium mapping table (block 720 ).
  • the operation of going down a level refers to finding the entry that corresponds to the medium range which underlies the selected medium range. It is noted that this underlying medium range may be broken up into multiple ranges and may have multiple entries in the medium mapping table corresponding to the selected medium range. For these cases, each range of the underlying medium range may be processed separately and the following steps of method 700 may continue in parallel for each of these ranges. Alternatively, the steps of method 700 may continue for a first entry corresponding to a first range, and when the steps of method 700 have been completed for the first entry, the second entry for the underlying medium range may be processed, then the third entry, and so on.
  • an offset may be applied to the original block numbers when going down a level from the original medium range to the underlying medium range.
  • the underlying medium range may now be referred to as the current medium range for the discussion pertaining to blocks 725 - 730 of method 700 .
  • the storage controller may determine if the current medium range is unmasked (conditional block 725 ). In one embodiment, this may be determined by checking the status of the current medium range in the medium mapping table. If the current medium range is unmasked (conditional block 725 , “yes” leg), then method 700 may return to block 715 and the storage controller may select another medium range that has a registered status.
  • the storage controller may select another medium range with a registered status. If the current medium range is masked (conditional block 725 , “no” leg), then the storage controller may go down a level in the medium mapping table by going to the medium range which underlies the current medium range (block 730 ). If the current medium range is masked, this indicates that the current medium range does not have any mappings assigned to itself and that all mappings are assigned to lower-level medium ranges.
  • Method 700 continues after block 730 by determining how many levels can be shortcut from the table, since in some cases, more than one level may be shortcut from the medium mapping table. It is noted that when going down a level in the medium mapping table, this lower-level medium range may be split up into multiple ranges which correspond to the current medium range. In these cases, the remaining steps of method 700 may be performed separately for each range of this lower-level medium range. Each range may be handled in a parallel or in a serial fashion. After block 730 , after going down a level in the medium mapping table, this underlying medium range (of the current medium range) may now be referred to as the current medium range for the remaining steps of method 700 .
  • the storage controller may determine if the current medium range has an underlying medium range (conditional block 735 ). If the current medium range has an underlying medium range (conditional block 735 , “yes” leg), then the storage controller may determine if the current medium range is unmasked (conditional block 740 ). If the current medium range does not have an underlying medium range (conditional block 735 , “no” leg), then the storage controller may create a new entry in the medium mapping table for the original medium range which points to the current medium range as its underlying medium range (block 745 ). This entry may be for a portion of the original medium range or for the entirety of the original medium range, depending on the circumstances detected when performing blocks 720 and 730 .
  • method 700 may return to block 715 to select a medium range with a registered status.
  • condition block 740 “yes” leg
  • the storage controller may create a new entry in the medium mapping table for the original medium range which points to the current medium range as its underlying medium range (block 745 ). If the current medium range is masked (conditional block 740 , “no” leg), then method 700 may return to block 730 to go down to a lower level in the medium mapping table. In this case, a shortcut of the table may span more than one level if each of the two or more intermediate levels (between the original medium range and the current medium range) have an underlying medium range and are masked.
  • FIG. 8 one embodiment of a method 800 for collapsing the medium mapping table is shown.
  • the components embodied in system 100 described above e.g., storage controller 110 ) may generally operate in accordance with method 800 .
  • the steps in this embodiment are shown in sequential order. However, some steps may occur in a different order than shown, some steps may be performed concurrently, some steps may be combined with other steps, and some steps may be absent in another embodiment.
  • the storage controller may initiate a medium mapping table collapsing process (block 805 ).
  • the conditions which trigger the initiation of the medium mapping table collapsing process may vary from embodiment to embodiment. For example, in one embodiment, if the storage controller determines that the number of mediums exceeds a programmable threshold, then the medium mapping table collapsing may be initiated. Alternatively, if the storage controller has idle processing resources, then the storage controller may utilize these processing resources to run the medium mapping table collapsing process. In other embodiments, the storage controller may initiate the medium mapping table collapsing process on a periodic basis (e.g., daily, twice a week).
  • the medium mapping table collapsing process may select a medium for consolidating entries in the address translation table (block 810 ).
  • a medium with a long chain of underlying mediums may be selected for consolidating entries.
  • the storage controller may select a given medium for consolidation if the given medium has a chain of underlying mediums that is greater than a programmable threshold.
  • Other techniques for selecting a medium for consolidation are possible and are contemplated.
  • the storage controller may select a given range of blocks of the selected medium (block 815 ).
  • the size of the range may vary depending on the characteristics of the selected medium.
  • the storage controller may begin with a first range of blocks starting with the smallest number and proceeding to higher numbers.
  • the storage controller may begin with the highest range of blocks based on block number and proceed to lower numbers on successive searches.
  • the storage controller may process the different ranges of blocks in any order.
  • condition block 820 If the selected medium is unmasked for the current range of blocks and if an address translation table lookup for the current range of blocks of the selected medium results in a hit (conditional block 820 , “yes” leg), then the storage controller may determine if the last range of blocks for the selected medium has already been reached (conditional block 825 ). If the given range of blocks has an entry assigned to the selected medium in the index of the address translation table, then the storage controller does not need to consolidate this entry since the entry is already assigned to the selected medium and can be found using a key generated from the selected medium.
  • method 800 may return to block 810 and select a new medium to consolidate. If the storage controller has not yet reached the last range of blocks for the selected medium (conditional block 825 , “no” leg), then the storage controller may move to the next range of blocks of the selected medium for searching (block 830 ), and then method 800 may return to conditional block 820 .
  • the storage controller may determine the underlying medium of the selected medium (block 835 ). In one embodiment, the storage controller may query the medium mapping table to determine the underlying medium of the selected medium. The storage controller may also determine if an offset should be applied to the block number in the underlying medium from the original block number of the selected medium. Next, the storage controller may determine if the underlying medium is unmasked for the current range of blocks and if an address translation table lookup for the current range of blocks using the underlying medium results in a hit (conditional block 840 ). The storage controller may use a key generated from the underlying medium and current range of blocks for the address translation table lookup (with an offset applied to the block numbers, if applicable).
  • condition block 840 the storage controller may determine if the selected medium is quiescent (conditional block 845 ). If the selected medium is quiescent (conditional block 845 , “yes” leg), then this indicates that there are no more underlying mediums to search, and method 800 may return to block 810 to select a new medium for consolidating entries. If the selected medium is not quiescent (conditional block 845 , “no” leg), then the storage controller may return to block 835 to determine the underlying medium of the current medium.
  • the storage controller may create one or more new entries in the address translation table for the current range of blocks and assign the new entries to the selected medium (block 850 ).
  • Each new entry may include a key generated from the selected medium ID and the range of blocks as mapped from the selected medium. This range of blocks may be different from the current range if an offset was applied to the underlying medium by the medium above it in the medium graph.
  • the storage controller may update the medium mapping table entry status to quiescent for the current range of blocks of the selected medium (block 855 ).
  • the storage controller may copy mappings from the existing entries (assigned to the underlying medium) of the address translation table to the new entries while retaining the existing entries (block 860 ). However, in some embodiments, the storage controller may reclaim the existing entries for the underlying medium if the region of the underlying medium corresponding to the existing entries is not reachable by any other mediums. After block 860 , method 800 may return to block 825 to determine if the last range of blocks has been reached for the selected medium.
  • method 700 (of FIG. 7 ) and method 800 may be performed concurrently in various combinations to merge mappings and consolidate mediums.
  • method 700 may be performed on a first medium to shortcut the medium mapping table while method 800 is being performed in parallel on a second medium to collapse the medium mapping table for the second medium.
  • method 700 may be performed on multiple different mediums in parallel while method 800 is also performed on multiple different mediums in parallel.
  • adjacent mappings may be merged if these adjacent mappings are compatible so as to further reduce the size of the medium mapping table.
  • Adjacent mappings may be considered compatible if they have the same attributes and can be combined into a single entry in the medium mapping table.
  • a given medium 900 may have two adjacent entries. The first of these entries may be for a range of 0-99, with the underlying medium set to 905, and with a status of quiescent. The second entry may be for a range of 100-399 with the underlying medium set to 905, and with a status of quiescent.
  • These two entries can be merged into a single with a range of 0-399, with the underlying medium set to 905, and with a status of quiescent.
  • a storage controller After a storage controller has promoted mappings to higher-level mediums while performing medium collapsing operations, it is customary to find many adjacent entries with the same attributes scattered throughout the medium mapping table. Accordingly, these adjacent entries may be merged to further reduce the size of the medium mapping table.
  • the above-described embodiments may comprise software.
  • the program instructions that implement the methods and/or mechanisms may be conveyed or stored on a computer readable medium.
  • a computer readable medium Numerous types of media which are configured to store program instructions are available and include hard disks, floppy disks, CD-ROM, DVD, flash memory, Programmable ROMs (PROM), random access memory (RAM), and various other forms of volatile or non-volatile storage.
  • resources may be provided over the Internet as services according to one or more various models.
  • models may include Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS).
  • IaaS Infrastructure as a Service
  • PaaS Platform as a Service
  • SaaS Software as a Service
  • IaaS computer infrastructure is delivered as a service.
  • the computing equipment is generally owned and operated by the service provider.
  • software tools and underlying equipment used by developers to develop software solutions may be provided as a service and hosted by the service provider.
  • SaaS typically includes a service provider licensing software as a service on demand. The service provider may host the software, or may deploy the software to a customer for a given period of time. Numerous combinations of the above models are possible and are contemplated.

Abstract

A system and method for performing search optimization operations. In order to optimize lookups for received data requests, a storage controller attempts to collapse the medium graph, thus reducing the number of mediums in which the storage controller must look to find a given block. One technique for collapsing the medium graph involves promoting individual mappings from a lower-level medium to a higher-level medium. Another technique for collapsing the medium graph involves pushing medium range pointers higher up in the medium mapping table to reduce the number of mediums that must be searched.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application No. 61/751,142, filed Jan. 10, 2013, the entirety of which is incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • This invention relates to optimizing snapshot lookup operations in a storage system.
  • 2. Description of the Related Art
  • As computer memory storage and data bandwidth increase, so does the amount and complexity of data that businesses daily manage. Large-scale distributed storage systems, such as data centers, typically run many business operations. A datacenter, which also may be referred to as a server room, is a centralized repository, either physical or virtual, for the storage, management, and dissemination of data pertaining to one or more businesses. A distributed storage system may be coupled to client computers interconnected by one or more networks. If any portion of the distributed storage system has poor performance, company operations may be impaired. A distributed storage system therefore maintains high standards for data availability and high-performance functionality.
  • Many techniques have been employed by storage systems to maintain high performance. For example, snapshots may be utilized to capture and store data at a particular point in time. A snapshot may be taken of a logical volume, and the snapshot may be stored to preserve the contents of the volume. If the data associated with the volume is later lost or corrupted, the volume can be restored from the snapshot.
  • As the amount of data corresponding to a snapshot increases, and as the overall number of snapshots being taken and stored increases, the storage utilization and processing overhead of the storage system likewise increases. The proliferation of snapshots in a storage system results in more processing overhead for the storage system as well as increasing numbers of lookups through multiple snapshots for some data requests. With time, the storage system may be required to manage a large number of snapshots, some of which may no longer be externally visible by clients.
  • In view of the above, systems and methods for optimizing snapshot lookup operations are desired.
  • SUMMARY OF THE INVENTION
  • Various embodiments of systems and methods for optimizing medium lookup operations are contemplated.
  • A storage system may include a storage controller and one or more storage devices. The storage system may be coupled to one or more host client systems. In one embodiment, the storage controller may utilize volumes and mediums to track client data that is stored in the storage system. A medium is defined as a logical grouping of data, and each medium has an identifier with which to identify the logical grouping of data. The storage controller may maintain a volume to medium mapping table to map each volume to a single medium, and this medium is referred to as the volume's anchor medium. Each medium may be mapped to any number of other mediums, and the storage controller may also maintain a medium mapping table to track relationships between anchor mediums and underlying mediums. The storage controller may also maintain an address translation table and corresponding indexes for mapping mediums and block numbers to physical storage locations.
  • In one embodiment, the storage controller may be configured to perform medium lookup optimizing operations on a periodic basis. The storage controller may search through all of the mediums in the system to find mediums that can be reduced or collapsed to make future searches of the system more efficient. The storage controller may look for active mediums in a read-write state which have one or more levels of underlying mediums beneath them, with each underlying medium representing a snapshot. These active mediums with one or more underlying mediums may be collapsed by the storage controller. The storage controller may not make changes that cause the result of a lookup of a block on a volume to change, but may make changes in mediums that are invisible from the point of view of all volumes on the storage controller.
  • In one embodiment, the storage controller may attempt to collapse the medium mapping table so as to reduce the number of mediums in which the storage controller will have to look to find data blocks for a given data request. The storage controller may collapse the medium mapping table by promoting individual mappings from lower-level mediums to higher-level mediums. The storage controller may also collapse the medium mapping table by pushing medium range pointers higher up in the table so as to reduce the number of mediums that need to be searched.
  • In one embodiment, the storage controller may promote individual mappings from existing entries in the address translation table into new entries for the destination medium. For a given medium with multiple levels of underlying mediums, all of the mappings for these levels of underlying mediums may be merged, resulting in a new version of the given medium. This process does not change the result of any individual mapping but rather improves the performance of the storage system by allowing the given medium to be converted into a quiesced medium that refers to itself. After receiving a data request targeting the given medium, the storage controller will have access to all of its mappings by performing a single lookup to the address translation table indexes for only the given medium.
  • In one embodiment, another technique may be utilized by the storage controller to reduce the number of lookups required by shortcutting the medium mapping table. In order to shortcut the medium mapping table, the system may collapse ranges of masked mediums together. A range of a medium is considered unmasked if one or more address translation table entries exist in the range for that medium, which indicates that the range has been previously written to. A range of a medium is considered masked if there are no entries in the range for that medium. If a first medium has multiple levels of underlying mediums below it, and its immediate underlying medium is masked, then a shortcut through the medium mapping table may be created, bypassing the medium directly below the first medium. In this case, a new entry may be created and entered into the medium mapping table, with the new entry mapping the first medium directly to a medium at least two levels below the first medium in the medium graph.
  • In another embodiment, another technique for collapsing the medium graph may be utilized by the storage controller. This technique may involve merging one or more mediums that cannot be referenced externally into the medium directly above them in the medium graph. If a second medium underlies a first medium and the second medium is no longer externally visible, the storage controller may merge the second medium with the first medium, combining the entries of the second medium and the first medium and renumbering the entries.
  • The combination of these techniques can dramatically shrink the number of mediums in use by the storage system, particularly in systems that take frequent checkpoints and then delete the volumes which access the checkpoints. Also, in systems with long chains of mediums, these techniques can help reduce these chains and reduce the number of medium lookups that are performed for received data requests.
  • These and other embodiments will become apparent upon consideration of the following description and accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a generalized block diagram illustrating one embodiment of a storage system.
  • FIG. 2 is a generalized block diagram of one embodiment of a directed acyclic graph (DAG) of mediums.
  • FIG. 3 illustrates one embodiment of a medium mapping table.
  • FIG. 4 illustrates a block diagram of one embodiment of an address translation table.
  • FIG. 5 is a generalized block diagram of one embodiment of a search optimization operation.
  • FIG. 6 illustrates one embodiment of a portion of a medium mapping table.
  • FIG. 7 is a generalized flow diagram illustrating one embodiment of a method for shortcutting a medium mapping table.
  • FIG. 8 is a generalized flow diagram illustrating one embodiment of a method for performing medium graph collapsing operations.
  • While the invention is susceptible to various modifications and alternative forms, specific embodiments are shown by way of example in the drawings and are herein described in detail. It should be understood, however, that drawings and detailed description thereto are not intended to limit the invention to the particular form disclosed, but on the contrary, the invention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention as defined by the appended claims.
  • DETAILED DESCRIPTION
  • In the following description, numerous specific details are set forth to provide a thorough understanding of the present invention. However, one having ordinary skill in the art should recognize that the invention might be practiced without these specific details. In some instances, well-known circuits, structures, signals, computer program instruction, and techniques have not been shown in detail to avoid obscuring the present invention.
  • Referring now to FIG. 1, a generalized block diagram of one embodiment of a storage system 100 is shown. Storage system 100 may include storage controller 110 and storage device groups 130 and 140, which are representative of any number of storage device groups (or data storage arrays). As shown, storage device group 130 includes storage devices 135A-N, which are representative of any number and type of storage devices (e.g., solid-state drives (SSDs)). Storage controller 110 may be coupled directly to client computer system 125, and storage controller 110 may be coupled remotely over network 120 to client computer system 115. Clients 115 and 125 are representative of any number of clients which may utilize storage controller 110 for storing and accessing data in system 100. It is noted that some systems may include only a single client, connected directly or remotely to storage controller 110.
  • Storage controller 110 may include software and/or hardware configured to provide access to storage devices 135A-N. Although storage controller 110 is shown as being separate from storage device groups 130 and 140, in some embodiments, storage controller 110 may be located within one or each of storage device groups 130 and 140. Storage controller 110 may include or be coupled to a base operating system (OS), a volume manager, and additional control logic for implementing the various techniques disclosed herein.
  • Storage controller 110 may include and/or execute on any number of processors and may include and/or execute on a single host computing device or be spread across multiple host computing devices, depending on the embodiment. In some embodiments, storage controller 110 may generally include or execute on one or more file servers and/or block servers. Storage controller 110 may use any of various techniques for replicating data across devices 135A-N to prevent loss of data due to the failure of a device or the failure of storage locations within a device. Storage controller 110 may also utilize any of various deduplication techniques for reducing the amount of data stored in devices 135A-N by deduplicating common data.
  • Storage controller 110 may also be configured to create and manage snapshots in system 100. A set of mediums may be recorded and maintained by storage controller 110. Most of the mediums may be read-only except for one or more selected mediums such as the most recent medium in use by a particular volume. Each medium logically comprises all of the blocks in the medium. However, only the blocks that were changed from the time the medium was created to the time the medium was closed are saved and mappings to these blocks may also be maintained with the medium.
  • In various embodiments, multiple mapping tables may be maintained by storage controller 110. These mapping tables may include a medium mapping table and a volume-to-medium mapping table. These tables may be utilized to record and maintain the mappings between mediums and underlying mediums and the mappings between volumes and mediums. Storage controller 110 may also include an address translation table with a plurality of entries, wherein each entry holds a virtual-to-physical mapping for a corresponding data component. This mapping table may be used to map logical read/write requests from each of the client computer systems 115 and 125 to physical locations in storage devices 135A-N. A “physical” pointer value may be read from the mappings associated with a given medium during a lookup operation corresponding to a received read/write request. This physical pointer value may then be used to locate a physical location within the storage devices 135A-N. It is noted the physical pointer value may be used to access another mapping table within a given storage device of the storage devices 135A-N. Consequently, one or more levels of indirection may exist between the physical pointer value and a target storage location.
  • In various embodiments, the address translation table may be accessed using a key comprising a medium ID, a logical or virtual address, a sector number, and so forth. A received read/write storage access request may identify a particular volume, sector, and length. The volume ID may be mapped to a medium ID using the volume to medium mapping table. A sector may be a logical block of data stored in a medium. Sectors may have different sizes on different mediums. The address translation table may map a medium in sector-size units. In other embodiments, other types of address translation tables may be utilized.
  • In one embodiment, the address translation table may map mediums to physical pointer values. For a given storage access request, an access may be made to the “mappings” of the address translation table. The term “mappings” is defined as the one or more entries of the address translation table which convert a given medium ID and block number into a physical pointer value. Depending on the embodiment, a physical pointer value may be a physical address or a logical address which the storage device maps to a physical location within the device. In one embodiment, an index may be utilized to access the address translation table. The index may identify locations of mappings within the address translation table. The index may be queried with a key value generated from a medium ID and sector number, and the index may be searched for one or more entries which match, or otherwise correspond to, the key value. Information from a matching entry may then be used to locate and retrieve a mapping which identifies a storage location which is the target of a received read or write request. In one embodiment, a hit in the index provides a corresponding virtual page ID identifying a page within the storage devices of the storage system, wherein the page stores both the key value and a corresponding physical pointer value. The page may then be searched with the key value to find the physical pointer value.
  • It is noted that in alternative embodiments, the number and type of client computers, storage controllers, networks, storage device groups, and data storage devices is not limited to those shown in FIG. 1. At various times one or more clients may operate offline. In addition, during operation, individual client computer connection types may change as users connect, disconnect, and reconnect to system 100. Further, the systems and methods described herein may be applied to directly attached storage systems or network attached storage systems and may include a host operating system configured to perform one or more aspects of the described methods. Numerous such alternatives are possible and are contemplated.
  • Network 120 may utilize a variety of techniques including wireless connection, direct local area network (LAN) connections, wide area network (WAN) connections such as the Internet, a router, storage area network, Ethernet, and others. Network 120 may comprise one or more LANs that may also be wireless. Network 120 may further include remote direct memory access (RDMA) hardware and/or software, transmission control protocol/internet protocol (TCP/IP) hardware and/or software, router, repeaters, switches, grids, and/or others. Protocols such as Fibre Channel, Fibre Channel over Ethernet (FCoE), iSCSI, and so forth may be used in network 120. The network 120 may interface with a set of communications protocols used for the Internet such as the Transmission Control Protocol (TCP) and the Internet Protocol (IP), or TCP/IP.
  • Client computer systems 115 and 125 are representative of any number of stationary or mobile computers such as desktop personal computers (PCs), servers, server farms, workstations, laptops, handheld computers, servers, personal digital assistants (PDAs), smart phones, and so forth. Generally speaking, client computer systems 115 and 125 include one or more processors comprising one or more processor cores. Each processor core includes circuitry for executing instructions according to a predefined general-purpose instruction set. For example, the x86 instruction set architecture may be selected. Alternatively, the ARM®, Alpha®, PowerPC®, SPARC®, or any other general-purpose instruction set architecture may be selected. The processor cores may access cache memory subsystems for data and computer program instructions. The cache subsystems may be coupled to a memory hierarchy comprising random access memory (RAM) and a storage device.
  • Referring now to FIG. 2, a block diagram illustrating a directed acyclic graph (DAG) 200 of mediums is shown. Also shown is a volume to medium mapping table 205 which shows which medium a volume maps to for each volume in use by a storage system. Volumes may be considered pointers into graph 200.
  • The term “medium” as is used herein is defined as a logical grouping of data. A medium may have a corresponding identifier with which to identify the logical grouping of data. Each medium may also include or be associated with mappings of logical block numbers to content location, deduplication entries, and other information. In one embodiment, medium identifiers may be used by the storage controller but medium identifiers may not be user-visible. A user (or client) may send a data request accompanied by a volume ID to specify which data is targeted by the request, and the storage controller may map the volume ID to a medium ID and then use the medium ID when processing the request.
  • The term medium is not to be confused with the terms “storage medium” or “computer readable storage medium”. A storage medium is defined as an actual physical device (e.g., SSD, HDD) that is utilized to store data. A computer readable storage medium (or non-transitory computer readable storage medium) is defined as a physical storage medium configured to store program instructions which are executable by a processor or other hardware device. Various types of program instructions that implement the methods and/or mechanisms described herein may be conveyed or stored on a computer readable medium. Numerous types of media which are configured to store program instructions are available and include hard disks, floppy disks, CD-ROM, DVD, flash memory, Programmable ROMs (PROM), random access memory (RAM), and various other forms of volatile or non-volatile storage.
  • It is also noted that the term “volume to medium mapping table” may refer to multiple tables rather than just a single table. Similarly, the term “medium mapping table” may also refer to multiple tables rather than just a single table. It is further noted that volume to medium mapping table 205 is only one example of a volume to medium mapping table. Other volume to medium mapping tables may have other numbers of entries for other numbers of volumes.
  • Each medium is depicted in graph 200 as three conjoined boxes, with the leftmost box showing the medium ID, the middle box showing the underlying medium, and the rightmost box displaying the status of the medium (RO—read-only) or (RW—read-write). Read-write mediums may be referred to as active mediums, while read-only mediums may represent previously taken snapshots. Within graph 200, a medium points to its underlying medium. For example, medium 20 points to medium 12 to depict that medium 12 is the underlying medium of medium 20. Medium 12 also points to medium 10, which in turn points to medium 5, which in turn points to medium 1. Some mediums are the underlying medium for more than one higher-level medium. For example, three separate mediums (12, 17, 11) point to medium 10, two separate mediums (18, 10) point to medium 5, and two separate mediums (6, 5) point to medium 1. Each of the mediums which is an underlying medium to at least one higher-level medium has a status of read-only.
  • The set of mediums on the bottom left of graph 200 is an example of a linear set. As depicted in graph 200, medium 3 was created first and then a snapshot was taken resulting in medium 3 becoming stable (i.e., the result of a lookup for a given block in medium 3 will always return the same value after this point). Medium 7 was created with medium 3 as its underlying medium. Any blocks written after medium 3 became stable were labeled as being in medium 7. Lookups to medium 7 return the value from medium 7 if one is found, but will look in medium 3 if a block is not found in medium 7. At a later time, a snapshot of medium 7 is taken, medium 7 becomes stable, and medium 14 is created. Lookups for blocks in medium 14 would check medium 7 and then medium 3 to find the targeted logical block. Eventually, a snapshot of medium 14 is taken and medium 14 becomes stable while medium 15 is created. At this point in graph 200, medium 14 is stable with writes to volume 102 going to medium 15.
  • Volume to medium mapping table 205 maps user-visible volumes to mediums. Each volume may be mapped to a single medium, also known as the anchor medium. This anchor medium, as with all other mediums, may take care of its own lookups. A medium on which multiple volumes depend (such as medium 10) tracks its own blocks independently of the volumes which depend on it. Each medium may also be broken up into ranges of blocks, and each range may be treated separately in medium DAG 200.
  • Referring now to FIG. 3, one embodiment of a medium mapping table 300 is shown. Any portion of or the entirety of medium mapping table 300 may be stored in storage controller 110 and/or in one or more of storage devices 135A-N. A volume identifier (ID) may be used to access volume to medium mapping table 205 to determine a medium ID corresponding to the volume ID. This medium ID may then be used to access medium mapping table 300. It is noted that table 300 is merely one example of a medium mapping table, and that in other embodiments, other medium mapping tables, with other numbers of entries, may be utilized. In addition, in other embodiments, a medium mapping table may include other attributes and be organized in a different manner than that shown in FIG. 3. It is also noted that any suitable data structure may be used to store the mapping table information in order to provide for efficient searches (e.g., b-trees, binary trees, hash tables, etc.). All such data structures are contemplated.
  • Each medium may be identified by a medium ID, as shown in the leftmost column of table 300. A range attribute may also be included in each entry of table 300, and the range may be in terms of data blocks. The size of a block of data (e.g., 4 KB, 8 KB) may vary depending on the embodiment. A medium may be broken up into multiple ranges, and each range of a medium may be treated as if it is an independent medium with its own attributes and mappings. Throughout this disclosure, there are various terms used to describe mediums as well as various techniques described for operating or performing actions on mediums. These terms and techniques also apply to ranges of mediums. For example, medium ID 2 has two separate ranges. Range 0-99 of medium ID 2 has a separate entry in table 300 from the entry for range 100-999 of medium ID 2.
  • Although both of these ranges of medium ID 2 map to underlying medium ID 1, it is possible for separate ranges of the same source medium to map to different underlying mediums. For example, separate ranges from medium ID 35 map to separate underlying mediums. For example, range 0-299 of medium ID 35 maps to underlying medium ID 18 with an offset of 400. This indicates that blocks 0-299 of medium ID 35 map to blocks 400-699 of medium ID 18. Additionally, range 300-499 of medium ID 35 maps to underlying medium ID 33 with an offset of −300 and range 500-899 of medium ID 35 maps to underlying medium ID 5 with an offset of −400. These entries indicate that blocks 300-499 of medium ID 35 map to blocks 0-199 of medium ID 33 while blocks 500-899 of medium ID 35 map to blocks 100-499 of medium ID 5. It is noted that in other embodiments, mediums may be broken up into more than three ranges.
  • The state column of table 300 records information that allows lookups for blocks to be performed more efficiently. A state of “Q” indicates the medium is quiescent, “R” indicates the medium is registered, and “U” indicates the medium is unmasked. In the quiescent state, a lookup is performed on exactly one or two mediums specified in table 300. In the registered state, a lookup is performed recursively. If an entry in table 300 is unmasked, then this indicates that a lookup should be performed in the basis medium. If an entry is masked, then the lookup should only be performed in the underlying medium. Although not shown in table 300 for any of the entries, another state “X” may be used to specify that the source medium is unmapped. The unmapped state indicates that the source medium contains no reachable data and can be discarded. This unmapped state may apply to a range of a source medium. If an entire medium is unmapped, then the medium ID may be entered into a sequence invalidation table and eventually discarded.
  • In one embodiment, when a medium is created, the medium is in the registered state if it has an underlying medium, or the medium is in the quiescent state if it is a brand-new volume with no pre-existing state. As the medium is written to, parts of it can become unmasked, with mappings existing both in the medium itself and the underlying medium. This may be done by splitting a single range into multiple range entries, some of which retain the original masked status, and others of which are marked as unmasked.
  • In addition, each entry in table 300 may include a basis attribute, which indicates the basis of the medium, which in this case points to the source medium itself. Each entry may also include an offset field, which specifies the offset that should be applied to the block address when mapping the source medium to an underlying medium. This allows mediums to map to other locations within an underlying medium rather than only being built on top of an underlying medium from the beginning block of the underlying medium. As shown in table 300, medium 8 has an offset of 500, which indicates that block 0 of medium 8 will map to block 500 of its underlying medium (medium 1). Therefore, a lookup of medium 1 via medium 8 will add an offset of 500 to the original block number of the request. The offset column allows a medium to be composed of multiple mediums. For example, in one embodiment, a medium may be composed of a “gold master” operating system image and per-VM (virtual machine) scratch space. Other flexible mappings are also possible and contemplated.
  • Each entry also includes an underlying medium attribute, which indicates the underlying medium of the source medium. If the underlying medium points to the source medium (as with medium 1), then this indicates that the source medium does not have an underlying medium, and all lookups will only be performed in the source medium. Each entry may also include a stable attribute, with “Y” (yes) indicating the medium is stable (or read-only), and with “N” (no) indicating the medium is read-write. In a stable medium, the data corresponding to a given block in the medium never changes, though the mapping that produces this data may change. For example, medium 2 is stable, but block 50 in medium 2 might be recorded in medium 2 or in medium 1, which may be searched logically in that order, though the searches may be done in parallel if desired. In one embodiment, a medium will be stable if the medium is used as an underlying medium by any medium other than itself.
  • Turning now to FIG. 4, a block diagram of one embodiment of an address translation table 400 is shown. In one embodiment, a given received read/write request received by a storage controller may identify a particular volume, sector (or block number), and length. The volume may be translated into a medium ID using the volume-to-medium mapping table. The medium ID and block number may then be used to access index 410 of address translation table 400 to locate an index entry corresponding to the specific medium ID and block number. The index entry may store a level ID and page ID of a corresponding entry in translation table 420. Using the level ID, page ID, and a key value generated from the medium ID and block number, the corresponding translation table entry may be located and a pointer to the storage location may be returned from this entry. The pointer may be used to identify or locate data stored in the storage devices of the storage system. It is noted that in various embodiments, the storage system may include storage devices (e.g., SSDs) which have internal mapping mechanisms. In such embodiments, the pointer in the mapping table entry may not be an actual physical address per se. Rather, the pointer may be a logical address which the storage device maps to a physical location within the device.
  • For the purposes of this discussion, the key value used to access entries in index 410 is the medium ID and block number corresponding to the data request. However, in other embodiments, other types of key values may be utilized. In these embodiments, a key generator may generate a key from the medium ID, block number, and/or one or more other requester data inputs, and the key may be used to access index 410 and locate a corresponding entry.
  • When index 410 is accessed with a query key value, index 410 may be searched for one or more entries which match, or otherwise correspond to, the key value. Attributes from the matching entry may then be used to locate and retrieve a mapping in translation table 420. In one embodiment, a hit in the index provides a corresponding level ID and page ID identifying a level and page within translation table 420 storing both the key value and a corresponding physical pointer value. The page identified by the corresponding page ID may be searched with the key value so as to retrieve the corresponding pointer.
  • Translation table 420 may comprise one or more levels. For example, in various embodiments, table 420 may comprise 16 to 64 levels, although another number of levels supported within a mapping table is possible and contemplated. Three levels labeled Level “N”, Level “N−1” and Level “N−2” are shown for ease of illustration. Each level within table 420 may include one or more partitions. In one embodiment, each partition is a 4 kilo-byte (KB) page. In one embodiment, a corresponding index 410 may be included in each level of translation table 420. In this embodiment, each level and each corresponding index 410 may be physically stored in a random-access manner within the storage devices.
  • In one embodiment, index 410 may be divided into partitions, such as partitions 412 a-412 b. In one embodiment, the size of the partitions may range from a 4 kilobyte (KB) page to 256 KB, though other sizes are possible and are contemplated. Each entry of index 410 may store a key value, and the key value may be based on the medium ID, block number, and other values. For the purposes of this discussion, the key value in each entry is represented by the medium ID and block number. This is shown merely to aid in the discussion of mapping between mediums and entries in index 410. In other embodiments, the key values of entries in index 410 may vary in how they are generated.
  • In various embodiments, portions of index 410 may be cached, or otherwise stored in a relatively fast access memory. In various embodiments, the entire index 410 may be cached. In some embodiments, where the primary index has become too large to cache in its entirety, or is otherwise larger than desired, secondary, tertiary, or other index portions may be used in the cache to reduce its size. In addition to the above, in various embodiments mapping pages corresponding to recent hits may be cached for at least some period of time. In this manner, processes which exhibit accesses with temporal locality can be serviced more rapidly (i.e., recently accessed locations will have their mappings cached and readily available).
  • In some embodiments, index 410 may be a secondary index which may be used to find a key value for accessing a primary index. The primary index may then be used for locating corresponding entries in address translation table 400. It is to be understood that any number of levels of indexes may be utilized in various embodiments. In addition, any number of levels of redirection may be utilized for performing the address translation of received data requests, depending on the embodiment. In another embodiment, index 410 may be a separate entity or entities from address translation table 400. It is noted that in other embodiments, other types of indexes and translation tables may be utilized to map medium IDs and block numbers to physical storage locations.
  • Referring now to FIG. 5, one embodiment of a search optimization operation is shown. The search optimization operation may be performed to collapse medium 54 into a medium that points to itself (i.e., has no underlying medium). This will ensure that future lookups to medium 54 are more efficient by only having to perform lookups to a single medium. As part of the search optimization operation, address translation table entries may be consolidated for medium 54, resulting in all of the blocks of medium 54 being mapped directly from medium 54 rather than from a lower level medium as was previously the case.
  • The left-side of FIG. 5 shows the status of medium 54 prior to the search optimization operation being performed. As shown in medium mapping table 505A, medium 54 has a range of 0-299 and an underlying medium of 38, and an offset of 200 is applied to blocks when going down to medium 38 from medium 54. It is noted that only the entries of the medium mapping table relevant to medium 54 are shown in table 505A, and the entries show only pertinent attributes to avoid cluttering the figure. Medium graph 510A illustrates the relationships between the mediums prior to the search optimization operation taking place. As can be seen from medium graph 510A, medium 54 points to medium 38, which in turn points to medium 37.
  • Also on the left-side of FIG. 5 is address translation table 500A. Table 500A shows only index entries associated with medium 54. These entries may actually be scattered throughout the overall address translation table but are shown as being adjacent entries in table 500A merely for ease of illustration. It is also noted that there may be other levels of indirection in the overall address translation table that convert a medium and block number to a corresponding storage location, but these other levels are not shown to avoid cluttering the figure. The index entries shown in table 500A represent all of the index entries in the address translation table associated with medium 54. While each entry in table 500A corresponds to a range size of 100 blocks, it is noted that in other embodiments, entries may correspond to other range sizes. In other embodiments, each entry in an address translation table may correspond to a smaller number of blocks. For example, in another embodiment, each entry in an address translation table may correspond to eight blocks.
  • It may be assumed for the purposes of this discussion that the storage controller performed a search to locate all index entries associated with medium 54, and the index entries shown in table 500A represent the result of this search. For block 0 of medium 54, there is not an index entry assigned to medium 54. Therefore, the storage controller would search for index entries corresponding to the underlying medium of medium 54, which in this case is medium 38. Since medium 54 maps to medium 38 with an offset of 200, block 0 of medium 54 translates to block 200 of medium 38, and so the storage controller would perform a search for block 200 of medium 38. It is assumed that this lookup also resulted in a miss, in which case the storage controller would search the underlying medium of 38, which in this case is medium 37. Since medium 38 maps to medium 37 with an offset of 500, block 200 of medium 38 translates to block 700 of medium 37. In this case, a lookup for block 700 of medium 37 results in a hit, and the entry corresponding to block 700 of medium 37 is shown in table 500A.
  • A search for block 100 of medium 54 would eventually locate the entry for block 300 of medium 38, and accordingly, an entry corresponding to block 300 of medium 38 is shown in table 500A. A search for block 200 of medium 54 would locate an entry assigned to medium 54, and therefore, an entry corresponding to block 200 of medium 54 is shown in table 500A. It will be assumed for the purposes of this discussion that the three entries shown in table 500A cover the entire address space of medium 54.
  • A storage controller may perform a search optimization operation for medium 54, and the results of this operation are shown on the right-side of FIG. 5. The first three entries of address translation table 500B are the same as the entries of table 500A. The entries assigned to medium 38 and medium 37 may be reclaimed by the storage controller if these blocks of medium 38 and medium 37 are no longer reachable (by upper-level mediums or by user volumes). However, these entries may remain in table 500B for a period of time prior to being reclaimed.
  • Two new entries have been added to table 500B as part of the search optimization operation, and these two entries are assigned to medium 54. These two new entries correspond to blocks 0 and 100 of medium 54, and attributes (e.g., page, level) may be copied from the corresponding existing entries and stored in the new entries. After the search optimization operation for medium 54 is performed, a lookup of table 500B for any block of medium 54 will result in a hit for this lookup. Therefore, future lookups of medium 54 will become more efficient. It is noted that other levels of address translation table 500B may also be updated as part of the search optimization operation for medium 54, but these updates are not shown in FIG. 5 to avoid cluttering the figure.
  • Medium mapping table 505B shows an updated entry for medium 54, which now points to itself as its own underlying medium. This is the case because a lookup of address translation table 500B for medium 54 will always result in a hit for any blocks of medium 54. Medium graph 510B also illustrates the new relationships between mediums after the search optimization operation, with medium 54 being its own underlying medium, and with medium 38 still pointing to medium 37.
  • Although not shown in FIG. 5, medium 54 may be the underlying medium of one or more other mediums. Any upper-level mediums that have medium 54 as their underlying medium may benefit from more efficient lookups after the search optimization operation is performed since they will have fewer underlying mediums to traverse now that medium 54 has been consolidated. In addition, mediums which have medium 54 as their underlying medium may be converted into quiesced mediums since lookups for these mediums will only have to search at most two mediums.
  • It is noted that in some embodiments, address translation tables 500A and 500B may have multiple levels of indirection and in these embodiments, each level of indirection which is accessed using a key generated from the medium ID and block number may be updated as part of the search optimization operation.
  • Although the example shown in FIG. 5 includes a small number of table entries and only two levels of underlying mediums beneath medium 54, in actual scenarios encountered by operational storage systems, the number of table entries and levels of underlying mediums may be substantially larger. In these embodiments, higher performance and efficiency gains can be attained by performing search optimization operations.
  • Turning now to FIG. 6, one embodiment of a portion of a medium mapping table 600 is shown. The entries for medium 5 shown in medium mapping table 600 illustrate another technique for medium graph collapsing by shortcutting the medium mapping table. Table 600 includes the same entries for medium 2 and medium 1 as shown in table 300 of FIG. 3. However, the entries for medium 5 have been updated to shortcut the table and reduce the number of lookups that are needed for medium 5.
  • Whereas medium 5 has a single entry in table 300 of FIG. 3, medium 5 now has two entries in table 600 after the shortcut has been created. Referring back to the single entry for medium 5 in table 300, the storage controller may detect that blocks 100-999 of medium 5 are only found in medium 2 (at the same offset), while these blocks in medium 2 are only found in medium 1. This condition may be detected by determining that blocks 100-999 of medium 2 are masked, which indicates that all blocks in this range are in the underlying medium (medium 1). Therefore, the storage controller may split up the single entry for medium 5 in table 300 into two range entries in table 600. The first new entry is for blocks 0-99 of medium 5, and this entry is still in the registered, unmasked (RU) state with underlying medium 2. The second new entry is for blocks 100-999 of medium 5, and this entry is in the quiescent, unmasked (QU) state with underlying medium 1. Lookups for blocks 100-999 of medium 5 may now be performed in a single underlying medium search (medium 1) rather than requiring the storage controller to traverse two underlying mediums (medium 2 and medium 1).
  • Another technique for collapsing the medium graph is to merge one or more mediums that cannot be referenced externally. These one or more mediums may be merged to the medium directly above them in the medium graph. For example, referring back to table 300 of FIG. 3, if medium 10 were no longer externally visible, the storage controller could merge medium 10 with medium 14 by coalescing entries in medium 10 into medium 14 and renumbering these entries. While the storage controller could merge the mappings in medium 14 into medium 25, the storage controller could not then delete medium 14 because medium 18 still uses medium 14 as its underlying medium, even though medium 14 might not be externally visible.
  • The combination of all of these techniques can dramatically shrink the number of mediums in use by a storage system. For storage systems that take frequent checkpoints and then forget them or lack external mappings to a portion of these checkpoints, these techniques can be especially beneficial for reducing the total number of mediums. In addition, for storage systems with long chains of mediums, these techniques can help ensure that lookups can be performed efficiently
  • Referring now to FIG. 7, one embodiment of a method 700 for shortcutting a medium mapping table is shown. The components embodied in system 100 described above (e.g., storage controller 110) may generally operate in accordance with method 700. In addition, the steps in this embodiment are shown in sequential order. However, some steps may occur in a different order than shown, some steps may be performed concurrently, some steps may be combined with other steps, and some steps may be absent in another embodiment.
  • The storage controller may initiate a process for shortcutting the medium mapping table (block 705). In one embodiment, this process for shortcutting the medium mapping table may only be one of multiple techniques for collapsing the medium graph and optimizing lookups that may be implemented. For example, this process for shortcutting may be combined with a process for consolidating entries in the medium mapping table (as described below in FIG. 8). Other techniques may also be combined with these methods to optimize lookups and reduce the number of mediums in the storage system.
  • Next, the storage controller may search for mediums to be read optimized (block 710). Next, the storage controller may select a medium range that has a status of registered (block 715). If a medium range has a registered status, this indicates that there are two or more levels of underlying medium ranges below this medium range. A status of registered also indicates that a search for blocks in that medium range may proceed recursively through multiple medium ranges. The medium range selected in block 715 may also be referred to as the original medium range for the purposes of this discussion. It is noted that in some cases, the selected medium may only have a single entry in the medium mapping table corresponding to a single range.
  • After selecting a medium range, the storage controller may go down a level in the medium mapping table (block 720). The operation of going down a level refers to finding the entry that corresponds to the medium range which underlies the selected medium range. It is noted that this underlying medium range may be broken up into multiple ranges and may have multiple entries in the medium mapping table corresponding to the selected medium range. For these cases, each range of the underlying medium range may be processed separately and the following steps of method 700 may continue in parallel for each of these ranges. Alternatively, the steps of method 700 may continue for a first entry corresponding to a first range, and when the steps of method 700 have been completed for the first entry, the second entry for the underlying medium range may be processed, then the third entry, and so on. Also, in some cases, an offset may be applied to the original block numbers when going down a level from the original medium range to the underlying medium range. After going down a level in the medium mapping table in block 720, the underlying medium range may now be referred to as the current medium range for the discussion pertaining to blocks 725-730 of method 700.
  • After block 720, the storage controller may determine if the current medium range is unmasked (conditional block 725). In one embodiment, this may be determined by checking the status of the current medium range in the medium mapping table. If the current medium range is unmasked (conditional block 725, “yes” leg), then method 700 may return to block 715 and the storage controller may select another medium range that has a registered status.
  • If the current medium is unmasked, this indicates that this medium has its own mappings for at least a portion of its associated data. Therefore, it is not possible to create a shortcut around this medium. Accordingly, the storage controller may select another medium range with a registered status. If the current medium range is masked (conditional block 725, “no” leg), then the storage controller may go down a level in the medium mapping table by going to the medium range which underlies the current medium range (block 730). If the current medium range is masked, this indicates that the current medium range does not have any mappings assigned to itself and that all mappings are assigned to lower-level medium ranges. Accordingly, a shortcutting of the medium mapping table can be implemented since all data blocks of the current medium range will end up being mapped through to a lower-level medium range. Method 700 continues after block 730 by determining how many levels can be shortcut from the table, since in some cases, more than one level may be shortcut from the medium mapping table. It is noted that when going down a level in the medium mapping table, this lower-level medium range may be split up into multiple ranges which correspond to the current medium range. In these cases, the remaining steps of method 700 may be performed separately for each range of this lower-level medium range. Each range may be handled in a parallel or in a serial fashion. After block 730, after going down a level in the medium mapping table, this underlying medium range (of the current medium range) may now be referred to as the current medium range for the remaining steps of method 700.
  • After block 730, the storage controller may determine if the current medium range has an underlying medium range (conditional block 735). If the current medium range has an underlying medium range (conditional block 735, “yes” leg), then the storage controller may determine if the current medium range is unmasked (conditional block 740). If the current medium range does not have an underlying medium range (conditional block 735, “no” leg), then the storage controller may create a new entry in the medium mapping table for the original medium range which points to the current medium range as its underlying medium range (block 745). This entry may be for a portion of the original medium range or for the entirety of the original medium range, depending on the circumstances detected when performing blocks 720 and 730. This new entry effectively shortcuts the medium mapping table and will make future lookups of the original medium range more efficient. The storage controller may use any of various techniques for determining when to reclaim the old entry which is being replaced by this new entry. After block 745, method 700 may return to block 715 to select a medium range with a registered status.
  • If the current medium range is unmasked (conditional block 740, “yes” leg), then the storage controller may create a new entry in the medium mapping table for the original medium range which points to the current medium range as its underlying medium range (block 745). If the current medium range is masked (conditional block 740, “no” leg), then method 700 may return to block 730 to go down to a lower level in the medium mapping table. In this case, a shortcut of the table may span more than one level if each of the two or more intermediate levels (between the original medium range and the current medium range) have an underlying medium range and are masked.
  • Turning now to FIG. 8, one embodiment of a method 800 for collapsing the medium mapping table is shown. The components embodied in system 100 described above (e.g., storage controller 110) may generally operate in accordance with method 800. In addition, the steps in this embodiment are shown in sequential order. However, some steps may occur in a different order than shown, some steps may be performed concurrently, some steps may be combined with other steps, and some steps may be absent in another embodiment.
  • The storage controller may initiate a medium mapping table collapsing process (block 805). The conditions which trigger the initiation of the medium mapping table collapsing process may vary from embodiment to embodiment. For example, in one embodiment, if the storage controller determines that the number of mediums exceeds a programmable threshold, then the medium mapping table collapsing may be initiated. Alternatively, if the storage controller has idle processing resources, then the storage controller may utilize these processing resources to run the medium mapping table collapsing process. In other embodiments, the storage controller may initiate the medium mapping table collapsing process on a periodic basis (e.g., daily, twice a week).
  • The medium mapping table collapsing process may select a medium for consolidating entries in the address translation table (block 810). In one embodiment, a medium with a long chain of underlying mediums may be selected for consolidating entries. For example, the storage controller may select a given medium for consolidation if the given medium has a chain of underlying mediums that is greater than a programmable threshold. Other techniques for selecting a medium for consolidation are possible and are contemplated.
  • Next, the storage controller may select a given range of blocks of the selected medium (block 815). The size of the range may vary depending on the characteristics of the selected medium. In one embodiment, the storage controller may begin with a first range of blocks starting with the smallest number and proceeding to higher numbers. Alternatively, the storage controller may begin with the highest range of blocks based on block number and proceed to lower numbers on successive searches. Alternatively, the storage controller may process the different ranges of blocks in any order.
  • If the selected medium is unmasked for the current range of blocks and if an address translation table lookup for the current range of blocks of the selected medium results in a hit (conditional block 820, “yes” leg), then the storage controller may determine if the last range of blocks for the selected medium has already been reached (conditional block 825). If the given range of blocks has an entry assigned to the selected medium in the index of the address translation table, then the storage controller does not need to consolidate this entry since the entry is already assigned to the selected medium and can be found using a key generated from the selected medium.
  • If the storage controller has already reached the last range of blocks for the selected medium (conditional block 825, “yes” leg), then method 800 may return to block 810 and select a new medium to consolidate. If the storage controller has not yet reached the last range of blocks for the selected medium (conditional block 825, “no” leg), then the storage controller may move to the next range of blocks of the selected medium for searching (block 830), and then method 800 may return to conditional block 820.
  • If the selected medium is masked for the current range of blocks or if the address translation table lookup for the current range of blocks using the selected medium results in a miss (conditional block 820, “no” leg), then the storage controller may determine the underlying medium of the selected medium (block 835). In one embodiment, the storage controller may query the medium mapping table to determine the underlying medium of the selected medium. The storage controller may also determine if an offset should be applied to the block number in the underlying medium from the original block number of the selected medium. Next, the storage controller may determine if the underlying medium is unmasked for the current range of blocks and if an address translation table lookup for the current range of blocks using the underlying medium results in a hit (conditional block 840). The storage controller may use a key generated from the underlying medium and current range of blocks for the address translation table lookup (with an offset applied to the block numbers, if applicable).
  • If the underlying medium is masked for the current range of blocks or if the address translation table lookup for the current range of blocks using the selected medium results in a miss (conditional block 840, “no” leg), then the storage controller may determine if the selected medium is quiescent (conditional block 845). If the selected medium is quiescent (conditional block 845, “yes” leg), then this indicates that there are no more underlying mediums to search, and method 800 may return to block 810 to select a new medium for consolidating entries. If the selected medium is not quiescent (conditional block 845, “no” leg), then the storage controller may return to block 835 to determine the underlying medium of the current medium.
  • If the underlying medium is unmasked for the current range of blocks and if an address translation table lookup for the current range of blocks using the underlying medium results in a hit (conditional block 840, “yes” leg), then the storage controller may create one or more new entries in the address translation table for the current range of blocks and assign the new entries to the selected medium (block 850). Each new entry may include a key generated from the selected medium ID and the range of blocks as mapped from the selected medium. This range of blocks may be different from the current range if an offset was applied to the underlying medium by the medium above it in the medium graph. Next, the storage controller may update the medium mapping table entry status to quiescent for the current range of blocks of the selected medium (block 855).
  • Next, the storage controller may copy mappings from the existing entries (assigned to the underlying medium) of the address translation table to the new entries while retaining the existing entries (block 860). However, in some embodiments, the storage controller may reclaim the existing entries for the underlying medium if the region of the underlying medium corresponding to the existing entries is not reachable by any other mediums. After block 860, method 800 may return to block 825 to determine if the last range of blocks has been reached for the selected medium.
  • In some embodiments, method 700 (of FIG. 7) and method 800 may be performed concurrently in various combinations to merge mappings and consolidate mediums. For example, method 700 may be performed on a first medium to shortcut the medium mapping table while method 800 is being performed in parallel on a second medium to collapse the medium mapping table for the second medium. Additionally, method 700 may be performed on multiple different mediums in parallel while method 800 is also performed on multiple different mediums in parallel. By combining method 700 and method 800 (and one or more other techniques), the number of mediums in the storage system may be reduced and future lookups to mediums may be optimized.
  • It is also noted that as part of the process of merging mappings and consolidating mediums, adjacent mappings may be merged if these adjacent mappings are compatible so as to further reduce the size of the medium mapping table. Adjacent mappings may be considered compatible if they have the same attributes and can be combined into a single entry in the medium mapping table. For example, a given medium 900 may have two adjacent entries. The first of these entries may be for a range of 0-99, with the underlying medium set to 905, and with a status of quiescent. The second entry may be for a range of 100-399 with the underlying medium set to 905, and with a status of quiescent. These two entries can be merged into a single with a range of 0-399, with the underlying medium set to 905, and with a status of quiescent. After a storage controller has promoted mappings to higher-level mediums while performing medium collapsing operations, it is customary to find many adjacent entries with the same attributes scattered throughout the medium mapping table. Accordingly, these adjacent entries may be merged to further reduce the size of the medium mapping table.
  • It is noted that the above-described embodiments may comprise software. In such an embodiment, the program instructions that implement the methods and/or mechanisms may be conveyed or stored on a computer readable medium. Numerous types of media which are configured to store program instructions are available and include hard disks, floppy disks, CD-ROM, DVD, flash memory, Programmable ROMs (PROM), random access memory (RAM), and various other forms of volatile or non-volatile storage.
  • In various embodiments, one or more portions of the methods and mechanisms described herein may form part of a cloud-computing environment. In such embodiments, resources may be provided over the Internet as services according to one or more various models. Such models may include Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS). In IaaS, computer infrastructure is delivered as a service. In such a case, the computing equipment is generally owned and operated by the service provider. In the PaaS model, software tools and underlying equipment used by developers to develop software solutions may be provided as a service and hosted by the service provider. SaaS typically includes a service provider licensing software as a service on demand. The service provider may host the software, or may deploy the software to a customer for a given period of time. Numerous combinations of the above models are possible and are contemplated.
  • Although the embodiments above have been described in considerable detail, numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.

Claims (20)

What is claimed is:
1. A computer system comprising:
one or more storage devices; and
a storage controller coupled to the one or more storage devices, wherein the storage controller is configured to:
maintain a medium mapping table to map underlying mediums to active mediums, wherein each underlying medium represents a snapshot; and
maintain an address translation table to map mediums to locations within the one or more storage devices;
wherein the storage controller is further configured to:
initiate a search optimization operation for a first range of a first medium, wherein the search optimization operation comprises collapsing the medium mapping table for the first range of the first medium, and wherein the first range of the first medium has at least one underlying medium;
create one or more first entries in the address translation table;
assign the one or more first entries to the first range of the first medium; and
copy attributes from one or more second entries in the address translation table to the first entries, wherein the second entries are assigned to a second medium, and wherein the second medium underlies the first range of the first medium.
2. The computer system as recited in claim 1, wherein responsive to determining a third medium underlies a second range of the first medium and a fourth medium underlies the third medium, the storage controller is configured to:
create a new entry for the second range of the first medium in the medium mapping table, wherein the new entry indicates the fourth medium underlies the second range of the first medium; and
discard an existing entry for the second range of the first medium in the medium mapping table, wherein the existing entry indicates the third medium underlies the second range of the first medium.
3. The computer system as recited in claim 2, wherein the storage controller is configured to create the new entry for the second range of the first medium in the medium mapping table responsive to detecting that there are no existing entries assigned to the third medium in the address translation table which correspond to an address space of the second range of the first medium.
4. The computer system as recited in claim 1, wherein the storage controller is further configured to copy attributes from one or more third entries of the address translation table to the first entries, wherein the third entries are assigned to a third medium, and wherein the third medium underlies the second medium.
5. The computer system as recited in claim 1, wherein the storage controller is further configured to apply an offset to block numbers copied to the first entries responsive to determining the first range of the first medium maps to the second medium using the offset.
6. The computer system as recited in claim 1, wherein the storage controller is configured to update the medium mapping table to indicate the first range of the first medium has no underlying medium responsive to determining mappings assigned to the first range of the first medium in the address translation table cover an entire address space of the first range of the first medium.
7. The computer system as recited in claim 1, wherein responsive to determining the entries in the second medium are no longer externally visible to any user volumes, the storage controller is configured to reclaim the second medium.
8. A method for use in a storage system, the method comprising:
maintaining a medium mapping table to map underlying mediums to active mediums, wherein each underlying medium represents a snapshot;
maintaining an address translation table to map mediums to locations within the one or more storage devices;
initiating a search optimization operation for a first range of a first medium, wherein the search optimization operation comprises collapsing the medium mapping table for the first range of the first medium, and wherein the first range of the first medium has at least one underlying medium;
creating one or more first entries in the address translation table;
assigning the one or more first entries to the first range of the first medium; and
copying attributes from one or more second entries in the address translation table to the first entries, wherein the second entries are assigned to a second medium, and wherein the second medium underlies the first range of the first medium
9. The method as recited in claim 8, wherein responsive to determining a third medium underlies a second range of the first medium and a fourth medium underlies the third medium, the method further comprising:
creating a new entry for the second range of the first medium in the medium mapping table, wherein the new entry indicates the fourth medium underlies the second range of the first medium; and
discarding an existing entry for the second range of the first medium in the medium mapping table, wherein the existing entry indicates the third medium underlies the second range of the first medium.
10. The method as recited in claim 9, further comprising creating the new entry for the second range of the first medium in the medium mapping table responsive to detecting that there are no existing entries assigned to the third medium in the address translation table which correspond to an address space of the second range of the first medium.
11. The method as recited in claim 8, further comprising copying attributes from one or more third entries of the address translation table to the first entries, wherein the third entries are assigned to a third medium, and wherein the third medium underlies the second medium.
12. The method as recited in claim 8, further comprising applying an offset to block numbers copied to the first entries responsive to determining the first range of the first medium maps to the second medium using the offset.
13. The method as recited in claim 8, further comprising updating the medium mapping table to indicate the first range of the first medium has no underlying medium responsive to determining mappings assigned to the first range of the first medium in the address translation table cover an entire address space of the first range of the first medium.
14. The method as recited in claim 8, further comprising reclaiming the second medium responsive to determining the entries in the second medium are no longer externally visible to any user volumes.
15. A non-transitory computer readable storage medium storing program instructions, wherein the program instructions are executable by a processor to
maintain a medium mapping table to map underlying mediums to active mediums, wherein each underlying medium represents a snapshot;
maintain an address translation table to map mediums to locations within the one or more storage devices;
initiate a search optimization operation for a first range of a first medium, wherein the search optimization operation comprises collapsing the medium mapping table for the first range of the first medium, and wherein the first range of the first medium has at least one underlying medium;
create one or more first entries in the address translation table;
assign the one or more first entries to the first range of the first medium; and
copy attributes from one or more second entries in the address translation table to the first entries, wherein the second entries are assigned to a second medium, and wherein the second medium underlies the first range of the first medium.
16. The computer readable storage medium as recited in claim 15, wherein responsive to determining a third medium underlies a second range of the first medium and a fourth medium underlies the third medium, the program instructions are executable by a processor to:
create a new entry for the second range of the first medium in the medium mapping table, wherein the new entry indicates the fourth medium underlies the second range of the first medium; and
discard an existing entry for the second range of the first medium in the medium mapping table, wherein the existing entry indicates the third medium underlies the second range of the first medium.
17. The computer readable storage medium as recited in claim 16, wherein the program instructions are further executable by a processor create the new entry for the second range of the first medium in the medium mapping table responsive to detecting that there are no existing entries assigned to the third medium in the address translation table which correspond to an address space of the second range of the first medium.
18. The computer readable storage medium as recited in claim 15, wherein the program instructions are further executable by a processor to copy attributes from one or more third entries of the address translation table to the first entries, wherein the third entries are assigned to a third medium, and wherein the third medium underlies the second medium.
19. The computer readable storage medium as recited in claim 15, wherein the program instructions are further executable by a processor to apply an offset to block numbers copied to the first entries responsive to determining the first range of the first medium maps to the second medium using the offset.
20. The computer readable storage medium as recited in claim 15, wherein the program instructions are further executable by a processor to update the medium mapping table to indicate the first range of the first medium has no underlying medium responsive to determining mappings assigned to the first range of the first medium in the address translation table cover an entire address space of the first range of the first medium.
US14/058,757 2013-01-10 2013-10-21 Optimizing snapshot lookups Abandoned US20140195551A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US14/058,757 US20140195551A1 (en) 2013-01-10 2013-10-21 Optimizing snapshot lookups
PCT/US2014/010885 WO2014110266A1 (en) 2013-01-10 2014-01-09 Optimizing snapshot lookups
US15/291,373 US10013317B1 (en) 2013-01-10 2016-10-12 Restoring a volume in a storage system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201361751142P 2013-01-10 2013-01-10
US14/058,757 US20140195551A1 (en) 2013-01-10 2013-10-21 Optimizing snapshot lookups

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/291,373 Continuation US10013317B1 (en) 2013-01-10 2016-10-12 Restoring a volume in a storage system

Publications (1)

Publication Number Publication Date
US20140195551A1 true US20140195551A1 (en) 2014-07-10

Family

ID=50030505

Family Applications (18)

Application Number Title Priority Date Filing Date
US14/046,872 Active 2033-12-24 US9063967B2 (en) 2013-01-10 2013-10-04 Performing copies in a storage system
US14/046,876 Expired - Fee Related US9589008B2 (en) 2013-01-10 2013-10-04 Deduplication of volume regions
US14/046,878 Active 2034-09-28 US9436720B2 (en) 2013-01-10 2013-10-04 Safety for volume operations
US14/046,870 Active 2035-06-06 US9646039B2 (en) 2013-01-10 2013-10-04 Snapshots in a storage system
US14/058,757 Abandoned US20140195551A1 (en) 2013-01-10 2013-10-21 Optimizing snapshot lookups
US14/737,927 Active US9361035B1 (en) 2013-01-10 2015-06-12 Performing copies in a storage system
US15/139,408 Active 2033-11-14 US9760313B1 (en) 2013-01-10 2016-04-27 Performing copies in a storage system
US15/291,373 Active US10013317B1 (en) 2013-01-10 2016-10-12 Restoring a volume in a storage system
US15/410,202 Active US9880779B1 (en) 2013-01-10 2017-01-19 Processing copy offload requests in a storage system
US15/416,665 Active US9891858B1 (en) 2013-01-10 2017-01-26 Deduplication of regions with a storage system
US15/484,243 Active US10235093B1 (en) 2013-01-10 2017-04-11 Restoring snapshots in a storage system
US15/875,189 Active 2033-11-11 US10585617B1 (en) 2013-01-10 2018-01-19 Buffering copy requests in a storage system
US16/276,451 Abandoned US20190179535A1 (en) 2013-01-10 2019-02-14 Efficient space utilization in a flash system using snapshots
US16/583,457 Active US11853584B1 (en) 2013-01-10 2019-09-26 Generating volume snapshots
US16/788,771 Active US11099769B1 (en) 2013-01-10 2020-02-12 Copying data without accessing the data
US17/406,421 Active US11662936B2 (en) 2013-01-10 2021-08-19 Writing data using references to previously stored data
US18/312,872 Pending US20230273743A1 (en) 2013-01-10 2023-05-05 Lightweight Copying Of Data Using Metadata References
US18/448,019 Pending US20230384963A1 (en) 2013-01-10 2023-08-10 Efficient Creation And Management Of Snapshots

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US14/046,872 Active 2033-12-24 US9063967B2 (en) 2013-01-10 2013-10-04 Performing copies in a storage system
US14/046,876 Expired - Fee Related US9589008B2 (en) 2013-01-10 2013-10-04 Deduplication of volume regions
US14/046,878 Active 2034-09-28 US9436720B2 (en) 2013-01-10 2013-10-04 Safety for volume operations
US14/046,870 Active 2035-06-06 US9646039B2 (en) 2013-01-10 2013-10-04 Snapshots in a storage system

Family Applications After (13)

Application Number Title Priority Date Filing Date
US14/737,927 Active US9361035B1 (en) 2013-01-10 2015-06-12 Performing copies in a storage system
US15/139,408 Active 2033-11-14 US9760313B1 (en) 2013-01-10 2016-04-27 Performing copies in a storage system
US15/291,373 Active US10013317B1 (en) 2013-01-10 2016-10-12 Restoring a volume in a storage system
US15/410,202 Active US9880779B1 (en) 2013-01-10 2017-01-19 Processing copy offload requests in a storage system
US15/416,665 Active US9891858B1 (en) 2013-01-10 2017-01-26 Deduplication of regions with a storage system
US15/484,243 Active US10235093B1 (en) 2013-01-10 2017-04-11 Restoring snapshots in a storage system
US15/875,189 Active 2033-11-11 US10585617B1 (en) 2013-01-10 2018-01-19 Buffering copy requests in a storage system
US16/276,451 Abandoned US20190179535A1 (en) 2013-01-10 2019-02-14 Efficient space utilization in a flash system using snapshots
US16/583,457 Active US11853584B1 (en) 2013-01-10 2019-09-26 Generating volume snapshots
US16/788,771 Active US11099769B1 (en) 2013-01-10 2020-02-12 Copying data without accessing the data
US17/406,421 Active US11662936B2 (en) 2013-01-10 2021-08-19 Writing data using references to previously stored data
US18/312,872 Pending US20230273743A1 (en) 2013-01-10 2023-05-05 Lightweight Copying Of Data Using Metadata References
US18/448,019 Pending US20230384963A1 (en) 2013-01-10 2023-08-10 Efficient Creation And Management Of Snapshots

Country Status (6)

Country Link
US (18) US9063967B2 (en)
EP (3) EP2943872A1 (en)
JP (4) JP6543195B2 (en)
KR (3) KR20150104604A (en)
CN (3) CN105027068B (en)
WO (5) WO2014110137A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160105502A1 (en) * 2013-06-20 2016-04-14 Huawei Technologies Co., Ltd. Data synchronization method, data synchronization apparatus, and distributed system
WO2016094145A1 (en) * 2014-12-11 2016-06-16 Pure Storage, Inc. Cloud alert to replica
CN106155569A (en) * 2015-04-07 2016-11-23 北京中科同向信息技术有限公司 A kind of storage snapping technique based on mirror image
US9823842B2 (en) 2014-05-12 2017-11-21 The Research Foundation For The State University Of New York Gang migration of virtual machines using cluster-wide deduplication
US10922240B2 (en) 2018-09-19 2021-02-16 Toshiba Memory Corporation Memory system, storage system and method of controlling the memory system

Families Citing this family (374)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8484162B2 (en) 2008-06-24 2013-07-09 Commvault Systems, Inc. De-duplication systems and methods for application-specific data
US8930306B1 (en) 2009-07-08 2015-01-06 Commvault Systems, Inc. Synchronized data deduplication
US11449394B2 (en) 2010-06-04 2022-09-20 Commvault Systems, Inc. Failover systems and methods for performing backup operations, including heterogeneous indexing and load balancing of backup and indexing resources
US8589625B2 (en) 2010-09-15 2013-11-19 Pure Storage, Inc. Scheduling of reconstructive I/O read operations in a storage environment
US8468318B2 (en) 2010-09-15 2013-06-18 Pure Storage Inc. Scheduling of I/O writes in a storage environment
US11275509B1 (en) 2010-09-15 2022-03-15 Pure Storage, Inc. Intelligently sizing high latency I/O requests in a storage environment
US11614893B2 (en) 2010-09-15 2023-03-28 Pure Storage, Inc. Optimizing storage device access based on latency
US8732426B2 (en) 2010-09-15 2014-05-20 Pure Storage, Inc. Scheduling of reactive I/O operations in a storage environment
US8589655B2 (en) 2010-09-15 2013-11-19 Pure Storage, Inc. Scheduling of I/O in an SSD environment
US8775868B2 (en) 2010-09-28 2014-07-08 Pure Storage, Inc. Adaptive RAID for an SSD environment
US9244769B2 (en) 2010-09-28 2016-01-26 Pure Storage, Inc. Offset protection data in a RAID array
US20120150818A1 (en) 2010-12-14 2012-06-14 Commvault Systems, Inc. Client-side repository in a networked deduplicated storage system
US9020900B2 (en) 2010-12-14 2015-04-28 Commvault Systems, Inc. Distributed deduplicated storage system
US8589640B2 (en) 2011-10-14 2013-11-19 Pure Storage, Inc. Method for maintaining multiple fingerprint tables in a deduplicating storage system
US11636031B2 (en) 2011-08-11 2023-04-25 Pure Storage, Inc. Optimized inline deduplication
US8719540B1 (en) * 2012-03-15 2014-05-06 Pure Storage, Inc. Fractal layout of data blocks across multiple devices
US9218376B2 (en) 2012-06-13 2015-12-22 Commvault Systems, Inc. Intelligent data sourcing in a networked storage system
US11032259B1 (en) 2012-09-26 2021-06-08 Pure Storage, Inc. Data protection in a storage system
US8745415B2 (en) 2012-09-26 2014-06-03 Pure Storage, Inc. Multi-drive cooperation to generate an encryption key
US10623386B1 (en) 2012-09-26 2020-04-14 Pure Storage, Inc. Secret sharing data protection in a storage system
US10908835B1 (en) * 2013-01-10 2021-02-02 Pure Storage, Inc. Reversing deletion of a virtual machine
US9063967B2 (en) 2013-01-10 2015-06-23 Pure Storage, Inc. Performing copies in a storage system
US11733908B2 (en) 2013-01-10 2023-08-22 Pure Storage, Inc. Delaying deletion of a dataset
US11768623B2 (en) 2013-01-10 2023-09-26 Pure Storage, Inc. Optimizing generalized transfers between storage systems
US9633033B2 (en) 2013-01-11 2017-04-25 Commvault Systems, Inc. High availability distributed deduplicated storage system
US10263770B2 (en) 2013-11-06 2019-04-16 Pure Storage, Inc. Data protection in a storage system using external secrets
US11630585B1 (en) 2016-08-25 2023-04-18 Pure Storage, Inc. Processing evacuation events in a storage array that includes a plurality of storage devices
US10365858B2 (en) 2013-11-06 2019-07-30 Pure Storage, Inc. Thin provisioning in a storage device
US11128448B1 (en) 2013-11-06 2021-09-21 Pure Storage, Inc. Quorum-aware secret sharing
US9516016B2 (en) 2013-11-11 2016-12-06 Pure Storage, Inc. Storage array password management
US9208086B1 (en) 2014-01-09 2015-12-08 Pure Storage, Inc. Using frequency domain to prioritize storage of metadata in a cache
US10628353B2 (en) 2014-03-08 2020-04-21 Diamanti, Inc. Enabling use of non-volatile media-express (NVMe) over a network
US11921658B2 (en) 2014-03-08 2024-03-05 Diamanti, Inc. Enabling use of non-volatile media-express (NVMe) over a network
US10380072B2 (en) 2014-03-17 2019-08-13 Commvault Systems, Inc. Managing deletions from a deduplication database
US10656864B2 (en) 2014-03-20 2020-05-19 Pure Storage, Inc. Data replication within a flash storage array
JP6260401B2 (en) * 2014-03-28 2018-01-17 富士通株式会社 Information processing system, information processing system control method, and information processing apparatus control program
US9563518B2 (en) 2014-04-02 2017-02-07 Commvault Systems, Inc. Information management by a media agent in the absence of communications with a storage manager
US9513820B1 (en) 2014-04-07 2016-12-06 Pure Storage, Inc. Dynamically controlling temporary compromise on data redundancy
US9779268B1 (en) 2014-06-03 2017-10-03 Pure Storage, Inc. Utilizing a non-repeating identifier to encrypt data
US9218244B1 (en) 2014-06-04 2015-12-22 Pure Storage, Inc. Rebuilding data across storage nodes
US11399063B2 (en) 2014-06-04 2022-07-26 Pure Storage, Inc. Network authentication for a storage system
US9218407B1 (en) * 2014-06-25 2015-12-22 Pure Storage, Inc. Replication and intermediate read-write state for mediums
US10496556B1 (en) 2014-06-25 2019-12-03 Pure Storage, Inc. Dynamic data protection within a flash storage system
US10296469B1 (en) 2014-07-24 2019-05-21 Pure Storage, Inc. Access control in a flash storage system
US9558069B2 (en) 2014-08-07 2017-01-31 Pure Storage, Inc. Failure mapping in a storage array
US9495255B2 (en) 2014-08-07 2016-11-15 Pure Storage, Inc. Error recovery in a storage cluster
US9864761B1 (en) 2014-08-08 2018-01-09 Pure Storage, Inc. Read optimization operations in a storage system
CN105376285A (en) * 2014-08-29 2016-03-02 纬创资通股份有限公司 Network storage deduplicating method and server
US10380026B2 (en) * 2014-09-04 2019-08-13 Sandisk Technologies Llc Generalized storage virtualization interface
US10430079B2 (en) 2014-09-08 2019-10-01 Pure Storage, Inc. Adjusting storage capacity in a computing system
US10031934B2 (en) * 2014-09-30 2018-07-24 International Business Machines Corporation Deleting tuples using separate transaction identifier storage
US10164841B2 (en) 2014-10-02 2018-12-25 Pure Storage, Inc. Cloud assist for storage systems
US20160100004A1 (en) 2014-10-06 2016-04-07 International Business Machines Corporation Data replication across servers
US9489132B2 (en) 2014-10-07 2016-11-08 Pure Storage, Inc. Utilizing unmapped and unknown states in a replicated storage system
US10430282B2 (en) 2014-10-07 2019-10-01 Pure Storage, Inc. Optimizing replication by distinguishing user and system write activity
US9575673B2 (en) 2014-10-29 2017-02-21 Commvault Systems, Inc. Accessing a file system using tiered deduplication
US9727485B1 (en) 2014-11-24 2017-08-08 Pure Storage, Inc. Metadata rewrite and flatten optimization
US10089011B1 (en) * 2014-11-25 2018-10-02 Scale Computing Zero memory buffer copying in a reliable distributed computing system
US9773007B1 (en) * 2014-12-01 2017-09-26 Pure Storage, Inc. Performance improvements in a storage system
US9588842B1 (en) 2014-12-11 2017-03-07 Pure Storage, Inc. Drive rebuild
US9864769B2 (en) 2014-12-12 2018-01-09 Pure Storage, Inc. Storing data utilizing repeating pattern detection
US10545987B2 (en) 2014-12-19 2020-01-28 Pure Storage, Inc. Replication to the cloud
US9569357B1 (en) 2015-01-08 2017-02-14 Pure Storage, Inc. Managing compressed data in a storage system
US10296354B1 (en) 2015-01-21 2019-05-21 Pure Storage, Inc. Optimized boot operations within a flash storage array
US11947968B2 (en) 2015-01-21 2024-04-02 Pure Storage, Inc. Efficient use of zone in a storage device
US10437784B2 (en) * 2015-01-30 2019-10-08 SK Hynix Inc. Method and system for endurance enhancing, deferred deduplication with hardware-hash-enabled storage device
US9715351B2 (en) * 2015-02-13 2017-07-25 Red Hat, Inc. Copy-offload on a device stack
US9710165B1 (en) 2015-02-18 2017-07-18 Pure Storage, Inc. Identifying volume candidates for space reclamation
US9880755B2 (en) 2015-02-25 2018-01-30 Western Digital Technologies, Inc. System and method for copy on write on an SSD
US10082985B2 (en) 2015-03-27 2018-09-25 Pure Storage, Inc. Data striping across storage nodes that are assigned to multiple logical arrays
US10178169B2 (en) 2015-04-09 2019-01-08 Pure Storage, Inc. Point to point based backend communication layer for storage processing
US10339106B2 (en) 2015-04-09 2019-07-02 Commvault Systems, Inc. Highly reusable deduplication database after disaster recovery
US10140149B1 (en) 2015-05-19 2018-11-27 Pure Storage, Inc. Transactional commits with hardware assists in remote memory
US20160350391A1 (en) 2015-05-26 2016-12-01 Commvault Systems, Inc. Replication using deduplicated secondary copy data
US9716755B2 (en) 2015-05-26 2017-07-25 Pure Storage, Inc. Providing cloud storage array services by a local storage array in a data center
US11102298B1 (en) 2015-05-26 2021-08-24 Pure Storage, Inc. Locally providing cloud storage services for fleet management
US9594678B1 (en) 2015-05-27 2017-03-14 Pure Storage, Inc. Preventing duplicate entries of identical data in a storage device
US9444822B1 (en) 2015-05-29 2016-09-13 Pure Storage, Inc. Storage array access control from cloud-based user authorization and authentication
US11503031B1 (en) 2015-05-29 2022-11-15 Pure Storage, Inc. Storage array access control from cloud-based user authorization and authentication
US10021170B2 (en) 2015-05-29 2018-07-10 Pure Storage, Inc. Managing a storage array using client-side services
US9300660B1 (en) 2015-05-29 2016-03-29 Pure Storage, Inc. Providing authorization and authentication in a cloud for a user of a storage array
US9588691B2 (en) 2015-06-10 2017-03-07 Pure Storage, Inc. Dynamically managing control information in a storage device
US9696931B2 (en) 2015-06-12 2017-07-04 International Business Machines Corporation Region-based storage for volume data and metadata
US9594512B1 (en) 2015-06-19 2017-03-14 Pure Storage, Inc. Attributing consumed storage capacity among entities storing data in a storage array
US10310740B2 (en) 2015-06-23 2019-06-04 Pure Storage, Inc. Aligning memory access operations to a geometry of a storage device
US9547441B1 (en) 2015-06-23 2017-01-17 Pure Storage, Inc. Exposing a geometry of a storage device
US10296236B2 (en) 2015-07-01 2019-05-21 Pure Storage, Inc. Offloading device management responsibilities from a storage device in an array of storage devices
US9892071B2 (en) 2015-08-03 2018-02-13 Pure Storage, Inc. Emulating a remote direct memory access (‘RDMA’) link between controllers in a storage array
US9851762B1 (en) 2015-08-06 2017-12-26 Pure Storage, Inc. Compliant printed circuit board (‘PCB’) within an enclosure
US9697034B2 (en) 2015-08-07 2017-07-04 Futurewei Technologies, Inc. Offloading probabilistic computations in data analytics applications
US11625181B1 (en) 2015-08-24 2023-04-11 Pure Storage, Inc. Data tiering using snapshots
US10198194B2 (en) 2015-08-24 2019-02-05 Pure Storage, Inc. Placing data within a storage device of a flash array
US11294588B1 (en) 2015-08-24 2022-04-05 Pure Storage, Inc. Placing data within a storage device
KR20170028825A (en) 2015-09-04 2017-03-14 퓨어 스토리지, 아이앤씨. Memory-efficient storage and searching in hash tables using compressed indexes
US11269884B2 (en) 2015-09-04 2022-03-08 Pure Storage, Inc. Dynamically resizable structures for approximate membership queries
US11341136B2 (en) 2015-09-04 2022-05-24 Pure Storage, Inc. Dynamically resizable structures for approximate membership queries
US10503653B2 (en) * 2015-09-11 2019-12-10 Toshiba Memory Corporation Memory system
US9843453B2 (en) 2015-10-23 2017-12-12 Pure Storage, Inc. Authorizing I/O commands with I/O tokens
US10514978B1 (en) 2015-10-23 2019-12-24 Pure Storage, Inc. Automatic deployment of corrective measures for storage arrays
US11360844B1 (en) 2015-10-23 2022-06-14 Pure Storage, Inc. Recovery of a container storage provider
US9384082B1 (en) 2015-10-23 2016-07-05 Pure Storage, Inc. Proactively providing corrective measures for storage arrays
US10284232B2 (en) 2015-10-28 2019-05-07 Pure Storage, Inc. Dynamic error processing in a storage device
US10374868B2 (en) 2015-10-29 2019-08-06 Pure Storage, Inc. Distributed command processing in a flash storage system
US9740414B2 (en) * 2015-10-29 2017-08-22 Pure Storage, Inc. Optimizing copy operations
US10353777B2 (en) 2015-10-30 2019-07-16 Pure Storage, Inc. Ensuring crash-safe forward progress of a system configuration update
US9760479B2 (en) 2015-12-02 2017-09-12 Pure Storage, Inc. Writing data in a storage system that includes a first type of storage device and a second type of storage device
US11762764B1 (en) 2015-12-02 2023-09-19 Pure Storage, Inc. Writing data in a storage system that includes a first type of storage device and a second type of storage device
US10326836B2 (en) 2015-12-08 2019-06-18 Pure Storage, Inc. Partially replicating a snapshot between storage systems
US11616834B2 (en) 2015-12-08 2023-03-28 Pure Storage, Inc. Efficient replication of a dataset to the cloud
US11347697B1 (en) 2015-12-15 2022-05-31 Pure Storage, Inc. Proactively optimizing a storage system
US10162835B2 (en) 2015-12-15 2018-12-25 Pure Storage, Inc. Proactive management of a plurality of storage arrays in a multi-array system
US10346043B2 (en) 2015-12-28 2019-07-09 Pure Storage, Inc. Adaptive computing for data compression
US20170193003A1 (en) 2015-12-30 2017-07-06 Commvault Systems, Inc. Redundant and robust distributed deduplication data storage system
US9886314B2 (en) 2016-01-28 2018-02-06 Pure Storage, Inc. Placing workloads in a multi-array system
US10572460B2 (en) 2016-02-11 2020-02-25 Pure Storage, Inc. Compressing data in dependence upon characteristics of a storage system
US9760297B2 (en) 2016-02-12 2017-09-12 Pure Storage, Inc. Managing input/output (‘I/O’) queues in a data storage system
EP3286648B1 (en) * 2016-02-12 2020-08-26 Hewlett-Packard Enterprise Development LP Assembling operating system volumes
US10372543B2 (en) 2016-03-04 2019-08-06 Toshiba Memory Corporation Memory system
US9959043B2 (en) 2016-03-16 2018-05-01 Pure Storage, Inc. Performing a non-disruptive upgrade of data in a storage system
US11016955B2 (en) 2016-04-15 2021-05-25 Hitachi Vantara Llc Deduplication index enabling scalability
US11112990B1 (en) 2016-04-27 2021-09-07 Pure Storage, Inc. Managing storage device evacuation
US9841921B2 (en) 2016-04-27 2017-12-12 Pure Storage, Inc. Migrating data in a storage array that includes a plurality of storage devices
US11809727B1 (en) 2016-04-27 2023-11-07 Pure Storage, Inc. Predicting failures in a storage system that includes a plurality of storage devices
US9811264B1 (en) 2016-04-28 2017-11-07 Pure Storage, Inc. Deploying client-specific applications in a storage system utilizing redundant system resources
US10133503B1 (en) 2016-05-02 2018-11-20 Pure Storage, Inc. Selecting a deduplication process based on a difference between performance metrics
US10303390B1 (en) 2016-05-02 2019-05-28 Pure Storage, Inc. Resolving fingerprint collisions in flash storage system
US10452297B1 (en) 2016-05-02 2019-10-22 Pure Storage, Inc. Generating and optimizing summary index levels in a deduplication storage system
US11231858B2 (en) 2016-05-19 2022-01-25 Pure Storage, Inc. Dynamically configuring a storage system to facilitate independent scaling of resources
US9507532B1 (en) 2016-05-20 2016-11-29 Pure Storage, Inc. Migrating data in a storage array that includes a plurality of storage devices and a plurality of write buffer devices
US10691567B2 (en) 2016-06-03 2020-06-23 Pure Storage, Inc. Dynamically forming a failure domain in a storage system that includes a plurality of blades
US11706895B2 (en) 2016-07-19 2023-07-18 Pure Storage, Inc. Independent scaling of compute resources and storage resources in a storage system
US10203903B2 (en) 2016-07-26 2019-02-12 Pure Storage, Inc. Geometry based, space aware shelf/writegroup evacuation
US10459652B2 (en) 2016-07-27 2019-10-29 Pure Storage, Inc. Evacuating blades in a storage array that includes a plurality of blades
US10474363B1 (en) 2016-07-29 2019-11-12 Pure Storage, Inc. Space reporting in a storage system
US11960348B2 (en) 2016-09-07 2024-04-16 Pure Storage, Inc. Cloud-based monitoring of hardware components in a fleet of storage systems
US10671439B1 (en) 2016-09-07 2020-06-02 Pure Storage, Inc. Workload planning with quality-of-service (‘QOS’) integration
US11531577B1 (en) 2016-09-07 2022-12-20 Pure Storage, Inc. Temporarily limiting access to a storage device
US10235229B1 (en) 2016-09-07 2019-03-19 Pure Storage, Inc. Rehabilitating storage devices in a storage array that includes a plurality of storage devices
US11886922B2 (en) 2016-09-07 2024-01-30 Pure Storage, Inc. Scheduling input/output operations for a storage system
US11481261B1 (en) 2016-09-07 2022-10-25 Pure Storage, Inc. Preventing extended latency in a storage system
US10908966B1 (en) 2016-09-07 2021-02-02 Pure Storage, Inc. Adapting target service times in a storage system
US10146585B2 (en) 2016-09-07 2018-12-04 Pure Storage, Inc. Ensuring the fair utilization of system resources using workload based, time-independent scheduling
US10331588B2 (en) 2016-09-07 2019-06-25 Pure Storage, Inc. Ensuring the appropriate utilization of system resources using weighted workload based, time-independent scheduling
US10747630B2 (en) 2016-09-30 2020-08-18 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, including operations by a master monitor node
US10613974B2 (en) 2016-10-04 2020-04-07 Pure Storage, Inc. Peer-to-peer non-volatile random-access memory
US10162523B2 (en) 2016-10-04 2018-12-25 Pure Storage, Inc. Migrating data between volumes using virtual copy operation
US10756816B1 (en) 2016-10-04 2020-08-25 Pure Storage, Inc. Optimized fibre channel and non-volatile memory express access
US10191662B2 (en) 2016-10-04 2019-01-29 Pure Storage, Inc. Dynamic allocation of segments in a flash storage system
US10481798B2 (en) 2016-10-28 2019-11-19 Pure Storage, Inc. Efficient flash management for multiple controllers
US10007459B2 (en) 2016-10-20 2018-06-26 Pure Storage, Inc. Performance tuning in a storage system that includes one or more storage devices
US11379132B1 (en) 2016-10-20 2022-07-05 Pure Storage, Inc. Correlating medical sensor data
US10185505B1 (en) 2016-10-28 2019-01-22 Pure Storage, Inc. Reading a portion of data to replicate a volume based on sequence numbers
US10359942B2 (en) 2016-10-31 2019-07-23 Pure Storage, Inc. Deduplication aware scalable content placement
CN106527987B (en) * 2016-11-04 2019-06-04 湖南国科微电子股份有限公司 A kind of SSD master control reliability lifting system and method without DRAM
US10162566B2 (en) 2016-11-22 2018-12-25 Pure Storage, Inc. Accumulating application-level statistics in a storage system
US11620075B2 (en) 2016-11-22 2023-04-04 Pure Storage, Inc. Providing application aware storage
US11550481B2 (en) 2016-12-19 2023-01-10 Pure Storage, Inc. Efficiently writing data in a zoned drive storage system
US10452290B2 (en) 2016-12-19 2019-10-22 Pure Storage, Inc. Block consolidation in a direct-mapped flash storage system
US10198205B1 (en) 2016-12-19 2019-02-05 Pure Storage, Inc. Dynamically adjusting a number of storage devices utilized to simultaneously service write operations
US11461273B1 (en) 2016-12-20 2022-10-04 Pure Storage, Inc. Modifying storage distribution in a storage system that includes one or more storage devices
US10489307B2 (en) 2017-01-05 2019-11-26 Pure Storage, Inc. Periodically re-encrypting user data stored on a storage device
US11307998B2 (en) 2017-01-09 2022-04-19 Pure Storage, Inc. Storage efficiency of encrypted host system data
US11093146B2 (en) 2017-01-12 2021-08-17 Pure Storage, Inc. Automatic load rebalancing of a write group
US10503700B1 (en) 2017-01-19 2019-12-10 Pure Storage, Inc. On-demand content filtering of snapshots within a storage system
US11340800B1 (en) 2017-01-19 2022-05-24 Pure Storage, Inc. Content masking in a storage system
US11163624B2 (en) 2017-01-27 2021-11-02 Pure Storage, Inc. Dynamically adjusting an amount of log data generated for a storage system
US11169727B1 (en) 2017-03-10 2021-11-09 Pure Storage, Inc. Synchronous replication between storage systems with virtualized storage
US11803453B1 (en) 2017-03-10 2023-10-31 Pure Storage, Inc. Using host connectivity states to avoid queuing I/O requests
US10503427B2 (en) 2017-03-10 2019-12-10 Pure Storage, Inc. Synchronously replicating datasets and other managed objects to cloud-based storage systems
US11675520B2 (en) 2017-03-10 2023-06-13 Pure Storage, Inc. Application replication among storage systems synchronously replicating a dataset
US11442825B2 (en) 2017-03-10 2022-09-13 Pure Storage, Inc. Establishing a synchronous replication relationship between two or more storage systems
US11941279B2 (en) 2017-03-10 2024-03-26 Pure Storage, Inc. Data path virtualization
US10454810B1 (en) 2017-03-10 2019-10-22 Pure Storage, Inc. Managing host definitions across a plurality of storage systems
US10521344B1 (en) 2017-03-10 2019-12-31 Pure Storage, Inc. Servicing input/output (‘I/O’) operations directed to a dataset that is synchronized across a plurality of storage systems
US11089105B1 (en) 2017-12-14 2021-08-10 Pure Storage, Inc. Synchronously replicating datasets in cloud-based storage systems
US10528488B1 (en) 2017-03-30 2020-01-07 Pure Storage, Inc. Efficient name coding
US9910618B1 (en) 2017-04-10 2018-03-06 Pure Storage, Inc. Migrating applications executing on a storage system
US10459664B1 (en) 2017-04-10 2019-10-29 Pure Storage, Inc. Virtualized copy-by-reference
US11403019B2 (en) 2017-04-21 2022-08-02 Pure Storage, Inc. Deduplication-aware per-tenant encryption
US10944671B2 (en) 2017-04-27 2021-03-09 Pure Storage, Inc. Efficient data forwarding in a networked device
CN107168648B (en) * 2017-05-04 2021-03-02 Oppo广东移动通信有限公司 File storage method and device and terminal
US11868629B1 (en) 2017-05-05 2024-01-09 Pure Storage, Inc. Storage system sizing service
US10789020B2 (en) 2017-06-12 2020-09-29 Pure Storage, Inc. Recovering data within a unified storage element
US10417092B2 (en) 2017-09-07 2019-09-17 Pure Storage, Inc. Incremental RAID stripe update parity calculation
US10552090B2 (en) 2017-09-07 2020-02-04 Pure Storage, Inc. Solid state drives with multiple types of addressable memory
US10853148B1 (en) 2017-06-12 2020-12-01 Pure Storage, Inc. Migrating workloads between a plurality of execution environments
US11442669B1 (en) 2018-03-15 2022-09-13 Pure Storage, Inc. Orchestrating a virtual storage system
US10613791B2 (en) 2017-06-12 2020-04-07 Pure Storage, Inc. Portable snapshot replication between storage systems
US11592991B2 (en) 2017-09-07 2023-02-28 Pure Storage, Inc. Converting raid data between persistent storage types
US11609718B1 (en) 2017-06-12 2023-03-21 Pure Storage, Inc. Identifying valid data after a storage system recovery
US10976962B2 (en) 2018-03-15 2021-04-13 Pure Storage, Inc. Servicing I/O operations in a cloud-based storage system
US11422731B1 (en) 2017-06-12 2022-08-23 Pure Storage, Inc. Metadata-based replication of a dataset
US11016824B1 (en) 2017-06-12 2021-05-25 Pure Storage, Inc. Event identification with out-of-order reporting in a cloud-based environment
US11210133B1 (en) 2017-06-12 2021-12-28 Pure Storage, Inc. Workload mobility between disparate execution environments
US10884636B1 (en) 2017-06-12 2021-01-05 Pure Storage, Inc. Presenting workload performance in a storage system
US11340939B1 (en) 2017-06-12 2022-05-24 Pure Storage, Inc. Application-aware analytics for storage systems
CN110720088A (en) 2017-06-12 2020-01-21 净睿存储股份有限公司 Accessible fast durable storage integrated into mass storage device
US11561714B1 (en) 2017-07-05 2023-01-24 Pure Storage, Inc. Storage efficiency driven migration
US11477280B1 (en) 2017-07-26 2022-10-18 Pure Storage, Inc. Integrating cloud storage services
US10402266B1 (en) 2017-07-31 2019-09-03 Pure Storage, Inc. Redundant array of independent disks in a direct-mapped flash storage system
US10831935B2 (en) 2017-08-31 2020-11-10 Pure Storage, Inc. Encryption management with host-side data reduction
US10776202B1 (en) 2017-09-22 2020-09-15 Pure Storage, Inc. Drive, blade, or data shard decommission via RAID geometry shrinkage
US10789211B1 (en) 2017-10-04 2020-09-29 Pure Storage, Inc. Feature-based deduplication
US10671434B1 (en) 2017-10-19 2020-06-02 Pure Storage, Inc. Storage based artificial intelligence infrastructure
US11455168B1 (en) 2017-10-19 2022-09-27 Pure Storage, Inc. Batch building for deep learning training workloads
US10452444B1 (en) 2017-10-19 2019-10-22 Pure Storage, Inc. Storage system with compute resources and shared storage resources
US11861423B1 (en) 2017-10-19 2024-01-02 Pure Storage, Inc. Accelerating artificial intelligence (‘AI’) workflows
US11494692B1 (en) 2018-03-26 2022-11-08 Pure Storage, Inc. Hyperscale artificial intelligence and machine learning infrastructure
US10360214B2 (en) 2017-10-19 2019-07-23 Pure Storage, Inc. Ensuring reproducibility in an artificial intelligence infrastructure
US11126451B2 (en) 2017-10-24 2021-09-21 Hewlett Packard Enterprise Development Lp Converting virtual volumes in place
US10884919B2 (en) 2017-10-31 2021-01-05 Pure Storage, Inc. Memory management in a storage system
US10671494B1 (en) 2017-11-01 2020-06-02 Pure Storage, Inc. Consistent selection of replicated datasets during storage system recovery
US10509581B1 (en) 2017-11-01 2019-12-17 Pure Storage, Inc. Maintaining write consistency in a multi-threaded storage system
US10467107B1 (en) 2017-11-01 2019-11-05 Pure Storage, Inc. Maintaining metadata resiliency among storage device failures
US10484174B1 (en) 2017-11-01 2019-11-19 Pure Storage, Inc. Protecting an encryption key for data stored in a storage system that includes a plurality of storage devices
US10817392B1 (en) 2017-11-01 2020-10-27 Pure Storage, Inc. Ensuring resiliency to storage device failures in a storage system that includes a plurality of storage devices
US10860475B1 (en) 2017-11-17 2020-12-08 Pure Storage, Inc. Hybrid flash translation layer
US10929226B1 (en) 2017-11-21 2021-02-23 Pure Storage, Inc. Providing for increased flexibility for large scale parity
US10990282B1 (en) 2017-11-28 2021-04-27 Pure Storage, Inc. Hybrid data tiering with cloud storage
US10936238B2 (en) 2017-11-28 2021-03-02 Pure Storage, Inc. Hybrid data tiering
US10795598B1 (en) 2017-12-07 2020-10-06 Pure Storage, Inc. Volume migration for storage systems synchronously replicating a dataset
US11036677B1 (en) 2017-12-14 2021-06-15 Pure Storage, Inc. Replicated data integrity
US10929031B2 (en) 2017-12-21 2021-02-23 Pure Storage, Inc. Maximizing data reduction in a partially encrypted volume
WO2019128166A1 (en) * 2017-12-29 2019-07-04 北京忆恒创源科技有限公司 Kv storage device and method of using kv storage device to provide file system
US10970395B1 (en) 2018-01-18 2021-04-06 Pure Storage, Inc Security threat monitoring for a storage system
US11144638B1 (en) 2018-01-18 2021-10-12 Pure Storage, Inc. Method for storage system detection and alerting on potential malicious action
US11010233B1 (en) 2018-01-18 2021-05-18 Pure Storage, Inc Hardware-based system monitoring
US10992533B1 (en) 2018-01-30 2021-04-27 Pure Storage, Inc. Policy based path management
US10467527B1 (en) 2018-01-31 2019-11-05 Pure Storage, Inc. Method and apparatus for artificial intelligence acceleration
US11036596B1 (en) 2018-02-18 2021-06-15 Pure Storage, Inc. System for delaying acknowledgements on open NAND locations until durability has been confirmed
US11494109B1 (en) 2018-02-22 2022-11-08 Pure Storage, Inc. Erase block trimming for heterogenous flash memory storage devices
US11150834B1 (en) 2018-03-05 2021-10-19 Pure Storage, Inc. Determining storage consumption in a storage system
US11861170B2 (en) 2018-03-05 2024-01-02 Pure Storage, Inc. Sizing resources for a replication target
US10521151B1 (en) 2018-03-05 2019-12-31 Pure Storage, Inc. Determining effective space utilization in a storage system
US10942650B1 (en) 2018-03-05 2021-03-09 Pure Storage, Inc. Reporting capacity utilization in a storage system
US10296258B1 (en) 2018-03-09 2019-05-21 Pure Storage, Inc. Offloading data storage to a decentralized storage network
US11210009B1 (en) 2018-03-15 2021-12-28 Pure Storage, Inc. Staging data in a cloud-based storage system
US11288138B1 (en) 2018-03-15 2022-03-29 Pure Storage, Inc. Recovery from a system fault in a cloud-based storage system
US10917471B1 (en) 2018-03-15 2021-02-09 Pure Storage, Inc. Active membership in a cloud-based storage system
US10924548B1 (en) 2018-03-15 2021-02-16 Pure Storage, Inc. Symmetric storage using a cloud-based storage system
US11048590B1 (en) 2018-03-15 2021-06-29 Pure Storage, Inc. Data consistency during recovery in a cloud-based storage system
US11171950B1 (en) 2018-03-21 2021-11-09 Pure Storage, Inc. Secure cloud-based storage system management
US11095706B1 (en) 2018-03-21 2021-08-17 Pure Storage, Inc. Secure cloud-based storage system management
US10838833B1 (en) 2018-03-26 2020-11-17 Pure Storage, Inc. Providing for high availability in a data analytics pipeline without replicas
US11934322B1 (en) 2018-04-05 2024-03-19 Pure Storage, Inc. Multiple encryption keys on storage drives
US11392553B1 (en) 2018-04-24 2022-07-19 Pure Storage, Inc. Remote data management
US11436344B1 (en) 2018-04-24 2022-09-06 Pure Storage, Inc. Secure encryption in deduplication cluster
US10678433B1 (en) 2018-04-27 2020-06-09 Pure Storage, Inc. Resource-preserving system upgrade
US11385792B2 (en) 2018-04-27 2022-07-12 Pure Storage, Inc. High availability controller pair transitioning
US11675503B1 (en) 2018-05-21 2023-06-13 Pure Storage, Inc. Role-based data access
US11455409B2 (en) 2018-05-21 2022-09-27 Pure Storage, Inc. Storage layer data obfuscation
US20190354628A1 (en) 2018-05-21 2019-11-21 Pure Storage, Inc. Asynchronous replication of synchronously replicated data
US11954220B2 (en) 2018-05-21 2024-04-09 Pure Storage, Inc. Data protection for container storage
US10871922B2 (en) 2018-05-22 2020-12-22 Pure Storage, Inc. Integrated storage management between storage systems and container orchestrators
US10678436B1 (en) 2018-05-29 2020-06-09 Pure Storage, Inc. Using a PID controller to opportunistically compress more data during garbage collection
US11436023B2 (en) 2018-05-31 2022-09-06 Pure Storage, Inc. Mechanism for updating host file system and flash translation layer based on underlying NAND technology
US10776046B1 (en) 2018-06-08 2020-09-15 Pure Storage, Inc. Optimized non-uniform memory access
US11281577B1 (en) 2018-06-19 2022-03-22 Pure Storage, Inc. Garbage collection tuning for low drive wear
US11869586B2 (en) 2018-07-11 2024-01-09 Pure Storage, Inc. Increased data protection by recovering data from partially-failed solid-state devices
US11416298B1 (en) 2018-07-20 2022-08-16 Pure Storage, Inc. Providing application-specific storage by a storage system
US11403000B1 (en) 2018-07-20 2022-08-02 Pure Storage, Inc. Resiliency in a cloud-based storage system
US11954238B1 (en) 2018-07-24 2024-04-09 Pure Storage, Inc. Role-based access control for a storage system
US11146564B1 (en) 2018-07-24 2021-10-12 Pure Storage, Inc. Login authentication in a cloud storage platform
US11632360B1 (en) 2018-07-24 2023-04-18 Pure Storage, Inc. Remote access to a storage device
KR102177489B1 (en) * 2018-08-17 2020-11-11 주식회사 마크베이스 Method and device of searching index for sensor tag data
US11194759B2 (en) 2018-09-06 2021-12-07 Pure Storage, Inc. Optimizing local data relocation operations of a storage device of a storage system
US11133076B2 (en) 2018-09-06 2021-09-28 Pure Storage, Inc. Efficient relocation of data between storage devices of a storage system
US11860820B1 (en) 2018-09-11 2024-01-02 Pure Storage, Inc. Processing data through a storage system in a data pipeline
US11016696B2 (en) 2018-09-14 2021-05-25 Commvault Systems, Inc. Redundant distributed data storage system
US10846216B2 (en) 2018-10-25 2020-11-24 Pure Storage, Inc. Scalable garbage collection
US10671302B1 (en) 2018-10-26 2020-06-02 Pure Storage, Inc. Applying a rate limit across a plurality of storage systems
US11113409B2 (en) 2018-10-26 2021-09-07 Pure Storage, Inc. Efficient rekey in a transparent decrypting storage array
US11526405B1 (en) 2018-11-18 2022-12-13 Pure Storage, Inc. Cloud-based disaster recovery
US10963189B1 (en) 2018-11-18 2021-03-30 Pure Storage, Inc. Coalescing write operations in a cloud-based storage system
US11340837B1 (en) 2018-11-18 2022-05-24 Pure Storage, Inc. Storage system management via a remote console
US11023179B2 (en) 2018-11-18 2021-06-01 Pure Storage, Inc. Cloud-based storage system storage management
US11010258B2 (en) 2018-11-27 2021-05-18 Commvault Systems, Inc. Generating backup copies through interoperability between components of a data storage management system and appliances for data storage and deduplication
US11200124B2 (en) 2018-12-06 2021-12-14 Commvault Systems, Inc. Assigning backup resources based on failover of partnered data storage servers in a data storage management system
US11698727B2 (en) 2018-12-14 2023-07-11 Commvault Systems, Inc. Performing secondary copy operations based on deduplication performance
US11650749B1 (en) 2018-12-17 2023-05-16 Pure Storage, Inc. Controlling access to sensitive data in a shared dataset
JP2020106999A (en) 2018-12-26 2020-07-09 株式会社日立製作所 Storage system and data management method for storage system
US11003369B1 (en) 2019-01-14 2021-05-11 Pure Storage, Inc. Performing a tune-up procedure on a storage device during a boot process
US11194473B1 (en) 2019-01-23 2021-12-07 Pure Storage, Inc. Programming frequently read data to low latency portions of a solid-state storage array
US11588633B1 (en) 2019-03-15 2023-02-21 Pure Storage, Inc. Decommissioning keys in a decryption storage system
US11042452B1 (en) 2019-03-20 2021-06-22 Pure Storage, Inc. Storage system data recovery using data recovery as a service
US11334254B2 (en) 2019-03-29 2022-05-17 Pure Storage, Inc. Reliability based flash page sizing
US11221778B1 (en) 2019-04-02 2022-01-11 Pure Storage, Inc. Preparing data for deduplication
US11775189B2 (en) 2019-04-03 2023-10-03 Pure Storage, Inc. Segment level heterogeneity
US11397674B1 (en) 2019-04-03 2022-07-26 Pure Storage, Inc. Optimizing garbage collection across heterogeneous flash devices
US10990480B1 (en) 2019-04-05 2021-04-27 Pure Storage, Inc. Performance of RAID rebuild operations by a storage group controller of a storage system
US11068162B1 (en) 2019-04-09 2021-07-20 Pure Storage, Inc. Storage management in a cloud data store
US20200327017A1 (en) 2019-04-10 2020-10-15 Commvault Systems, Inc. Restore using deduplicated secondary copy data
US11099986B2 (en) 2019-04-12 2021-08-24 Pure Storage, Inc. Efficient transfer of memory contents
US11463264B2 (en) 2019-05-08 2022-10-04 Commvault Systems, Inc. Use of data block signatures for monitoring in an information management system
US11327676B1 (en) 2019-07-18 2022-05-10 Pure Storage, Inc. Predictive data streaming in a virtual storage system
US11126364B2 (en) 2019-07-18 2021-09-21 Pure Storage, Inc. Virtual storage system architecture
US11392555B2 (en) 2019-05-15 2022-07-19 Pure Storage, Inc. Cloud-based file services
US11853266B2 (en) 2019-05-15 2023-12-26 Pure Storage, Inc. Providing a file system in a cloud environment
CN113316766B (en) * 2019-05-29 2023-09-26 Lg电子株式会社 Digital device for performing start-up procedure and control method thereof
US11487665B2 (en) 2019-06-05 2022-11-01 Pure Storage, Inc. Tiered caching of data in a storage system
US11100092B2 (en) 2019-06-17 2021-08-24 Bank Of America Corporation Database tool
US11269861B2 (en) 2019-06-17 2022-03-08 Bank Of America Corporation Database tool
US11281394B2 (en) 2019-06-24 2022-03-22 Pure Storage, Inc. Replication across partitioning schemes in a distributed storage system
US10929046B2 (en) 2019-07-09 2021-02-23 Pure Storage, Inc. Identifying and relocating hot data to a cache determined with read velocity based on a threshold stored at a storage device
US11526408B2 (en) 2019-07-18 2022-12-13 Pure Storage, Inc. Data recovery in a virtual storage system
US11861221B1 (en) 2019-07-18 2024-01-02 Pure Storage, Inc. Providing scalable and reliable container-based storage services
US11093139B1 (en) 2019-07-18 2021-08-17 Pure Storage, Inc. Durably storing data within a virtual storage system
US11487715B1 (en) 2019-07-18 2022-11-01 Pure Storage, Inc. Resiliency in a cloud-based storage system
US11550514B2 (en) 2019-07-18 2023-01-10 Pure Storage, Inc. Efficient transfers between tiers of a virtual storage system
US11422751B2 (en) 2019-07-18 2022-08-23 Pure Storage, Inc. Creating a virtual storage system
US11086713B1 (en) 2019-07-23 2021-08-10 Pure Storage, Inc. Optimized end-to-end integrity storage system
US10909001B1 (en) * 2019-08-23 2021-02-02 EMC IP Holding Company LLC Storage system with snapshot group split functionality
US11086553B1 (en) 2019-08-28 2021-08-10 Pure Storage, Inc. Tiering duplicated objects in a cloud-based object store
US11693713B1 (en) 2019-09-04 2023-07-04 Pure Storage, Inc. Self-tuning clusters for resilient microservices
US11963321B2 (en) 2019-09-11 2024-04-16 Pure Storage, Inc. Low profile latching mechanism
US11797569B2 (en) 2019-09-13 2023-10-24 Pure Storage, Inc. Configurable data replication
US11625416B1 (en) 2019-09-13 2023-04-11 Pure Storage, Inc. Uniform model for distinct types of data replication
US11573864B1 (en) 2019-09-16 2023-02-07 Pure Storage, Inc. Automating database management in a storage system
US11403185B2 (en) 2019-09-27 2022-08-02 Amazon Technologies, Inc. Network-accessible block-level snapshots
US11360856B2 (en) * 2019-09-27 2022-06-14 Amazon Technologies, Inc. Manifest index for block-level snapshots
US11669386B1 (en) 2019-10-08 2023-06-06 Pure Storage, Inc. Managing an application's resource stack
US11403043B2 (en) 2019-10-15 2022-08-02 Pure Storage, Inc. Efficient data compression by grouping similar data within a data segment
US11720714B2 (en) 2019-11-22 2023-08-08 Pure Storage, Inc. Inter-I/O relationship based detection of a security threat to a storage system
US11687418B2 (en) 2019-11-22 2023-06-27 Pure Storage, Inc. Automatic generation of recovery plans specific to individual storage elements
US11941116B2 (en) 2019-11-22 2024-03-26 Pure Storage, Inc. Ransomware-based data protection parameter modification
US11341236B2 (en) 2019-11-22 2022-05-24 Pure Storage, Inc. Traffic-based detection of a security threat to a storage system
US11675898B2 (en) 2019-11-22 2023-06-13 Pure Storage, Inc. Recovery dataset management for security threat monitoring
US11645162B2 (en) 2019-11-22 2023-05-09 Pure Storage, Inc. Recovery point determination for data restoration in a storage system
US11651075B2 (en) 2019-11-22 2023-05-16 Pure Storage, Inc. Extensible attack monitoring by a storage system
US11657155B2 (en) 2019-11-22 2023-05-23 Pure Storage, Inc Snapshot delta metric based determination of a possible ransomware attack against data maintained by a storage system
US11520907B1 (en) 2019-11-22 2022-12-06 Pure Storage, Inc. Storage system snapshot retention based on encrypted data
US11625481B2 (en) 2019-11-22 2023-04-11 Pure Storage, Inc. Selective throttling of operations potentially related to a security threat to a storage system
US11720692B2 (en) 2019-11-22 2023-08-08 Pure Storage, Inc. Hardware token based management of recovery datasets for a storage system
US11755751B2 (en) 2019-11-22 2023-09-12 Pure Storage, Inc. Modify access restrictions in response to a possible attack against data stored by a storage system
US11500788B2 (en) 2019-11-22 2022-11-15 Pure Storage, Inc. Logical address based authorization of operations with respect to a storage system
US11615185B2 (en) 2019-11-22 2023-03-28 Pure Storage, Inc. Multi-layer security threat detection for a storage system
US20210173811A1 (en) 2019-12-04 2021-06-10 Commvault Systems, Inc. Optimizing the restoration of deduplicated data stored in multi-node replicated file systems
US11868318B1 (en) 2019-12-06 2024-01-09 Pure Storage, Inc. End-to-end encryption in a storage system with multi-tenancy
US11720497B1 (en) 2020-01-13 2023-08-08 Pure Storage, Inc. Inferred nonsequential prefetch based on data access patterns
US11733901B1 (en) 2020-01-13 2023-08-22 Pure Storage, Inc. Providing persistent storage to transient cloud computing services
US11709636B1 (en) 2020-01-13 2023-07-25 Pure Storage, Inc. Non-sequential readahead for deep learning training
US11868622B2 (en) 2020-02-25 2024-01-09 Pure Storage, Inc. Application recovery across storage systems
US11637896B1 (en) 2020-02-25 2023-04-25 Pure Storage, Inc. Migrating applications to a cloud-computing environment
US11955174B2 (en) * 2020-02-26 2024-04-09 Arista Networks, Inc. Selectively connectable content-addressable memory
US11550816B1 (en) 2020-03-02 2023-01-10 Amazon Technologies, Inc. Variable replication levels for an object of a snapshot of a block storage volume
US11321006B1 (en) 2020-03-25 2022-05-03 Pure Storage, Inc. Data loss prevention during transitions from a replication source
US11099956B1 (en) 2020-03-26 2021-08-24 Commvault Systems, Inc. Snapshot-based disaster recovery orchestration of virtual machine failover and failback operations
US11301152B1 (en) 2020-04-06 2022-04-12 Pure Storage, Inc. Intelligently moving data between storage systems
US11630598B1 (en) 2020-04-06 2023-04-18 Pure Storage, Inc. Scheduling data replication operations
US11494267B2 (en) 2020-04-14 2022-11-08 Pure Storage, Inc. Continuous value data redundancy
US11921670B1 (en) 2020-04-20 2024-03-05 Pure Storage, Inc. Multivariate data backup retention policies
US20210365587A1 (en) * 2020-05-20 2021-11-25 EMC IP Holding Company LLC Data masking in a microservice architecture
US11468193B2 (en) 2020-05-20 2022-10-11 EMC IP Holding Company LLC Data masking in a microservice architecture
US11687424B2 (en) 2020-05-28 2023-06-27 Commvault Systems, Inc. Automated media agent state management
US11431488B1 (en) 2020-06-08 2022-08-30 Pure Storage, Inc. Protecting local key generation using a remote key management service
US11349917B2 (en) 2020-07-23 2022-05-31 Pure Storage, Inc. Replication handling among distinct networks
US11442652B1 (en) 2020-07-23 2022-09-13 Pure Storage, Inc. Replication handling during storage system transportation
JP7121079B2 (en) 2020-07-31 2022-08-17 株式会社日立製作所 Storage system and data replication method in storage system
CN112052193B (en) * 2020-09-28 2023-04-07 成都佰维存储科技有限公司 Garbage recycling method and device, readable storage medium and electronic equipment
US11640339B2 (en) 2020-11-23 2023-05-02 International Business Machines Corporation Creating a backup data set
US11397545B1 (en) 2021-01-20 2022-07-26 Pure Storage, Inc. Emulating persistent reservations in a cloud-based storage system
US11853285B1 (en) 2021-01-22 2023-12-26 Pure Storage, Inc. Blockchain logging of volume-level events in a storage system
US11645175B2 (en) 2021-02-12 2023-05-09 Commvault Systems, Inc. Automatic failover of a storage manager
CN112965783A (en) * 2021-02-24 2021-06-15 上海英方软件股份有限公司 System and method for using storage snapshot backup virtual machine
US20220365827A1 (en) 2021-05-12 2022-11-17 Pure Storage, Inc. Rebalancing In A Fleet Of Storage Systems Using Data Science
US11816129B2 (en) 2021-06-22 2023-11-14 Pure Storage, Inc. Generating datasets using approximate baselines
JP7429214B2 (en) 2021-10-07 2024-02-07 株式会社日立製作所 Storage system and data replication method in storage system
US11714723B2 (en) 2021-10-29 2023-08-01 Pure Storage, Inc. Coordinated snapshots for data stored across distinct storage environments
US11893263B2 (en) 2021-10-29 2024-02-06 Pure Storage, Inc. Coordinated checkpoints among storage systems implementing checkpoint-based replication
US11914867B2 (en) 2021-10-29 2024-02-27 Pure Storage, Inc. Coordinated snapshots among storage systems implementing a promotion/demotion model
US11922052B2 (en) 2021-12-15 2024-03-05 Pure Storage, Inc. Managing links between storage objects
US11847071B2 (en) 2021-12-30 2023-12-19 Pure Storage, Inc. Enabling communication between a single-port device and multiple storage system controllers
US11860780B2 (en) 2022-01-28 2024-01-02 Pure Storage, Inc. Storage cache management
US11886295B2 (en) 2022-01-31 2024-01-30 Pure Storage, Inc. Intra-block error correction

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5403639A (en) * 1992-09-02 1995-04-04 Storage Technology Corporation File server having snapshot application data groups
US20020178335A1 (en) * 2000-06-19 2002-11-28 Storage Technology Corporation Apparatus and method for dynamically changeable virtual mapping scheme
US20030140209A1 (en) * 2001-12-10 2003-07-24 Richard Testardi Fast path caching
US20050216535A1 (en) * 2004-03-29 2005-09-29 Nobuyuki Saika Backup method, storage system, and program for backup
US20050223154A1 (en) * 2004-04-02 2005-10-06 Hitachi Global Storage Technologies Netherlands B.V. Method for controlling disk drive
US20100057673A1 (en) * 2008-09-04 2010-03-04 Boris Savov Reusable mapping rules for data to data transformation
US20120054419A1 (en) * 2010-08-27 2012-03-01 Via Technologies, Inc. Memory device and operating method thereof
US20120330903A1 (en) * 2011-06-21 2012-12-27 Subramaniam Periyagaram Deduplication in an extent-based architecture
US20130042052A1 (en) * 2011-08-11 2013-02-14 John Colgrove Logical sector mapping in a flash storage array
US20130046949A1 (en) * 2011-08-16 2013-02-21 John Colgrove Mapping in a storage system
US20140201150A1 (en) * 2013-01-11 2014-07-17 Commvault Systems, Inc. Single snapshot for multiple agents
US20150143065A1 (en) * 2012-07-17 2015-05-21 Huawei Technologies Co., Ltd. Data Processing Method and Apparatus, and Shared Storage Device

Family Cites Families (186)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5208813A (en) 1990-10-23 1993-05-04 Array Technology Corporation On-line reconstruction of a failed redundant array system
DE9310582U1 (en) 1993-07-15 1993-09-23 Hettich Paul Gmbh & Co LOCKING DEVICE FOR DRAWERS OR THE LIKE
US6412045B1 (en) 1995-05-23 2002-06-25 Lsi Logic Corporation Method for transferring data from a host computer to a storage media using selectable caching strategies
US6721941B1 (en) * 1996-08-27 2004-04-13 Compuware Corporation Collection of timing and coverage data through a debugging interface
US5832529A (en) 1996-10-11 1998-11-03 Sun Microsystems, Inc. Methods, apparatus, and product for distributed garbage collection
US5940838A (en) 1997-07-11 1999-08-17 International Business Machines Corporation Parallel file system and method anticipating cache usage patterns
US6038639A (en) 1997-09-09 2000-03-14 Storage Technology Corporation Data file storage management system for snapshot copy operations
US6799283B1 (en) 1998-12-04 2004-09-28 Matsushita Electric Industrial Co., Ltd. Disk array device
JP2000181803A (en) 1998-12-18 2000-06-30 Fujitsu Ltd Electronic data keeping device with key management function and method therefor
US8127121B2 (en) * 1999-01-28 2012-02-28 Ati Technologies Ulc Apparatus for executing programs for a first computer architechture on a computer of a second architechture
US6834298B1 (en) 1999-09-21 2004-12-21 Siemens Information And Communication Networks, Inc. System and method for network auto-discovery and configuration
US6912537B2 (en) 2000-06-20 2005-06-28 Storage Technology Corporation Dynamically changeable virtual mapping scheme
US20070192863A1 (en) * 2005-07-01 2007-08-16 Harsh Kapoor Systems and methods for processing data flows
US20110213869A1 (en) * 2000-09-25 2011-09-01 Yevgeny Korsunsky Processing data flows with a data flow processor
JP2002108573A (en) 2000-09-28 2002-04-12 Nec Corp Disk array device and method for controlling its error and recording medium with its control program recorded thereon
US6757769B1 (en) 2000-11-28 2004-06-29 Emc Corporation Cooperative lock override procedure
US6718448B1 (en) 2000-11-28 2004-04-06 Emc Corporation Queued locking of a shared resource using multimodal lock types
US6857059B2 (en) * 2001-01-11 2005-02-15 Yottayotta, Inc. Storage virtualization system and methods
US6850938B1 (en) 2001-02-08 2005-02-01 Cisco Technology, Inc. Method and apparatus providing optimistic locking of shared computer resources
WO2002065309A1 (en) 2001-02-13 2002-08-22 Candera, Inc. System and method for policy based storage provisioning and management
US6968350B2 (en) * 2001-04-07 2005-11-22 Microsoft Corporation Method for establishing a virtual hard drive for an emulated computer system running on a host computer system
US6973549B1 (en) 2001-12-10 2005-12-06 Incipient, Inc. Locking technique for control and synchronization
US7373364B1 (en) * 2002-03-05 2008-05-13 Network Appliance, Inc. System and method for creating a point-in-time restoration of a database file
US7100089B1 (en) * 2002-09-06 2006-08-29 3Pardata, Inc. Determining differences between snapshots
US7260628B2 (en) 2002-09-06 2007-08-21 Hitachi, Ltd. Event notification in storage networks
US7216164B1 (en) 2002-10-09 2007-05-08 Cisco Technology, Inc. Methods and apparatus for determining the performance of a server
JP4717317B2 (en) * 2002-11-05 2011-07-06 シャープ株式会社 File management apparatus, recording / reproducing apparatus, recording medium, recording / reproducing method, and program
US7028218B2 (en) 2002-12-02 2006-04-11 Emc Corporation Redundant multi-processor and logical processor configuration for a file server
US7231544B2 (en) * 2003-02-27 2007-06-12 Hewlett-Packard Development Company, L.P. Restoring data from point-in-time representations of the data
CN1795440A (en) 2003-04-07 2006-06-28 艾特拉克斯公司 Network security system based on physical location
US8521789B2 (en) * 2003-05-27 2013-08-27 Oracle International Corporation Undrop objects and dependent objects in a database system
US6996586B2 (en) * 2003-06-18 2006-02-07 International Business Machines Corporation Method, system, and article for incremental virtual copy of a data block
US7424498B1 (en) * 2003-06-30 2008-09-09 Data Domain, Inc. Probabilistic summary data structure based encoding for garbage collection
CN101566929B (en) * 2003-08-14 2013-10-16 克姆佩棱特科技公司 Virtual disk drive system and method
US9489150B2 (en) * 2003-08-14 2016-11-08 Dell International L.L.C. System and method for transferring data between different raid data storage types for current data and replay data
US7865485B2 (en) 2003-09-23 2011-01-04 Emc Corporation Multi-threaded write interface and methods for increasing the single file read and write throughput of a file server
US7529782B2 (en) * 2003-11-13 2009-05-05 Commvault Systems, Inc. System and method for performing a snapshot and for restoring data
US7205787B1 (en) * 2003-11-24 2007-04-17 Neascape, Inc. On-chip termination for a high-speed single-ended interface
JP4426262B2 (en) 2003-11-26 2010-03-03 株式会社日立製作所 Disk array device and failure avoiding method for disk array device
US8560747B1 (en) 2007-02-16 2013-10-15 Vmware, Inc. Associating heartbeat data with access to shared resources of a computer system
DE602004026422D1 (en) * 2004-02-25 2010-05-20 Microsoft Corp DATABASE DATA RESTORATION SYSTEM AND METHOD
US7424482B2 (en) 2004-04-26 2008-09-09 Storwize Inc. Method and system for compression of data for block mode access storage
US20060074940A1 (en) 2004-10-05 2006-04-06 International Business Machines Corporation Dynamic management of node clusters to enable data sharing
US7363444B2 (en) 2005-01-10 2008-04-22 Hewlett-Packard Development Company, L.P. Method for taking snapshots of data
US7315917B2 (en) * 2005-01-20 2008-01-01 Sandisk Corporation Scheduling of housekeeping operations in flash memory systems
US20060174074A1 (en) 2005-02-03 2006-08-03 International Business Machines Corporation Point-in-time copy operation
US7913300B1 (en) 2005-04-08 2011-03-22 Netapp, Inc. Centralized role-based access control for storage servers
US8200887B2 (en) 2007-03-29 2012-06-12 Violin Memory, Inc. Memory management system and method
US8364845B2 (en) 2005-05-19 2013-01-29 Wyse Technology Inc. Method and system for thin client configuration
WO2006123416A1 (en) 2005-05-19 2006-11-23 Fujitsu Limited Disk failure recovery method and disk array device
US7933936B2 (en) 2005-06-10 2011-04-26 Network Appliance, Inc. Method and system for automatic management of storage space
US20080229415A1 (en) * 2005-07-01 2008-09-18 Harsh Kapoor Systems and methods for processing data flows
US7979613B2 (en) 2005-07-15 2011-07-12 International Business Machines Corporation Performance of a storage system
US7869525B2 (en) * 2005-08-01 2011-01-11 Ati Technologies, Inc. Dynamic bus inversion method and system
JP4424282B2 (en) * 2005-08-22 2010-03-03 日本電気株式会社 File management apparatus, file management method, and program
US7426618B2 (en) * 2005-09-06 2008-09-16 Dot Hill Systems Corp. Snapshot restore method and apparatus
JP2007087036A (en) 2005-09-21 2007-04-05 Hitachi Ltd Snapshot maintenance device and method
JP4662548B2 (en) * 2005-09-27 2011-03-30 株式会社日立製作所 Snapshot management apparatus and method, and storage system
ITVA20050061A1 (en) 2005-11-08 2007-05-09 St Microelectronics Srl METHOD OF MANAGEMENT OF A NON-VOLATILE MEMORY AND RELATIVE MEMORY DEVICE
JP4927408B2 (en) * 2006-01-25 2012-05-09 株式会社日立製作所 Storage system and data restoration method thereof
JP2007199953A (en) 2006-01-25 2007-08-09 Fujitsu Ltd Disk array device and disk array control method
US7743197B2 (en) * 2006-05-11 2010-06-22 Emulex Design & Manufacturing Corporation System and method for virtualizing PCIe devices
JP2007233903A (en) 2006-03-03 2007-09-13 Hitachi Ltd Storage controller and data recovery method for storage controller
US8832247B2 (en) 2006-03-24 2014-09-09 Blue Coat Systems, Inc. Methods and systems for caching content at multiple levels
US20070277012A1 (en) * 2006-05-23 2007-11-29 Hitachi, Ltd. Method and apparatus for managing backup data and journal
US8307148B2 (en) * 2006-06-23 2012-11-06 Microsoft Corporation Flash management techniques
US7987438B2 (en) 2006-08-10 2011-07-26 International Business Machines Corporation Structure for initializing expansion adapters installed in a computer system having similar expansion adapters
US7555599B2 (en) 2006-09-06 2009-06-30 International Business Machines Corporation System and method of mirrored RAID array write management
US7475215B2 (en) 2006-09-08 2009-01-06 Lsi Corporation Identification of uncommitted memory blocks during an initialization procedure
JP4573898B2 (en) 2006-11-27 2010-11-04 富士通株式会社 Server management program, mail server management program, server management system, management server, and server management method
US8694712B2 (en) * 2006-12-05 2014-04-08 Microsoft Corporation Reduction of operational costs of virtual TLBs
US20080155191A1 (en) 2006-12-21 2008-06-26 Anderson Robert J Systems and methods for providing heterogeneous storage systems
US8370562B2 (en) 2007-02-25 2013-02-05 Sandisk Il Ltd. Interruptible cache flushing in flash memory systems
US9632870B2 (en) 2007-03-29 2017-04-25 Violin Memory, Inc. Memory system with multiple striping of raid groups and method for performing the same
JP4900811B2 (en) 2007-03-30 2012-03-21 株式会社日立製作所 Storage system and storage control method
JP4529990B2 (en) 2007-03-30 2010-08-25 ブラザー工業株式会社 Image processing program and image processing apparatus
US7975115B2 (en) 2007-04-11 2011-07-05 Dot Hill Systems Corporation Method and apparatus for separating snapshot preserved and write data
US8086652B1 (en) 2007-04-27 2011-12-27 Netapp, Inc. Storage system-based hole punching for reclaiming unused space from a data container
US7958303B2 (en) 2007-04-27 2011-06-07 Gary Stephen Shuster Flexible data storage system
US7991942B2 (en) 2007-05-09 2011-08-02 Stmicroelectronics S.R.L. Memory block compaction method, circuit, and system in storage devices based on flash memories
EP2154815A4 (en) 2007-05-30 2012-11-14 Fujitsu Ltd Image encrypting device, image decrypting device, method and program
US7765426B2 (en) * 2007-06-07 2010-07-27 Micron Technology, Inc. Emerging bad block detection
US8204858B2 (en) 2007-06-25 2012-06-19 Dot Hill Systems Corporation Snapshot reset method and apparatus
US8874854B2 (en) 2007-07-30 2014-10-28 International Business Machines Corporation Method for selectively enabling and disabling read caching in a storage subsystem
JP5330702B2 (en) * 2008-01-31 2013-10-30 株式会社日立製作所 Storage device system having a plurality of controller modules
US7970994B2 (en) 2008-03-04 2011-06-28 International Business Machines Corporation High performance disk array rebuild
US8352540B2 (en) 2008-03-06 2013-01-08 International Business Machines Corporation Distinguishing data streams to enhance data storage efficiency
US8606982B2 (en) * 2008-03-10 2013-12-10 Qimonda Ag Derivative logical output
US7873619B1 (en) 2008-03-31 2011-01-18 Emc Corporation Managing metadata
JP2009251725A (en) * 2008-04-02 2009-10-29 Hitachi Ltd Storage controller and duplicated data detection method using storage controller
US8621241B1 (en) 2008-04-25 2013-12-31 Netapp, Inc. Storage and recovery of cryptographic key identifiers
US8296398B2 (en) * 2008-04-29 2012-10-23 Overland Storage, Inc. Peer-to-peer redundant file server system and methods
US8117464B1 (en) 2008-04-30 2012-02-14 Netapp, Inc. Sub-volume level security for deduplicated data
US9678879B2 (en) 2008-05-29 2017-06-13 Red Hat, Inc. Set partitioning for encoding file system allocation metadata
US8347059B2 (en) 2008-08-15 2013-01-01 International Business Machines Corporation Management of recycling bin for thinly-provisioned logical volumes
US8296547B2 (en) * 2008-08-27 2012-10-23 International Business Machines Corporation Loading entries into a TLB in hardware via indirect TLB entries
US20100077205A1 (en) 2008-09-19 2010-03-25 Ekstrom Joseph J System and Method for Cipher E-Mail Protection
US8756369B2 (en) 2008-09-26 2014-06-17 Netapp, Inc. Priority command queues for low latency solid state drives
US7910688B2 (en) 2008-10-22 2011-03-22 Evonik Stockhausen Inc. Recycling superabsorbent polymer fines
JP4399021B1 (en) 2008-10-29 2010-01-13 株式会社東芝 Disk array control device and storage device
US7945733B2 (en) 2008-12-12 2011-05-17 Lsi Corporation Hierarchical storage management (HSM) for redundant array of independent disks (RAID)
US8200922B2 (en) 2008-12-17 2012-06-12 Netapp, Inc. Storage system snapshot assisted by SSD technology
CN102257482B (en) 2008-12-19 2015-06-03 惠普开发有限公司 Redundant data storage for uniform read latency
US8312204B2 (en) 2009-01-23 2012-11-13 Seagate Technology Llc System and method for wear leveling in a data storage device
JP4869368B2 (en) 2009-03-12 2012-02-08 株式会社東芝 Storage device and virtualization device
US7941584B2 (en) 2009-03-26 2011-05-10 Arm Limited Data processing apparatus and method for performing hazard detection
US8205065B2 (en) 2009-03-30 2012-06-19 Exar Corporation System and method for data deduplication
US8560787B2 (en) 2009-03-30 2013-10-15 International Business Machines Corporation Incremental backup of source to target storage volume
US8234468B1 (en) * 2009-04-29 2012-07-31 Netapp, Inc. System and method for providing variable length deduplication on a fixed block file system
TWI397009B (en) 2009-04-30 2013-05-21 Inventec Corp Data processing apparatus of basic input/output system
US8180955B2 (en) 2009-05-06 2012-05-15 Via Telecom, Inc. Computing systems and methods for managing flash memory device
US8290911B1 (en) 2009-06-30 2012-10-16 Symantec Corporation System and method for implementing data deduplication-aware copying of data
EP2302636B1 (en) * 2009-09-21 2014-11-05 STMicroelectronics (Rousset) SAS Method for reading a non volatile memory using metadata and a correspondence table
JP2011070345A (en) * 2009-09-25 2011-04-07 Hitachi Ltd Computer system, management device for the same and management method for the same
US8589851B2 (en) * 2009-12-15 2013-11-19 Memoir Systems, Inc. Intelligent memory system compiler
US8510569B2 (en) * 2009-12-16 2013-08-13 Intel Corporation Providing integrity verification and attestation in a hidden execution environment
US9134918B2 (en) 2009-12-31 2015-09-15 Sandisk Technologies Inc. Physical compression of data with flat or systematic pattern
US8452932B2 (en) 2010-01-06 2013-05-28 Storsimple, Inc. System and method for efficiently creating off-site data volume back-ups
US9176853B2 (en) * 2010-01-29 2015-11-03 Symantec Corporation Managing copy-on-writes to snapshots
US8627000B2 (en) * 2010-02-08 2014-01-07 Microsoft Corporation Virtual disk manipulation operations
WO2011104663A1 (en) 2010-02-23 2011-09-01 Confidato Security Solutions Ltd Method and computer program product for order preserving symbol based encryption
JP4892072B2 (en) 2010-03-24 2012-03-07 株式会社東芝 Storage device that eliminates duplicate data in cooperation with host device, storage system including the storage device, and deduplication method in the system
US8825982B2 (en) * 2010-06-10 2014-09-02 Global Supercomputing Corporation Storage unsharing
US8396843B2 (en) 2010-06-14 2013-03-12 Dell Products L.P. Active file instant cloning
US8738970B2 (en) 2010-07-23 2014-05-27 Salesforce.Com, Inc. Generating performance alerts
US8713268B2 (en) 2010-08-05 2014-04-29 Ut-Battelle, Llc Coordinated garbage collection for raid array of solid state disks
US8468318B2 (en) 2010-09-15 2013-06-18 Pure Storage Inc. Scheduling of I/O writes in a storage environment
US8589625B2 (en) 2010-09-15 2013-11-19 Pure Storage, Inc. Scheduling of reconstructive I/O read operations in a storage environment
US8775868B2 (en) 2010-09-28 2014-07-08 Pure Storage, Inc. Adaptive RAID for an SSD environment
US20120117029A1 (en) 2010-11-08 2012-05-10 Stephen Gold Backup policies for using different storage tiers
US20120124329A1 (en) * 2010-11-17 2012-05-17 Mccombs Edward M Translation Lookaside Buffer Structure Including a Data Array Having an Integrated Multiplexer
US8458145B2 (en) * 2011-01-20 2013-06-04 Infinidat Ltd. System and method of storage optimization
US8966184B2 (en) 2011-01-31 2015-02-24 Intelligent Intellectual Property Holdings 2, LLC. Apparatus, system, and method for managing eviction of data
US20120233416A1 (en) * 2011-03-08 2012-09-13 International Business Machines Corporation Multi-target, point-in-time-copy architecture with data deduplication
US9563555B2 (en) 2011-03-18 2017-02-07 Sandisk Technologies Llc Systems and methods for storage allocation
US8433683B2 (en) * 2011-06-08 2013-04-30 Oracle International Corporation Systems and methods of data replication of a file system
US8706703B2 (en) 2011-06-27 2014-04-22 International Business Machines Corporation Efficient file system object-based deduplication
US8595267B2 (en) * 2011-06-27 2013-11-26 Amazon Technologies, Inc. System and method for implementing a scalable data storage service
US8751463B1 (en) 2011-06-30 2014-06-10 Emc Corporation Capacity forecasting for a deduplicating storage system
CN102866935B (en) * 2011-07-07 2014-11-12 北京飞杰信息技术有限公司 Instant replication method and storage system based on ISCSI (internet small computer system interface)
US8527544B1 (en) 2011-08-11 2013-09-03 Pure Storage Inc. Garbage collection in a storage system
US8793467B2 (en) 2011-09-30 2014-07-29 Pure Storage, Inc. Variable length encoding in a storage system
JP5768587B2 (en) 2011-08-17 2015-08-26 富士通株式会社 Storage system, storage control device, and storage control method
US8527732B2 (en) * 2011-09-16 2013-09-03 Hitachi, Ltd. Storage system and method of controlling storage system
US8700875B1 (en) 2011-09-20 2014-04-15 Netapp, Inc. Cluster view for storage devices
US9542413B2 (en) 2011-10-06 2017-01-10 Hitachi, Ltd. Stored data deduplication method, stored data deduplication apparatus, and deduplication program
US8825605B2 (en) 2011-10-11 2014-09-02 Netapp, Inc. Deduplication aware scheduling of requests to access data blocks
US8600939B2 (en) * 2011-10-24 2013-12-03 Oracle International Corporation Writable snapshots
US9635132B1 (en) * 2011-12-15 2017-04-25 Amazon Technologies, Inc. Service and APIs for remote volume-based block storage
US8918579B2 (en) 2012-02-06 2014-12-23 Sandisk Technologies Inc. Storage device and method for selective data compression
US9075710B2 (en) 2012-04-17 2015-07-07 SanDisk Technologies, Inc. Non-volatile key-value store
US9519647B2 (en) * 2012-04-17 2016-12-13 Sandisk Technologies Llc Data expiry in a non-volatile device
US8996881B2 (en) 2012-04-23 2015-03-31 International Business Machines Corporation Preserving redundancy in data deduplication systems by encryption
US8793466B2 (en) 2012-04-27 2014-07-29 Netapp, Inc. Efficient data object storage and retrieval
US9645177B2 (en) 2012-05-04 2017-05-09 Seagate Technology Llc Retention-drift-history-based non-volatile memory read threshold optimization
US8874850B1 (en) 2012-05-10 2014-10-28 Netapp, Inc. Hierarchically tagged cache
US20130318314A1 (en) 2012-05-25 2013-11-28 Red Hat, Inc. Managing copies of data on multiple nodes using a data controller node to avoid transaction deadlock
KR20150021117A (en) 2012-06-18 2015-02-27 액티피오 인크. Enhanced data management virtualization system
CN102779080B (en) * 2012-06-18 2015-05-27 上海爱数软件有限公司 Method for generating snapshot, method and device for data recovery by using snapshot
US9223659B1 (en) * 2012-06-28 2015-12-29 Emc International Company Generating and accessing a virtual volume snapshot in a continuous data protection system
US8959305B1 (en) 2012-06-29 2015-02-17 Emc Corporation Space reclamation with virtually provisioned devices
US9489293B2 (en) 2012-08-17 2016-11-08 Netapp, Inc. Techniques for opportunistic data storage
US8799602B2 (en) * 2012-08-22 2014-08-05 International Business Machines Corporation Automated migration to a new target volume via merged bitmaps to maintain consistency
US9176822B2 (en) 2012-08-31 2015-11-03 Cleversafe, Inc. Adjusting dispersed storage error encoding parameters
JP5954081B2 (en) 2012-09-26 2016-07-20 富士通株式会社 Storage control device, storage control method, and storage control program
US9348757B2 (en) * 2012-10-08 2016-05-24 International Business Machines Corporation System supporting multiple partitions with differing translation formats
WO2014076743A1 (en) 2012-11-19 2014-05-22 Hitachi, Ltd. Storage system
US9348840B2 (en) 2012-12-14 2016-05-24 Intel Corporation Adaptive data striping and replication across multiple storage clouds for high availability and performance
US9063967B2 (en) 2013-01-10 2015-06-23 Pure Storage, Inc. Performing copies in a storage system
US9652376B2 (en) 2013-01-28 2017-05-16 Radian Memory Systems, Inc. Cooperative flash memory control
US9335932B2 (en) 2013-03-15 2016-05-10 Bracket Computing, Inc. Storage unit selection for virtualized storage units
KR20140114515A (en) * 2013-03-15 2014-09-29 삼성전자주식회사 Nonvolatile memory device and deduplication method thereof
US9519575B2 (en) * 2013-04-25 2016-12-13 Sandisk Technologies Llc Conditional iteration for a non-volatile device
US10263770B2 (en) 2013-11-06 2019-04-16 Pure Storage, Inc. Data protection in a storage system using external secrets
US9516016B2 (en) 2013-11-11 2016-12-06 Pure Storage, Inc. Storage array password management
JP6233086B2 (en) 2014-02-20 2017-11-22 富士通株式会社 Storage control device, storage system, and control program
US9798596B2 (en) 2014-02-27 2017-10-24 Commvault Systems, Inc. Automatic alert escalation for an information management system
US10656864B2 (en) 2014-03-20 2020-05-19 Pure Storage, Inc. Data replication within a flash storage array
US9361469B2 (en) 2014-03-26 2016-06-07 Amazon Technologies, Inc. Electronic communication with secure screen sharing of sensitive information
US9513820B1 (en) 2014-04-07 2016-12-06 Pure Storage, Inc. Dynamically controlling temporary compromise on data redundancy
US9563509B2 (en) 2014-07-15 2017-02-07 Nimble Storage, Inc. Methods and systems for storing data in a redundant manner on a plurality of storage units of a storage system
US10430282B2 (en) 2014-10-07 2019-10-01 Pure Storage, Inc. Optimizing replication by distinguishing user and system write activity
US9489132B2 (en) 2014-10-07 2016-11-08 Pure Storage, Inc. Utilizing unmapped and unknown states in a replicated storage system
US9552248B2 (en) 2014-12-11 2017-01-24 Pure Storage, Inc. Cloud alert to replica
JP2016129003A (en) * 2015-01-09 2016-07-14 富士通株式会社 Snapshot management method, management unit, and snapshot management program
US10459657B2 (en) 2016-09-16 2019-10-29 Hewlett Packard Enterprise Development Lp Storage system with read cache-on-write buffer

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5403639A (en) * 1992-09-02 1995-04-04 Storage Technology Corporation File server having snapshot application data groups
US20020178335A1 (en) * 2000-06-19 2002-11-28 Storage Technology Corporation Apparatus and method for dynamically changeable virtual mapping scheme
US20030140209A1 (en) * 2001-12-10 2003-07-24 Richard Testardi Fast path caching
US20050216535A1 (en) * 2004-03-29 2005-09-29 Nobuyuki Saika Backup method, storage system, and program for backup
US20050223154A1 (en) * 2004-04-02 2005-10-06 Hitachi Global Storage Technologies Netherlands B.V. Method for controlling disk drive
US20100057673A1 (en) * 2008-09-04 2010-03-04 Boris Savov Reusable mapping rules for data to data transformation
US20120054419A1 (en) * 2010-08-27 2012-03-01 Via Technologies, Inc. Memory device and operating method thereof
US20120330903A1 (en) * 2011-06-21 2012-12-27 Subramaniam Periyagaram Deduplication in an extent-based architecture
US20130042052A1 (en) * 2011-08-11 2013-02-14 John Colgrove Logical sector mapping in a flash storage array
US20130046949A1 (en) * 2011-08-16 2013-02-21 John Colgrove Mapping in a storage system
US20150143065A1 (en) * 2012-07-17 2015-05-21 Huawei Technologies Co., Ltd. Data Processing Method and Apparatus, and Shared Storage Device
US20140201150A1 (en) * 2013-01-11 2014-07-17 Commvault Systems, Inc. Single snapshot for multiple agents

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160105502A1 (en) * 2013-06-20 2016-04-14 Huawei Technologies Co., Ltd. Data synchronization method, data synchronization apparatus, and distributed system
US9823842B2 (en) 2014-05-12 2017-11-21 The Research Foundation For The State University Of New York Gang migration of virtual machines using cluster-wide deduplication
US10156986B2 (en) 2014-05-12 2018-12-18 The Research Foundation For The State University Of New York Gang migration of virtual machines using cluster-wide deduplication
WO2016094145A1 (en) * 2014-12-11 2016-06-16 Pure Storage, Inc. Cloud alert to replica
US9552248B2 (en) 2014-12-11 2017-01-24 Pure Storage, Inc. Cloud alert to replica
US10235065B1 (en) * 2014-12-11 2019-03-19 Pure Storage, Inc. Datasheet replication in a cloud computing environment
US11061786B1 (en) 2014-12-11 2021-07-13 Pure Storage, Inc. Cloud-based disaster recovery of a storage system
US11775392B2 (en) 2014-12-11 2023-10-03 Pure Storage, Inc. Indirect replication of a dataset
CN106155569A (en) * 2015-04-07 2016-11-23 北京中科同向信息技术有限公司 A kind of storage snapping technique based on mirror image
US10922240B2 (en) 2018-09-19 2021-02-16 Toshiba Memory Corporation Memory system, storage system and method of controlling the memory system

Also Published As

Publication number Publication date
US9891858B1 (en) 2018-02-13
US9646039B2 (en) 2017-05-09
WO2014110311A1 (en) 2014-07-17
US11853584B1 (en) 2023-12-26
US20210382625A1 (en) 2021-12-09
EP2943872A1 (en) 2015-11-18
WO2014110158A1 (en) 2014-07-17
US11099769B1 (en) 2021-08-24
JP2016508273A (en) 2016-03-17
JP2016510458A (en) 2016-04-07
WO2014110343A1 (en) 2014-07-17
US9436720B2 (en) 2016-09-06
CN105027068B (en) 2018-10-26
CN105027069A (en) 2015-11-04
US20140195754A1 (en) 2014-07-10
US20190179535A1 (en) 2019-06-13
US9063967B2 (en) 2015-06-23
US10013317B1 (en) 2018-07-03
US20230273743A1 (en) 2023-08-31
EP2943870A1 (en) 2015-11-18
CN105027068A (en) 2015-11-04
US9361035B1 (en) 2016-06-07
WO2014110137A1 (en) 2014-07-17
US20140195749A1 (en) 2014-07-10
US20230384963A1 (en) 2023-11-30
US9760313B1 (en) 2017-09-12
US9589008B2 (en) 2017-03-07
CN105027069B (en) 2020-05-12
US20140195755A1 (en) 2014-07-10
CN105027070B (en) 2018-05-29
US20140195762A1 (en) 2014-07-10
US9880779B1 (en) 2018-01-30
JP2019096355A (en) 2019-06-20
KR20150104604A (en) 2015-09-15
EP2943871A1 (en) 2015-11-18
KR20150104606A (en) 2015-09-15
US10585617B1 (en) 2020-03-10
CN105027070A (en) 2015-11-04
JP6543195B2 (en) 2019-07-10
KR20150104605A (en) 2015-09-15
US10235093B1 (en) 2019-03-19
WO2014110266A1 (en) 2014-07-17
JP6445458B2 (en) 2018-12-26
JP2016508275A (en) 2016-03-17
US11662936B2 (en) 2023-05-30
JP6767115B2 (en) 2020-10-14

Similar Documents

Publication Publication Date Title
US10013317B1 (en) Restoring a volume in a storage system
US11775392B2 (en) Indirect replication of a dataset
US11803567B1 (en) Restoration of a dataset from a cloud
EP3120235B1 (en) Remote replication using mediums
US10430282B2 (en) Optimizing replication by distinguishing user and system write activity
US10482061B1 (en) Removing invalid data from a dataset in advance of copying the dataset
US9864761B1 (en) Read optimization operations in a storage system

Legal Events

Date Code Title Description
AS Assignment

Owner name: PURE STORAGE, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:COLGROVE, JOHN;MILLER, ETHAN;HAYES, JOHN;AND OTHERS;SIGNING DATES FROM 20131009 TO 20131021;REEL/FRAME:031478/0063

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION