US11429495B2 - Data recovery mechanisms in deduplication-enabled storage facilities - Google Patents
Data recovery mechanisms in deduplication-enabled storage facilities Download PDFInfo
- Publication number
- US11429495B2 US11429495B2 US16/819,225 US202016819225A US11429495B2 US 11429495 B2 US11429495 B2 US 11429495B2 US 202016819225 A US202016819225 A US 202016819225A US 11429495 B2 US11429495 B2 US 11429495B2
- Authority
- US
- United States
- Prior art keywords
- copy
- source
- references
- data
- referencing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/14—Error detection or correction of the data by redundancy in operation
- G06F11/1402—Saving, restoring, recovering or retrying
- G06F11/1446—Point-in-time backing up or restoration of persistent data
- G06F11/1448—Management of the data involved in backup or backup restore
- G06F11/1453—Management of the data involved in backup or backup restore using de-duplication of the data
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/14—Error detection or correction of the data by redundancy in operation
- G06F11/1402—Saving, restoring, recovering or retrying
- G06F11/1446—Point-in-time backing up or restoration of persistent data
- G06F11/1458—Management of the backup or restore process
- G06F11/1469—Backup restoration techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/14—Error detection or correction of the data by redundancy in operation
- G06F11/1402—Saving, restoring, recovering or retrying
- G06F11/1471—Saving, restoring, recovering or retrying involving logging of persistent data for recovery
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2053—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
- G06F11/2056—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2053—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
- G06F11/2056—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring
- G06F11/2087—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring with a common controller
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/21—Design, administration or maintenance of databases
- G06F16/215—Improving data quality; Data cleansing, e.g. de-duplication, removing invalid entries or correcting typographical errors
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2201/00—Indexing scheme relating to error detection, to error correction, and to monitoring
- G06F2201/82—Solving problems relating to consistency
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2201/00—Indexing scheme relating to error detection, to error correction, and to monitoring
- G06F2201/84—Using snapshots, i.e. a logical point-in-time copy of the data
Definitions
- Storage facilities may for example be based on a pool of multiple volumes or several pools grouped together in a storage system, or multiple such storage systems.
- Deduplication in a storage facility is the mechanism of storing references to source grains in metadata whenever a new write's IO data matches data already written by a grain, instead of carrying out the write conventionally which would result in the same data being stored again.
- a grain that contains a reference to a source is called a referrer.
- an arbitrary number of referrers from zero upward may point to a common source.
- a set of grains linked in this way is referred to herein as a “referencing domain.”
- a recovery mechanism After a disaster in a storage facility, a recovery mechanism will be applied to identify corrupted source data in the disaster domain and to restore as much of it as possible.
- the principal resources available for a recovery are a snapshot of the pool and a journal.
- a journal is a sequential log of IO operations carried out on the pool. The journal may be replayed from the time when the snapshot was taken to rebuild the system state closer to the time of the disaster.
- a standard approach to recovery is to rewrite a full set of the volume's data from a copy, such as a mirror copy or a backup copy.
- the deduplication metadata references are lost and data recovery is restricted to what is in the backup or mirror copy.
- the method may comprise identifying a first source in a data storage facility that, following a disaster event, is pointing to data which is corrupt.
- the method may further comprise determining that a first copy services relationship exists between a first referencing domain of the first source and a second referencing domain, where the first copy services relationship indicates the second referencing domain will have a second copy of the corrupted data.
- the method may further comprise determining whether the second copy is valid.
- the method may further comprise writing the second copy to replace the corrupted data in response to determining that the second copy is valid.
- Some embodiments of the present disclosure can also be illustrated as a computer program product comprising a computer readable storage medium having program instructions embodied therewith, the program instructions executable by a computer to cause the computer to perform the method discussed above.
- the CPU may identify a first source in a data storage facility that, following a disaster event, is pointing to data which is corrupt.
- the CPU may further determine that a first copy services relationship exists between a first referencing domain of the first source and a second referencing domain, where the first copy services relationship indicates the second referencing domain will have a second copy of the corrupted data.
- the CPU may further determine whether the second copy is valid.
- the CPU may further write the second copy to replace the corrupted data in response to determining that the second copy is valid.
- FIG. 1 depicts a generic backup process using a storage controller and two storage disks VDISK 1 and VDISK 2 according to several embodiments of the present disclosure
- FIG. 2 illustrates an example volume recovery method consistent with several embodiments of the present disclosure
- FIG. 3A shows a first example system with first and second pools according to several embodiments of the present disclosure
- FIG. 3B shows the first example system at an intermediate stage in recovery according to several embodiments of the present disclosure
- FIG. 3C shows the first example system at a final stage in recovery according to several embodiments of the present disclosure
- FIG. 4A shows a second example system with three pools, Pool 0 , Pool 1 , and Pool 2 , according to several embodiments of the present disclosure
- FIG. 4B shows the second example system in an intermediate stage in recovery according to several embodiments of the present disclosure
- FIG. 4C shows the second example system at a final stage in recovery according to several embodiments of the present disclosure
- FIG. 5A shows a third example system with three pools, Pool 0 , Pool 1 , and Pool 2 , according to several embodiments of the present disclosure
- FIG. 5B shows the third example system in an intermediate stage in recovery according to several embodiments of the present disclosure
- FIG. 5C shows the third example system in a next stage in recovery according to several embodiments of the present disclosure
- FIG. 5D shows the third example system in another stage in recovery according to several embodiments of the present disclosure
- FIG. 5E shows the third example system at a final stage in recovery according to several embodiments of the present disclosure
- FIG. 6 depicts a cloud computing environment according to an embodiment of the present disclosure
- FIG. 7 depicts abstraction model layers according to an embodiment of the present disclosure.
- FIG. 8 illustrates a high-level block diagram of an example computer system that may be used in implementing embodiments of the present disclosure.
- aspects of the present disclosure relate to data recovery mechanisms in storage facilities that operate with deduplication. More particular aspects of the present disclosure relate to methods and systems for performing recovery of corrupted data in deduplication-enabled storage volumes in which intact and valid references can be preserved, thereby avoiding the need for a full rewrite and loss of these still-valid references.
- a “source,” as used herein, a source refers to metadata containing the location of written data and a count of the number of its referrers.
- a “referrer,” as used herein, refers to metadata containing a reference to source metadata for the data that has been written to that grain.
- a “reference,” as used herein, refers to an address in metadata.
- Methodadata refers to data about the data written to a specific location, specifically in this disclosure in the context of the reference and the referrer count of each grain.
- the reference can either be the address of the location of the host data, in which case the grain metadata is a source. Or the reference can be to the source metadata which has previously written this host data, in which case the grain metadata is a referrer.
- Gram refers to a unit of address space that is the subject of IO operations, e.g. writes and reads.
- a grain is effectively the unit of granularity of the deduplication engine.
- Referencing domain refers to a set of grains that refer to a common source grain including the source grain and its referring grains.
- a “(storage) volume,” as used herein, refers to a singular logical block address (LBA) range presented either to a host or internally.
- LBA logical block address
- volume copy refers to a copy of a volume, e.g. a copy generated by mirroring, backup, or other copy service.
- a “copy service,” as used herein, refers to a service that maintains multiple copies of data across domains or systems.
- a “(storage) pool,” as used herein, refers to a grouping of volumes using the same backend or internal storage.
- a “snapshot,” as used herein, refers to a copy of a volume or other LBA range taken at a given point in time.
- a “storage system,” as used herein, refers to a cluster for one or more storage pools.
- the recovery mechanisms described herein presuppose the existence of one or more copy services relationships that maintain copies of the same data at two or more different address ranges, e.g. in two different volumes or pools, such as may be provided by mirroring or by making backup copies.
- Copy services allow storage facilities to have, amongst other things, disaster recovery integrated into the IO path.
- IBM Spectrum Virtualize features multiple layers of copy services such as Volume Mirroring, Remote Copy and Flash Copy. Volume Mirroring and Remote Copy maintain ongoing copies of the source data, Flash Copy maintains snapshots of the source volume.
- FIG. 1 depicts a system 100 including a storage controller 108 and two storage disks VDISK 1 ( 110 ) and VDISK 2 ( 112 ) according to several embodiments of the present disclosure.
- the disks 110 and 112 could form part of a larger array of disks and may form part of an enterprise storage solution.
- the disks 110 and 112 could be part of a storage solution relating to a commercial website, for example.
- the storage controller 108 may be provided with a replication function for data backup.
- the replication function may back up local data in a non-disruptive way to another set of local storage devices by using mirroring or point-in-time copies.
- the replication function may backup the data to a remote site.
- Terminology in the art refers to a primary site and a secondary site for data storage, where the primary site is where the original or master copy is located and the secondary site is where the backup copy is located.
- Terminology in the art also refers to a source volume and a target volume, where data is transferred from the source to the target when performing a backup or mirroring operation.
- the term destination volume is a synonym for target volume.
- Example storage controllers with a replication function are the IBM SAN Volume Controller (SVC) or Storage RAID arrays such as the IBM Storwize (IBM trademark) products.
- Example mirroring or point-in-time copy technology is FlashCopy (IBM trademark) and Global Mirror with Change Volumes (GMCV).
- Example remote site data backup technology is HyperSwap (IBM trademark), Global Mirror (IBM trademark) or Metro Mirror (IBM trademark). Storage will typically be organized in virtualized storage arrays implemented with, for example the IBM SVC.
- An example system running a copy service with an in-sync relationship may encounter a recovery event.
- the system may have several deduplication volumes in a referencing domain as well as a journal.
- the system may further support deduplication metadata and data.
- FIG. 2 illustrates an example volume recovery method 200 consistent with several embodiments of the present disclosure.
- Method 200 includes identifying corrupt grains at operation 202 .
- Operation 202 may include, for example, performing a journal replay to rebuild forward lookup metadata.
- the journal entries may be replayed from the time of the last snapshot and processed by a Deduplication and Data Reduction recovery mechanism in order to rebuild the metadata.
- Operation 202 may also include reading a data disk as part of this rebuild mechanism to identify corrupt address ranges. Corrupt grains are thus identified for which the source data is corrupt but the metadata is valid.
- Method 200 further includes identifying a referencing domain of each corrupt grain at operation 204 .
- the referencing domain for each of these grains is identified, i.e. the set of all referrers and source that are linked and point to corrupt source data. If a grain is the source, any references to the grain have been created subsequent to the write. Thus, the journal recovery mechanism may identify corrupt sources as well as their referrers, as the references created need to be replayed too. Under the proposed scheme, entries to the journal may be added documenting the referrer virtual address and volume-id when incrementing/decrementing reference counts for a given chunk. Scanning ahead in the journal may thus enable identifying the referencing domain for any given source.
- Operation 204 may include a heuristic approach, where volumes document at a relatively granular level which volumes they refer to. Such an approach may allow the identification of certain referencing domain members, for example by checking whether the same virtual address on both volumes contains deduplication matches as may be the case for certain virtual desktop infrastructure (VDI) environments.
- VDI virtual desktop infrastructure
- operation 204 may be made trivial, albeit at the cost of additional metadata overheads.
- Method 200 further includes finding a valid copy of the source data at operation 206 .
- Operation 206 may include, for example, determining whether there is a valid copy of the source data for this referencing domain, i.e. the source grain of the referencing domain, by searching of all volumes in the referencing domain and their respective grains. Checking for a valid copy can be implemented by checking that the volume copy LBA ranges that are referred to by the referencing domain for this copy are valid. This is a test that the data on that copy is up to date (in copy services terms one would say ‘synchronized’) and not corrupt itself. Having identified the volumes, copy services can provide the identity of the other copy. The valid data could be a previously recovered grain in a different deduplication pool.
- Method 200 further includes writing a valid copy to address space accessible by the reference domain at operation 208 .
- Operation 208 may include reading and writing (i.e., copying) the valid copy of the grain maintained by the copy service which has been found to address space that is accessible by the referencing domain.
- the valid copy may be written to the data disk of the referencing domain such as the volume where the corrupted referencing domain stores its data.
- Method 200 further includes changing the source reference point to the valid copy at operation 210 .
- the source reference for the referencing domain may be changed (at operation 210 ) to reference this new address location of the valid data that has been copied across. All other metadata can be left intact.
- the system may then record that the grain has been recovered with valid data. Recording the recovery of the referencing domain may allow subsequent iterations of the same recovery mechanism to use the recovered referencing domain to repair other referencing domains that have not yet been recovered with which the recovered referencing domain has a copy services relationship.
- this step of changing the source reference pointer can be a useful option, since when copying across from the valid data copy, it ensures that the data is written to fresh extents (i.e. address ranges) on the data disk to avoid the risk of overwriting valid data belonging to other volumes.
- operation 210 may be omitted and the data overwritten in place, i.e. using the same address range.
- an overwrite-in-place procedure should additionally be tested, e.g. using checksums, to avoid overwriting any data that itself contains references.
- the recovery mechanism as described in method 200 above is applied to each corrupt grain in turn.
- the recovery mechanism can be repeated as many times as useful, e.g. until it is detected that the latest iteration has resulted in no further recovery actions, or until a particular piece of key data is recovered, or until the system is instructed to abandon the recovery attempt.
- the recovery mechanism may be able to recover the volume copy completely and in addition may be able to partially recover other volume data for volumes that are part of the referencing domain. It should however be noted that the recovery mechanism relies upon the ability to read a copy of the data during the recovery procedure. In the situation that valid data for the volume is located in other copies within the system, then the recovery mechanism may be able to recover the volume copy completely. On the other hand, if the corrupted grain is not associated with other copies of data, i.e. the referencing domain solely consists of the corrupted grain, then the recovery mechanism may not be able to recover the grain. In addition, the recovery mechanism may not be able to recover the copy completely if the copy services relationship was not completely synced at the time of the disaster or was itself partially corrupt.
- the recovery mechanism may still be useful, since it may recover some grains, since the recovery mechanism can still be applied in cases where there is only a partially complete copy available.
- the main goal of the recovery is to restore a particular, highly specific part of the data, so partial recovery is still useful.
- the recovery mechanism can be extended across multiple pools with corruption. As long as there is an in-sync copy of the data without data corruption in one of the pools, then the data can be restored across all the pools and hence the data for that grain in each referencing domain for those pools. This is applicable in the scenario where the system has two or more failure domains, basically any dependency between any number of failure domains is possible so long as no metadata or overlapping source data corruption has occurred in the system.
- FIGS. 3A to 3C show sequential stages in a first example implementation of the recovery mechanism according to the disclosure.
- This is a simple example in which a volume copy using volume mirroring is a referrer to a deduplicated volume for a specific grain. If during recovery: (i) the data for that grain is found to be corrupted, (ii) the metadata is found to be valid, and (iii) the mirrored volume copy was clean prior to this recovery (i.e. in sync, with no data lost by the disaster), then the system will read that grain from the valid data copy and write the valid data. The deduplication metadata for that grain is left intact and the source metadata is updated to point at the recovered valid data. Effectively this will recover data for both the referrer and source grains and hence the corresponding portion of the source volume.
- FIG. 3A shows a first example system 300 with first and second pools according to several embodiments of the present disclosure.
- Pool 0 302 has a Referrer A 308 which is in a Volume A 305 pointing with a reference to a Source B 306 in Volume B 303 which in turn points to data 312 which is corrupt, e.g. as a result of a failure event.
- Pool 1 304 has a Source C 310 which is in a Volume C 307 pointing to data 314 which is valid.
- a copy services relationship 340 as indicated by the dashed outline, e.g. mirroring, exists between Volume A 305 and Volume C 307 .
- the recovery mechanism identifies that the data 312 is corrupt.
- FIG. 3B shows the first example system 300 at an intermediate stage in recovery according to several embodiments of the present disclosure.
- the recovery mechanism having identified that data 312 is corrupt, looks for copies of data 312 . It does this by undertaking a scan by following reverse paths from the deduplication reference to look through the referencing domain to check if any referencing domain members (i.e. grains) are in a volume that has a copy services relationship with another volume, in which case there should be a copy of data 312 .
- the recovery mechanism finds that Volume A 305 of Referrer A 308 in Pool 0 302 has a copy services relationship 340 with Volume C 307 of Source C 310 in Pool 1 304 where a valid data copy of data 312 exists in the form of data 314 .
- FIG. 3C shows the first example system 300 at a final stage in recovery according to several embodiments of the present disclosure.
- the valid data copy of data 314 is copied across to Pool 0 302 to replace data 312 .
- the corruption of data 312 is thus healed.
- FIGS. 4A to 4C show sequential stages in a second example implementation of the recovery mechanism according to the disclosure.
- the second example involves moving valid data across two pools instead of one to heal a corruption.
- FIG. 4A shows a second example system 400 with three pools, Pool 0 402 , Pool 1 404 , and Pool 2 406 according to several embodiments of the present disclosure.
- Data 416 and data 418 (in Pool 0 402 and Pool 1 404 , respectively) are corrupt, whereas data 420 in Pool 3 406 is valid.
- Pool 0 402 has a Referrer A 410 which is in Volume A 405 pointing with a reference to a Source B 408 in Volume B 403 which in turn points to data 416 .
- Pool 2 406 has a Source D 414 pointing to data 420 .
- FIG. 4B shows the second example system 400 in an intermediate stage in recovery according to several embodiments of the present disclosure.
- the recovery mechanism having identified that data 416 is corrupt, looks for copies of data 416 . It does this in the same way as in the first example and follows the referencing domain to Referrer A 410 , finds the first copy services relationship 440 to Volume C 407 of Pool 1 404 , but then identifies that the copy in data 418 is also corrupt so cannot be used for recovery.
- the recovery mechanism finds the second copy services relationship 460 between Volume C 407 of Pool 1 404 and Volume D 409 of Pool 2 406 , and follows that, and identifies that the further data copy data 420 is valid.
- FIG. 4C shows the second example system at a final stage in recovery according to several embodiments of the present disclosure.
- Data 418 is repaired from data 420 by copying and then data 416 is repaired from data 418 by copying again.
- FIGS. 5A to 5E show sequential stages in a third example implementation of the recovery mechanism according to the disclosure.
- This third example is similar to the second example (i.e., the example depicted in FIGS. 4A-4C ) but is one level more complicated in that there is no copy services relationship directly between the data copies in Pool 1 and Pool 2 .
- FIG. 5A shows a third example system 500 with three pools, Pool 0 502 , Pool 1 504 , and Pool 2 506 , according to several embodiments of the present disclosure.
- Data 518 and data 520 (in Pool 0 502 and Pool 1 504 , respectively) are corrupt, whereas data 522 in Pool 3 506 is valid.
- Pool 0 502 has a Referrer A 510 which is in Volume A 505 pointing with a reference to a Source B 508 in Volume B 503 which in turn points to data 518 .
- Pool 1 504 has a Referrer D 514 which is in Volume D 509 pointing with a reference to a Source C 512 which in turn points to data 520 .
- Pool 2 506 has a Source E 516 which points to valid data 522 .
- Copy services relationship 560 indicates that one of the referrers, Referrer D 514 , referring to the source grain, Source C 512 , is in a copy services relationship 560 with Source E 516 in Pool 2 506 .
- the valid data 522 in Pool 2 506 can then be copied across to Pool 1 504 and to Pool 0 502 similar to in the second example (i.e., that depicted in FIGS. 4A-4C , above), as described in further detail below.
- FIG. 5B shows the third example system 500 in an intermediate stage in recovery according to several embodiments of the present disclosure.
- the recovery mechanism having identified that data 518 is corrupt, looks for copies of data 518 . It may do this in the same way as in the previous examples, following the referencing domain to Referrer A 510 , finding the first copy services relationship 540 to Volume C 507 of Pool 1 504 , but then identifies that the copy in data 520 is also corrupt so cannot be used for recovery.
- FIG. 5C shows the third example system 500 in a next stage in recovery according to several embodiments of the present disclosure.
- the recovery mechanism may find a second copy services relationship 560 between Volume D 509 of Pool 1 504 and Volume E 511 of Pool 2 506 . Following relationship 560 , the recovery mechanism may identify that the further data copy data 522 is valid.
- FIG. 5D shows the third example system 500 in another stage in recovery according to several embodiments of the present disclosure.
- data 520 is repaired from data 522 by copying, similar to the repair of data 312 from 314 with reference to FIG. 3C .
- FIG. 5E shows the third example system 500 at a final stage in recovery according to several embodiments of the present disclosure.
- Data 518 is repaired from the newly-repaired data 520 by copying.
- the recovery mechanism according to the disclosure is compatible with and can be used in combination with a recovery mechanism that utilizes journal replay to find the most up-to-date version of data associated with a given grain that survives on the data disk after a disaster. This may be achieved by only allowing journal replay updates to be played where the associated data is valid, if the prior metadata referenced valid data. This is a powerful recovery mechanism, as it is preferable to return old data than no data at all in many circumstances.
- a similar mechanism can be combined with the recovery mechanism proposed in the present disclosure to attempt to recover based-off copy services, i.e. other copies of data available in another location from copy services (e.g. another copy of the volume in a different pool).
- the recovery mechanism according to the present disclosure can be used to discover valid data in other copies across a larger domain (e.g. across multiple pools) as well as within a single domain (e.g. a single pool).
- the recovery mechanism proposed in the present disclosure particularly when combined other recovery mechanisms, allows for the recovery of newer data than would otherwise be possible.
- Cloud computing is a model of service delivery for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, network bandwidth, servers, processing, memory, storage, applications, virtual machines, and services) that can be rapidly provisioned and released with minimal management effort or interaction with a provider of the service.
- This cloud model may include at least five characteristics, at least three service models, and at least four deployment models.
- On-demand self-service a cloud consumer can unilaterally provision computing capabilities, such as server time and network storage, as needed automatically without requiring human interaction with the service's provider.
- Resource pooling the provider's computing resources are pooled to serve multiple consumers using a multi-tenant model, with different physical and virtual resources dynamically assigned and reassigned according to demand There is a sense of location independence in that the consumer generally has no control or knowledge over the exact location of the provided resources but may be able to specify location at a higher level of abstraction (e.g., country, state, or datacenter).
- Rapid elasticity capabilities can be rapidly and elastically provisioned, in some cases automatically, to quickly scale out and rapidly released to quickly scale in. To the consumer, the capabilities available for provisioning often appear to be unlimited and can be purchased in any quantity at any time.
- Measured service cloud systems automatically control and optimize resource use by leveraging a metering capability at some level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth, and active user accounts). Resource usage can be monitored, controlled, and reported, providing transparency for both the provider and consumer of the utilized service.
- level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth, and active user accounts).
- SaaS Software as a Service: the capability provided to the consumer is to use the provider's applications running on a cloud infrastructure.
- the applications are accessible from various client devices through a thin client interface such as a web browser (e.g., web-based e-mail).
- a web browser e.g., web-based e-mail
- the consumer does not manage or control the underlying cloud infrastructure including network, servers, operating systems, storage, or even individual application capabilities, with the possible exception of limited user-specific application configuration settings.
- PaaS Platform as a Service
- the consumer does not manage or control the underlying cloud infrastructure including networks, servers, operating systems, or storage, but has control over the deployed applications and possibly application hosting environment configurations.
- IaaS Infrastructure as a Service
- the consumer does not manage or control the underlying cloud infrastructure but has control over operating systems, storage, deployed applications, and possibly limited control of select networking components (e.g., host firewalls).
- Private cloud the cloud infrastructure is operated solely for an organization. It may be managed by the organization or a third party and may exist on-premises or off-premises.
- Public cloud the cloud infrastructure is made available to the general public or a large industry group and is owned by an organization selling cloud services.
- Hybrid cloud the cloud infrastructure is a composition of two or more clouds (private, community, or public) that remain unique entities but are bound together by standardized or proprietary technology that enables data and application portability (e.g., cloud bursting for load-balancing between clouds).
- a cloud computing environment is service oriented with a focus on statelessness, low coupling, modularity, and semantic interoperability.
- An infrastructure that includes a network of interconnected nodes.
- cloud computing environment 600 comprises one or more cloud computing nodes 610 with which local computing devices used by cloud consumers, such as, for example, personal digital assistant (PDA) or cellular telephone 640 A, desktop computer 640 B, laptop computer 640 C, and/or automobile computer system 640 N may communicate.
- Nodes 610 may communicate with one another. They may be grouped (not shown) physically or virtually, in one or more networks, such as Private, Community, Public, or Hybrid clouds as described hereinabove, or a combination thereof.
- This allows cloud computing environment 600 to offer infrastructure, platforms and/or software as services for which a cloud consumer does not need to maintain resources on a local computing device.
- computing devices 640 A-N shown in FIG. 6 are intended to be illustrative only and that computing nodes 610 and cloud computing environment 600 can communicate with any type of computerized device over any type of network and/or network addressable connection (e.g., using a web browser).
- FIG. 7 a set of functional abstraction layers provided by cloud computing environment 600 ( FIG. 6 ) is shown. It should be understood in advance that the components, layers, and functions shown in FIG. 7 are intended to be illustrative only and embodiments of the invention are not limited thereto. As depicted, the following layers and corresponding functions are provided:
- Hardware and software layer 760 includes hardware and software components.
- hardware components include: mainframes 761 ; RISC (Reduced Instruction Set Computer) architecture based servers 762 ; servers 763 ; blade servers 764 ; storage devices 765 ; and networks and networking components 766 .
- software components include network application server software 767 and database software 768 .
- Virtualization layer 770 provides an abstraction layer from which the following examples of virtual entities may be provided: virtual servers 771 ; virtual storage 772 ; virtual networks 773 , including virtual private networks; virtual applications and operating systems 774 ; and virtual clients 775 .
- management layer 780 may provide the functions described below.
- Resource provisioning 781 provides dynamic procurement of computing resources and other resources that are utilized to perform tasks within the cloud computing environment.
- Metering and Pricing 782 provide cost tracking as resources are utilized within the cloud computing environment, and billing or invoicing for consumption of these resources. In one example, these resources may comprise application software licenses.
- Security provides identity verification for cloud consumers and tasks, as well as protection for data and other resources.
- User portal 783 provides access to the cloud computing environment for consumers and system administrators.
- Service level management 784 provides cloud computing resource allocation and management such that required service levels are met.
- Service Level Agreement (SLA) planning and fulfillment 785 provide pre-arrangement for, and procurement of, cloud computing resources for which a future requirement is anticipated in accordance with an SLA.
- SLA Service Level Agreement
- Workloads layer 790 provides examples of functionality for which the cloud computing environment may be utilized. Examples of workloads and functions which may be provided from this layer include: mapping and navigation 791 ; software development and lifecycle management 792 ; virtual classroom education delivery 793 ; data analytics processing 794 ; transaction processing 795 ; and storage facility disaster recovery 796 .
- FIG. 8 shown is a high-level block diagram of an example computer system 800 that may be configured to perform various aspects of the present disclosure, including, for example, method 200 .
- the example computer system 800 may be used in implementing one or more of the methods or modules, and any related functions or operations, described herein (e.g., using one or more processor circuits or computer processors of the computer), in accordance with embodiments of the present disclosure.
- the major components of the computer system 800 may comprise one or more CPUs 802 , a memory subsystem 808 , a terminal interface 816 , a storage interface 818 , an I/O (Input/Output) device interface 820 , and a network interface 822 , all of which may be communicatively coupled, directly or indirectly, for inter-component communication via a memory bus 806 , an I/O bus 814 , and an I/O bus interface unit 812 .
- CPUs 802 the major components of the computer system 800 may comprise one or more CPUs 802 , a memory subsystem 808 , a terminal interface 816 , a storage interface 818 , an I/O (Input/Output) device interface 820 , and a network interface 822 , all of which may be communicatively coupled, directly or indirectly, for inter-component communication via a memory bus 806 , an I/O bus 814 , and an I/O bus interface unit 812 .
- the computer system 800 may contain one or more general-purpose programmable central processing units (CPUs) 802 , some or all of which may include one or more cores 804 A, 804 B, 804 C, and 804 D, herein generically referred to as the CPU 802 .
- the computer system 800 may contain multiple processors typical of a relatively large system; however, in other embodiments the computer system 800 may alternatively be a single CPU system.
- Each CPU 802 may execute instructions stored in the memory subsystem 808 on a CPU core 804 and may comprise one or more levels of on-board cache.
- the memory subsystem 808 may comprise a random-access semiconductor memory, storage device, or storage medium (either volatile or non-volatile) for storing data and programs.
- the memory subsystem 808 may represent the entire virtual memory of the computer system 800 and may also include the virtual memory of other computer systems coupled to the computer system 800 or connected via a network.
- the memory subsystem 808 may be conceptually a single monolithic entity, but, in some embodiments, the memory subsystem 808 may be a more complex arrangement, such as a hierarchy of caches and other memory devices.
- memory may exist in multiple levels of caches, and these caches may be further divided by function, so that one cache holds instructions while another holds non-instruction data, which is used by the processor or processors.
- Memory may be further distributed and associated with different CPUs or sets of CPUs, as is known in any of various so-called non-uniform memory access (NUMA) computer architectures.
- NUMA non-uniform memory access
- the main memory or memory subsystem 804 may contain elements for control and flow of memory used by the CPU 802 . This may include a memory controller 810 .
- the memory bus 806 is shown in FIG. 8 as a single bus structure providing a direct communication path among the CPU 802 , the memory subsystem 808 , and the I/O bus interface 812
- the memory bus 806 may, in some embodiments, comprise multiple different buses or communication paths, which may be arranged in any of various forms, such as point-to-point links in hierarchical, star or web configurations, multiple hierarchical buses, parallel and redundant paths, or any other appropriate type of configuration.
- the I/O bus interface 812 and the I/O bus 814 are shown as single respective units, the computer system 800 may, in some embodiments, contain multiple I/O bus interface units 812 , multiple I/O buses 814 , or both.
- multiple I/O interface units are shown, which separate the I/O bus 814 from various communications paths running to the various I/O devices, in other embodiments some or all of the I/O devices may be connected directly to one or more system I/O buses.
- the computer system 800 may be a multi-user mainframe computer system, a single-user system, or a server computer or similar device that has little or no direct user interface but receives requests from other computer systems (clients). Further, in some embodiments, the computer system 800 may be implemented as a desktop computer, portable computer, laptop or notebook computer, tablet computer, pocket computer, telephone, smart phone, mobile device, or any other appropriate type of electronic device.
- FIG. 8 is intended to depict the representative major components of an exemplary computer system 800 . In some embodiments, however, individual components may have greater or lesser complexity than as represented in FIG. 8 , components other than or in addition to those shown in FIG. 8 may be present, and the number, type, and configuration of such components may vary.
- the present invention may be a system, a method, and/or a computer program product at any possible technical detail level of integration
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, configuration data for integrated circuitry, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++, or the like, and procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
- the functions noted in the blocks may occur out of the order noted in the Figures.
- two blocks shown in succession may, in fact, be accomplished as one step, executed concurrently, substantially concurrently, in a partially or wholly temporally overlapping manner, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- a method for recovering data after a disaster event which has corrupted data in a storage facility of the kind that both operates with deduplication and also that maintains multiple copies of data through establishing copy services relationships between different storage volumes wherein the deduplication generates referencing domains which are formed through references between grains, each referencing domain having a source grain containing a reference pointing to a location where data is stored and at least one referrer pointing to the source, the data recovery method comprising identifying a source that following a disaster event is pointing to data which is corrupt, establishing whether a copy services relationship exists between the referencing domain of the source with corrupted data and another referencing domain in which case the other referencing domain will have a copy of the corrupted data, checking that the copy is valid, and, if ‘yes’, writing the copy to replace the corrupted data.
- establishing whether a copy services relationship exists between the referencing domain of the source with corrupted data and another referencing domain includes checking for a copy services relationship directly between the source with the corrupted data and another referencing domain.
- establishing whether a copy services relationship exists between the referencing domain of the source with corrupted data and another referencing domain includes scanning the references to the source with the corrupted data to check whether a copy services relationship exists between any one of its referrers and another referencing domain.
- the storage facility may implement deduplication in that the references consist only of forward references, a forward reference being held by each referrer to point to its source, and wherein scanning the references to crawl over the referencing domain involves reversing the forward references and following the reversed forward references.
- the storage facility may implement deduplication in that the references include both forward and backward references, a forward reference being held by each referrer to point to its source, and backward references being held by each source to point to its referrers, wherein scanning the references to crawl over the referencing domain involves following the backward references.
- the method further comprises: establishing whether a copy services relationship exists between the referencing domain of the copy which is also corrupt and a still further referencing domain, in which case the still further referencing domain will have a further copy of the corrupted data; checking that the further copy is valid; if ‘yes’, writing the further copy to replace the corrupted data of both the copy and the originally identified source; and repairing the source references to point to the written copies.
- the data recovery method may further comprise repairing the source reference to point to the written copy.
- the data recovery method may be implemented such that said checking that the copy is valid comprises checking that address ranges that are referred to by the referencing domain for this copy are valid.
- the copy services relationship may be between storage at different levels in the storage hierarchy.
- the storage may be in a three-level hierarchy of volumes, pools and storage systems and a copy services relationship, such as remote copy, may exist between storage volumes, between storage pools, and/or between storage systems.
- a copy services relationship between different storage systems will allow recovery of data from another storage system in cases where there are extra system copies of valid data.
- “Storage system,” as used herein, typically refers to sets of nodes forming a cluster that manage a localized pool of resources.
- a customer can have copies of volumes between multiple clusters, for example different physical sites or power domains. In this case, as long as the copy was in sync, i.e. valid, prior to the disaster then data can still be recovered according to methods described in the present disclosure.
- An example here is remote copy or global mirror. The data copies can be within the same cluster or across multiple clusters.
- Corrupted source data can be recovered in embodiments of the disclosure by examining a copy service that has been running for the disaster domain to find a (non-corrupted) copy of the corrupt data in the referencing domain, writing the data from the non-corrupted data to the pool where the corrupt data is located as restored data and amending the metadata in the source to point to the restored data instead of the corrupted data.
- Amending the metadata is generally required, since it will not be safe to overwrite the corrupted data with the non-corrupted data, but rather the non-corrupted data should be copied into a different, unallocated address space range.
- the proposed approach thus provides a recovery mechanism that allow recovery of data in a deduplicated system via reads to volume copies established by a copy service in another storage pool or JO group. This is achieved in a fashion that restores data and maintains valid deduplication references in order to recover data on all volumes in the same referencing domain as the volume copy.
- a benefit of a recovery mechanism consistent with the present disclosure is its ability to recover corrupted source data without performing a full resync or restore. Since the referencing metadata is “a priori” valid, since that is a prerequisite of applying the approach described herein, source data can be recovered for any volumes where referrers are located. With this approach, a storage facility operating with deduplication and configured to have multiple, i.e. at least two, separate failure domains between which copies of source data are maintained would be able to recover completely from a disaster that was confined to source data corruption.
- a computer program stored on a computer readable medium and loadable into the internal memory of a computer, comprising software code portions, when said program is run on a computer, for performing the above-described method.
- a computer program product storing the computer program may also be provided.
- a storage facility comprising a plurality of storage volumes that are configured jointly to maintain copies of data between them through copy services relationships, each storage volume being configured to operate with deduplication so that in use referencing domains are formed through references between grains, each referencing domain having a source grain containing a reference pointing to a location where data is stored and at least one referrer pointing to the source, the storage facility having a data recovery mechanism operable recover data after a disaster event that is configured to identify a source that following a disaster event is pointing to data which is corrupt, establish whether a copy services relationship exists between the referencing domain of the source with corrupted data and another referencing domain in which case the other referencing domain will have a copy of the corrupted data, check that the copy is valid, and, if ‘yes’, write the copy to replace the corrupted data.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Quality & Reliability (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims (16)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/819,225 US11429495B2 (en) | 2020-03-16 | 2020-03-16 | Data recovery mechanisms in deduplication-enabled storage facilities |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/819,225 US11429495B2 (en) | 2020-03-16 | 2020-03-16 | Data recovery mechanisms in deduplication-enabled storage facilities |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20210286687A1 US20210286687A1 (en) | 2021-09-16 |
| US11429495B2 true US11429495B2 (en) | 2022-08-30 |
Family
ID=77664848
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/819,225 Active 2040-12-11 US11429495B2 (en) | 2020-03-16 | 2020-03-16 | Data recovery mechanisms in deduplication-enabled storage facilities |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US11429495B2 (en) |
Citations (17)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100070478A1 (en) * | 2008-09-15 | 2010-03-18 | International Business Machines Corporation | Retrieval and recovery of data chunks from alternate data stores in a deduplicating system |
| US8677178B2 (en) | 2008-11-20 | 2014-03-18 | International Business Machines Corporation | Recovery control in mirrored disks |
| US20140143786A1 (en) * | 2010-12-09 | 2014-05-22 | International Business Machines Corporation | Management of copy services relationships via policies specified on resource groups |
| US9367398B1 (en) | 2014-03-28 | 2016-06-14 | Emc Corporation | Backing up journal data to a memory of another node |
| US20160203055A1 (en) * | 2015-01-14 | 2016-07-14 | International Business Machines Corporation | Synchronized flashcopy backup restore of a raid protected array |
| US20170083250A1 (en) * | 2015-09-21 | 2017-03-23 | International Business Machines Corporation | Copy-redirect on write |
| US20170083404A1 (en) * | 2015-09-21 | 2017-03-23 | International Business Machines Corporation | Point-in-time copy on write for golden image |
| US10037154B2 (en) | 2011-08-01 | 2018-07-31 | Actifio, Inc. | Incremental copy performance between data stores |
| US10176190B2 (en) | 2015-01-29 | 2019-01-08 | SK Hynix Inc. | Data integrity and loss resistance in high performance and high capacity storage deduplication |
| US10210062B2 (en) | 2017-06-08 | 2019-02-19 | International Business Machines Corporation | Data storage system comprising an array of drives |
| US10255137B1 (en) | 2013-12-16 | 2019-04-09 | EMC IP Holding Company LLC | Point-in-time recovery on deduplicated storage |
| US10339112B1 (en) * | 2013-04-25 | 2019-07-02 | Veritas Technologies Llc | Restoring data in deduplicated storage |
| US20200097179A1 (en) * | 2018-09-20 | 2020-03-26 | Hitachi, Ltd. | Storage controller and storage control method |
| US20200241753A1 (en) * | 2019-01-25 | 2020-07-30 | International Business Machines Corporation | Recovering from data loss using copy services relationships between volumes |
| US20210064247A1 (en) * | 2019-08-30 | 2021-03-04 | International Business Machines Corporation | Managing capacity in a storage system using copy services |
| US20210173811A1 (en) * | 2019-12-04 | 2021-06-10 | Commvault Systems, Inc. | Optimizing the restoration of deduplicated data stored in multi-node replicated file systems |
| US20210279141A1 (en) * | 2019-04-10 | 2021-09-09 | Commvault Systems, Inc. | Restore using deduplicated secondary copy data |
-
2020
- 2020-03-16 US US16/819,225 patent/US11429495B2/en active Active
Patent Citations (25)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8290915B2 (en) | 2008-09-15 | 2012-10-16 | International Business Machines Corporation | Retrieval and recovery of data chunks from alternate data stores in a deduplicating system |
| US20100070478A1 (en) * | 2008-09-15 | 2010-03-18 | International Business Machines Corporation | Retrieval and recovery of data chunks from alternate data stores in a deduplicating system |
| US8677178B2 (en) | 2008-11-20 | 2014-03-18 | International Business Machines Corporation | Recovery control in mirrored disks |
| US20140143786A1 (en) * | 2010-12-09 | 2014-05-22 | International Business Machines Corporation | Management of copy services relationships via policies specified on resource groups |
| US10037154B2 (en) | 2011-08-01 | 2018-07-31 | Actifio, Inc. | Incremental copy performance between data stores |
| US10339112B1 (en) * | 2013-04-25 | 2019-07-02 | Veritas Technologies Llc | Restoring data in deduplicated storage |
| US10255137B1 (en) | 2013-12-16 | 2019-04-09 | EMC IP Holding Company LLC | Point-in-time recovery on deduplicated storage |
| US9367398B1 (en) | 2014-03-28 | 2016-06-14 | Emc Corporation | Backing up journal data to a memory of another node |
| US20160203055A1 (en) * | 2015-01-14 | 2016-07-14 | International Business Machines Corporation | Synchronized flashcopy backup restore of a raid protected array |
| US10176190B2 (en) | 2015-01-29 | 2019-01-08 | SK Hynix Inc. | Data integrity and loss resistance in high performance and high capacity storage deduplication |
| US10209910B2 (en) * | 2015-09-21 | 2019-02-19 | International Business Machines Corporation | Copy-redirect on write |
| US9886349B2 (en) * | 2015-09-21 | 2018-02-06 | International Business Machines Corporation | Point-in-time copy on write for golden image |
| US20180067815A1 (en) * | 2015-09-21 | 2018-03-08 | International Business Machines Corporation | Point-in-time copy on write for golden image |
| US10042714B2 (en) * | 2015-09-21 | 2018-08-07 | International Business Machines Corporation | Point-in-time copy on write for golden image |
| US20170083250A1 (en) * | 2015-09-21 | 2017-03-23 | International Business Machines Corporation | Copy-redirect on write |
| US20170083404A1 (en) * | 2015-09-21 | 2017-03-23 | International Business Machines Corporation | Point-in-time copy on write for golden image |
| US20180165027A1 (en) * | 2015-09-21 | 2018-06-14 | International Business Machines Corporation | Copy-redirect on write |
| US9940041B2 (en) * | 2015-09-21 | 2018-04-10 | International Business Machines Corporation | Copy-redirect on write |
| US10210062B2 (en) | 2017-06-08 | 2019-02-19 | International Business Machines Corporation | Data storage system comprising an array of drives |
| US20200097179A1 (en) * | 2018-09-20 | 2020-03-26 | Hitachi, Ltd. | Storage controller and storage control method |
| US10817209B2 (en) * | 2018-09-20 | 2020-10-27 | Hitachi, Ltd. | Storage controller and storage control method |
| US20200241753A1 (en) * | 2019-01-25 | 2020-07-30 | International Business Machines Corporation | Recovering from data loss using copy services relationships between volumes |
| US20210279141A1 (en) * | 2019-04-10 | 2021-09-09 | Commvault Systems, Inc. | Restore using deduplicated secondary copy data |
| US20210064247A1 (en) * | 2019-08-30 | 2021-03-04 | International Business Machines Corporation | Managing capacity in a storage system using copy services |
| US20210173811A1 (en) * | 2019-12-04 | 2021-06-10 | Commvault Systems, Inc. | Optimizing the restoration of deduplicated data stored in multi-node replicated file systems |
Non-Patent Citations (2)
| Title |
|---|
| Mell et al., "The NIST Definition of Cloud Computing," Recommendations of the National Institute of Standards and Technology, U.S. Department of Commerce, Special Publication 800-145, Sep. 2011, 7 pages. |
| Rostagni et al., "Data Validation During Data Recovery in a Log-Structured Array Storage System," U.S. Appl. No. 16/548,474, filed Aug. 22, 2019, 36 pages. |
Also Published As
| Publication number | Publication date |
|---|---|
| US20210286687A1 (en) | 2021-09-16 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11132264B2 (en) | Point-in-time copy restore | |
| US10725976B2 (en) | Fast recovery using self-describing replica files in a distributed storage system | |
| US8856472B2 (en) | Restore in cascaded copy environment | |
| US11030060B2 (en) | Data validation during data recovery in a log-structured array storage system | |
| US10216429B2 (en) | Performing post-processing operations for log file writes | |
| US11003559B2 (en) | Underperforming drive handling in redundant arrays | |
| US9760449B2 (en) | Restoring a point-in-time copy | |
| US20190188309A1 (en) | Tracking changes in mirrored databases | |
| US9760450B2 (en) | Restoring a clone point-in-time copy | |
| US12045173B2 (en) | Stale data recovery using virtual storage metadata | |
| US9158712B2 (en) | Instantaneous save/restore of virtual machines with persistent memory | |
| US11429495B2 (en) | Data recovery mechanisms in deduplication-enabled storage facilities | |
| US11175999B2 (en) | Management of backup volume extents via a tiered storage mechanism | |
| US11640339B2 (en) | Creating a backup data set | |
| WO2022188613A1 (en) | Recording changes to records whilst preserving record immutability | |
| US11593237B2 (en) | Fast recovery with enhanced raid protection | |
| US10740203B2 (en) | Aggregation of updated tracks to be copied to a backup volume for physically contiguous storage on a RAID stride | |
| CN110188005B (en) | Host creating method, data backup method, device, electronic equipment and storage medium | |
| US20200110819A1 (en) | Low cost fast recovery index in storage class memory | |
| US20200097372A1 (en) | Persistent storage segment cache for recovery |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TOMKINS, DOMINIC;BARTLETT, ERIC JOHN;MULHOLLAND, MILES;AND OTHERS;REEL/FRAME:052119/0701 Effective date: 20200313 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |