WO2016028757A2 - Journal de transactions multifilière pour le domaine primaire et le rétablissement/le renseignement - Google Patents

Journal de transactions multifilière pour le domaine primaire et le rétablissement/le renseignement Download PDF

Info

Publication number
WO2016028757A2
WO2016028757A2 PCT/US2015/045659 US2015045659W WO2016028757A2 WO 2016028757 A2 WO2016028757 A2 WO 2016028757A2 US 2015045659 W US2015045659 W US 2015045659W WO 2016028757 A2 WO2016028757 A2 WO 2016028757A2
Authority
WO
WIPO (PCT)
Prior art keywords
data
node
primary
restore
access
Prior art date
Application number
PCT/US2015/045659
Other languages
English (en)
Other versions
WO2016028757A3 (fr
Inventor
Kumar Kanteti
Paula Long
Kannan Sasi
Gregory James MCHALE
Original Assignee
DataGravity, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by DataGravity, Inc. filed Critical DataGravity, Inc.
Publication of WO2016028757A2 publication Critical patent/WO2016028757A2/fr
Publication of WO2016028757A3 publication Critical patent/WO2016028757A3/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1471Saving, restoring, recovering or retrying involving logging of persistent data for recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/815Virtual
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/84Using snapshots, i.e. a logical point-in-time copy of the data

Definitions

  • HA High Availability
  • the analytics may enable other functions such as data intelligence.
  • primary data is read from and written to a primary storage pool. As the data is written to the primary pool it is automatically mirrored and also tracked for data protection to a recovery pool.
  • the mirror can also be used for intelligence including analytics stored as discovery points.
  • the techniques disclosed herein relate to a system that merges primary data storage, data protection, and intelligence into a single unified system.
  • the unified system provides primary and restore data, analytics, and analytics-based data protection without requiring separate solutions for each aspect.
  • Intelligence is provided through inline data analytics, with additional data intelligence and analytics gathered on protected data and prior analytics, and stored in discovery points, all without impacting performance of primary storage.
  • the disclosed system implements: multi-threaded log writes across primary and restore nodes; nested virtual machine directories, where subdirectories are associated with a virtual structure that corresponds to a file system for snapshot purposes; file system clone available on demand with background metadata and data migration; and /or write gathering across file systems/nodes.
  • that multi-threaded log writes are implemented at a protection and analytics (PART) node.
  • the PART node receives access requests from multiple concurrently executing threads, and assigns a transaction identifier (ID) to the access requests.
  • the PART then collects the access requests in a random access, multithreaded log before sending them to both a primary and a restore storage system. Subsequently, the PART forwards the access requests from the PART node to the primary node and restore node.
  • ID transaction identifier
  • the PART may further optionally determine when a number of access requests in the random access, multithreaded log reaches a predetermined number. At that time, the PART issues a synchronization command to the primary and restore nodes which causes data to be flushed from respective temporary caches to a persistent file system in each of the primary and restore. Once data is confirmed as having been flushed in both the primary and restore nodes, the PART may then release entries in the random access, multithreaded log.
  • the PART maintains a set of file system level objects, one for each subdirectory in a directory tree created by an application, such as a hypervisor.
  • the PART intercepts a make directory request from the application to store a new a file system level object for each subdirectory in the tree.
  • the file system level object contains access information for the corresponding subdirectory, such that multiple make directory requests result in storing a corresponding multiple number of file system level objects as a virtual file system.
  • a file system object located with a subdirectory are then serviced by the primary and restore nodes using only the virtual file system level object information and not the subdirectory directly. This ensures that the virtual file system objects remain transparent to the application.
  • a property may be associated with two or more virtual file system objects to indicate that an access request applies to two or more subdirectories as a consistency group.
  • the data-intelligent storage system intercepts a request to clone a data object.
  • a clone object is first thin provisioned and opened for access. Data is copied to the clone object only upon the first to occur of either (a) a subsequent access request for the clone object, or (b) as part of a background restore process.
  • Thin provisioning may involve creating a bitmap data object containing a bit for each one of a plurality data chunks in the data object.
  • Bits may be set in the bitmap corresponding to data chunks referred to in the subsequent access request for the clone. In such an instance, the bitmap is updated as data chunks are copied to the clone object.
  • a separate process for handling temporary clone objects uses the bitmaps to determine when to access the original object, the clone, or a snapshot.
  • FIG. 1 is a diagram showing interaction between a Primary Node, Intelligence Node, and Remote Intelligence Node, and connected storage pools.
  • FIG. 2 is a view of an appliance device containing Primary and Intelligence Nodes.
  • FIG. 3 is a diagram showing the components of a Primary Node.
  • FIG. 4 is a diagram showing the components of an Intelligence Node.
  • FIG. 5 is a diagram showing the analytics flow process.
  • FIG. 6A is a diagram showing the structure of a change catalog.
  • FIG. 6B shows a discovery point
  • FIG. 7 is a diagram illustrating a multi-threaded log.
  • FIG. 8 shows a process flow for handling an access request at the PART.
  • FIG. 9 shows a process flow for synchronizing requests to release entries in the multithreaded log.
  • FIG. 10 shows write access gathering at the PART.
  • FIG. 11 shows virtual file system objects representing virtual machine subdirectories created by a hypervisor application can be submitted to a snapshot process.
  • FIG. 12 shows a more general case where an action is applied in a restore node across subdirectory trees and file objects stored within those directories.
  • FIG. 13 is a high level diagram of a system that provides clone on demand with background migration of data and metadata.
  • FIG. 14 illustrates a file object and corresponding clone bitmap.
  • FIG. 15 is a process flow for creating a clone of a file object.
  • FIG. 16 is a process flow for access a cloned file object.
  • FIG. 17 shows a directory tree object and its corresponding clone.
  • FIG. 18 is a process flow for creating a directory object.
  • FIG. 19 is a process flow for accessing a temporary clone directory object.
  • Primary Storage networked storage accessible to multiple computers/workstations.
  • the storage can be accessed via any networked device, either as files or blocks.
  • “primary storage” refers to both blocks and files.
  • Intelligence Storage secondary storage containing gathered intelligence, discovery points, and a redundant real-time copy of files and block data contained in Primary Storage.
  • Primary Node includes access protocols to communicate with an Intelligence Node, Remote Sites, and Expansion Nodes; access protocols layer (for example, NFS, SMB, iSCSI); protection and analytics in real-time ("PART") layer; file and block storage layer (file system, block volume); and connection to storage devices (RAID, DISK, etc.).
  • a Primary Node appears to system users as Primary Storage, and provides an interface and controls to act as the access to Intelligence Storage.
  • Intelligence Node includes access protocols to communicate with a Primary Node, Remote Sites, and Expansion Nodes; data intelligence storage layer (intelligent data services & rules processing); file and block storage layer (file system, block volume); and connection to storage devices (RAID, long-term storage).
  • intelligence node data is accessed by users through a Primary Node, but in alternate embodiments Intelligence Nodes may be directly accessed by users.
  • a discovery point created from a mirrored (high availability) copy of primary data, contains data analytics for accessed and changed primary data since a prior discovery point.
  • a discovery point may contain the changed data, providing for a virtually full but physically sparse copy of the primary data captured at a user-specified point in time or dynamically based on change rate or other analytics.
  • analytics metadata stored in a discovery point can be expanded as deeper levels of user data analysis are performed and more analytics are gathered. Tracked primary data changes can be retained for the life of the discovery point or can be removed at scheduled or dynamic intervals, such as after deep data analysis is complete and desired analytics metadata is obtained. Removing primary data allows for more efficient space utilization, while retaining primary data enables point-in-time recovery of that version of data.
  • Change Catalog an ordered set of real-time access and change information related to a data object, tracked at a discovery point granularity.
  • a change catalog tracks who, how, when, and where aspects of a data object being accessed and/or modified. There is one change catalog for every discovery point.
  • Remote Site one or more off-site nodes in communication with local site primary or intelligence nodes.
  • Pool the collection of data storage connected to a node.
  • Object a file, directory, share, volume, region within a volume, or an embedded object.
  • Objects can be complex, containing other embedded objects.
  • a file can be a container containing other files, or a volume can have a file system on top of it which in turn contains files.
  • the system is capable of recognizing complex objects and tracking changes at finer embedded object granularity.
  • Selective Restore an automatic (policy based) or manual (customer initiated) restore at an object level.
  • Site Restore a manually initiated process to recreate primary or intelligence pool content using a previously protected version of the data being restored.
  • Container objects which may have other embedded objects, such as a file, directory, file system, or volume.
  • Expansion Nodes appliance having a processor, memory (RAM), network connectivity, and storage devices, and connected to one or more primary or intelligence nodes scaling the processing power and/or storage for connected nodes.
  • the disclosed high availability (HA) storage system provides primary storage, analytics, and live restore functions.
  • Live restore is a technique used to optimize data restoration. It can be used to recover user data in case of a failure or to recover previous versions of the user data.
  • the system provides primary storage access as block and/or file level storage while avoiding single points of failure.
  • the system collects analytics in realtime while also protecting data in real-time on separate physical media, and includes options for off-site data protection.
  • the system implements deep analytics enabling restore, storage, and data intelligence, and protects both customer data and associated analytics.
  • the system provides traditional file based and custom API methods for extracting analytics metadata.
  • the system employs Live Restore techniques at a file and at a block level to recover in case of a failure or to recover a previous version of user data.
  • a file or block level Live Restore uses previously gathered analytics to prioritize data to be restored, while allowing user I/O access to the data during restoration.
  • Primary Node 100 of the system connects within a network to provide block and/or file level storage access to connected computing devices (not shown), real-time data protection, and real-time analytics of primary data.
  • Primary data is read from and written to primary storage pool 110.
  • the data can be written or read as files or blocks depending on the access protocol being used.
  • As the data is written it is automatically mirrored and tracked for data protection as part of a HA process for the primary node.
  • the mirrored cache of the data is created for Intelligence Node 120.
  • the Intelligence Node enables data protection, analytics, and recovery.
  • the Intelligence Node stores a real-time copy of primary data, analytics and discovery points within intelligence pool 130. Discovery points are automatically or manually created at any point by the Intelligence Node, and based on fine grained change data enabling action to be taken immediately with no need to copy the underlying primary data or do any post processing to determine what has changed since any prior discovery point.
  • each Node is capable as acting as either a Primary Node, an Intelligence Node, or both. For reliability and performance reasons, separate Primary and Intelligence Nodes are desirable. In case of failure of either node, the other may take over operation of both. Implementation without dual-capability (that is, operating solely a Primary Node and solely an Intelligence Node) is possible but loss of service (to either primary or intelligence storage) would occur on failure of such a node.
  • each one of the Nodes has a processor and local memory for storing and executing Node software, a connection to physical storage media, and one or more network connections including at least a dedicated high bandwidth and low latency communication path to other Nodes.
  • the Primary Node and Intelligence Node are physically housed within a single device, creating a user impression of a single appliance.
  • FIG. 2 shows one such example, with Primary Node 100 and Intelligence Node 120 housed together to appear as a single physical appliance.
  • Implementation may be with any number of disks, for example such as a four rack units (4U) housing containing up to twenty-four hard drives, with separate physical storage devices connected to the system.
  • 4U rack units
  • each node is completely separated from the other with the exception of a backplane, with each node having a dedicated (not shared) power supply, processor, memory, network connection, operating media and optionally non-volatile memory. Separation enables continued operation, for example the Intelligence Node may continue operating should the Primary Node fail, and vice versa, but shared resource implementation is also possible.
  • a node actively operating as Primary Node 100 operates storage protocol server software 300, for example Common Internet File System (CIFS), Network File System (NFS), Server Message Block (SMB), or Internet Small Computer System Interface (iSCSI), so the Primary Node will appear as primary storage to network- connected computer devices.
  • the storage protocol server software also communicates with a protection and analytics in real-time process (PART) 310 which intercepts and takes action on every data access.
  • CIFS Common Internet File System
  • NFS Network File System
  • SMB Server Message Block
  • iSCSI Internet Small Computer System Interface
  • the PART 310 performs three main roles after intercepting any data access request: mirroring primary data for HA, gathering in-line data analytics on primary data, and storing primary data.
  • the examples explained herein are directed to a file access perspective, but the PART can similarly process block level accesses.
  • the PART can identify embedded objects and perform the same analysis that is applied to file-level accesses.
  • Intercepted access requests include read, modify (write data or alter attributes, such as renaming, moving, or changing permissions), create, and delete.
  • the PART tracks and mirrors the request (and data) to the Intelligence Node. Communication with the Intelligence Node is through synchronous or asynchronous inter-process communication (IPC) 340 depending on configuration.
  • IPC inter-process communication
  • IPC may including any suitable protocols or connections, such as Remote Procedure Call (RPC) or a Board-to-Board (B2B) high performance, low latency communication path that may be hardware specific.
  • RPC Remote Procedure Call
  • B2B Board-to-Board
  • Any data included with a data access request, such as included in write operations, is also mirrored to the Intelligence Node as part of HA system operation. This mirroring establishes data protection through real-time redundancy of primary storage.
  • the PART executes in-line analysis of primary data, gathering real-time analytics.
  • the PART sends gathered real-time analytics to the Intelligence Node, where the analytics are added to a change catalog maintained by the Intelligence Node.
  • the PART directs the request to an actual file system, for example Fourth Extended File System (EXT4) or Z File System (ZFS), or block volume for file or block storage access 330 to physical storage devices.
  • EXT4 Fourth Extended File System
  • ZFS Z File System
  • the storage access function 330 (be it file system level or block level) performs the access request on storage media, and returns the result to the PART for return to the requesting system.
  • the storage media includes disks attached to the system, but other storage media solutions are possible.
  • the Primary Node also includes the software necessary to operate as an Intelligence Node in case of Intelligence Node failure.
  • the Primary Node also operates management software.
  • the management software provides system administrators access to configure and manage system users and access discovery points for the restore process.
  • a node actively operating as Intelligence Node 120 operates Inter Process Communication (IPC) communication software 400 capable of communicating with the Primary Node.
  • the communication software includes an API to receive real time analytics (change catalog entries) from the Primary Node, data change and access requests (read, modify, create, delete) from the Primary Node, data protection and intelligence control commands, and data restore commands.
  • Data protection and intelligence control commands include commands for creating discovery points, setting up management rules for managing discovery points (including deletion), and searching and restoring content that has been backed up.
  • Data restore commands include commands for accessing previously backed up data.
  • Data change requests that are received at the Intelligence Node are applied to that node's copy of current data, thereby maintaining a real-time mirror of primary storage. This implements real-time data protection for the current data.
  • the Intelligence Node maintains a change catalog 600 containing real-time analytics gathered from accessed and changed data since the last discovery point 650.
  • a discovery point is also created by associating and storing a change catalog together with reference to the mirrored copy of changed primary data since the last discovery point as maintained in the intelligence pool.
  • the Intelligence Node implements file or block- level access 430 to its own pool 130 of physical storage. This intelligence storage pool retains the real-time copy of primary data and discovery points.
  • the stored intelligence data within discovery points includes in-line analytics (change catalog) as received from the Primary Node and additional analytics 410 executed by the Intelligence Node.
  • the real-time copy of primary data also enables distributed response processing between the Primary and Intelligence Nodes. For example, load balancing between the Primary and Intelligence Nodes may enable greater scalability. As both have real-time copies of primary data, read requests may be balanced between the nodes, or alternatively directed to both nodes with the fastest-to-respond used for the response.
  • the Primary Node may act as a controller for such distributed processing, or a separate controller may be used.
  • Primary 110 and Intelligence Data 130 reside on the same appliance, they can be distributed to multiple discrete appliances deploying all the same techniques with the exception that the communication method is performed over a network transport instead of using the HA mechanisms within an array.
  • Intelligence is at the core of the system. There are four types of intelligence functions in the system: Data, Operational, Storage, and Recovery. All four use the same processing engine and common analytics metadata to provide analysis both at fixed points and as gathered over time.
  • Data Intelligence 452 allows for intelligent user content management.
  • Operational Intelligence 456 analyzes the behavior of the system and application logs stored on the system to provide insight into applications and security of the system.
  • Storage Intelligence 454 allows for intelligent storage system resource management, including automatic storage allocation and reallocation including dynamically growing and shrinking storage pools.
  • Recovery Intelligence 450 allows for intelligent data protection and data restore. All types of intelligence may be used for, or enable operation in conjunction with, different types of analytics, such as, but not limited to, collaboration, trending, e-discovery, audits, scoring, and similarity.
  • Analytics begin at the Primary Node, which tracks data access and data modifications, system behavior, change rates, and other real-time analytics. It provides this real-time analytics information to the Intelligence Node. Intelligence gathering determines time and owner relationships with the data for collaboration and contextual information about the data. The gathered intelligence is used for later search and reporting, and is tracked in change catalogs associated with the data.
  • change catalogs 600 are created as part of inline real-time analytics 500 performed by the Primary Node 100, but change catalogs 600 are then also further expanded by the Intelligence Node 120 performing further data processing, and create the foundation for later search.
  • the change catalog data is initially created in realtime at the Primary Node (such as via PART 310) and includes extended information about the specific data access, for example, allowing complete tracking of who/how/when/where accessed, created, modified, or deleted a file or other data object.
  • Traditional file metadata includes only an owner, group, path, access rights, file size, and last modified timestamp. This provides some, but not complete, information about a file.
  • the PART operated by the Primary Node, intercepts every file access event.
  • the Primary Node has the ability to track extended metadata about a file - including identification of every modification and every access, even those which do not modify the file, by timestamp, user, and type of access.
  • this extended metadata is stored as a change catalog entry 610 that identifies the object being accessed, the actor (user performing an operation), and the operation being performed. Additional information which may be in a change catalog entry includes, but is not limited to, object name, owner, access control lists, and time of operation.
  • the change catalog 600 contains this extended metadata information, and serves as the foundation of further analytics, such as performed later by the Intelligence Node.
  • the change catalog entry may also include security information, such as permission rights for access, associated with the object.
  • An administrator may configure the degree of tracking, or even enable/disable tracking on a file location, user, group-specific, or other basis, and the Primary Node is capable of incorporating all details of every file access into the change catalog entries.
  • These change catalog entries of enhanced metadata are gathered by the Primary Node and transmitted to the Intelligence Node for storing and expanding with further analytics.
  • the change catalog metadata tracks incremental changes which are also linked to a discovery point 650. Every time a new discovery point is created the current change catalog is closed off and stored within the discovery point.
  • the system may be configured to retain a copy of the discovery point analytics metadata at the Intelligence Node even if that discovery point is migrated off the Intelligence Node, enabling more efficient query processing.
  • a discovery point 650 is created by associating and storing a change catalog together with the mirrored copy of changed primary data since the last discovery point in the intelligence pool. After a discovery point creation, a new change catalog 600 is created allowing gathering of new real-time analytics on primary data. Change catalogs and discovery points are preferably maintained per volume or file system in primary storage, but may also span multiple volumes or file systems. Discovery points allow deeper analytics on a point in time version of primary data, and can also be used to recover a prior version of primary data.
  • a discovery point contains data analytics for accessed and changed data since a prior discovery point. When created, a discovery point also contains a virtually full but physically sparse copy of primary data at the time of creation of that discovery point.
  • the system uses data visible within discovery points to perform deeper data processing, creating more analytics metadata.
  • the analysis is done on accessed and changed data since a previous discovery point, using the real-time analytics reflected in the change catalog. These newly gathered deeper analytics are also stored within the discovery point.
  • Primary data may be retained for the life of the discovery point, or may be removed earlier, such as after the deep data analysis is complete and desired analytics metadata obtained. Removing the primary data allows for more efficient space utilization, while retaining the primary data enables recovery of primary data at the point in time of the creation of the discovery point. From one discovery point until the creation of a next discovery point, file changes, deletions, renames, creations and such are tracked as cumulative modifications to from the prior discovery point, so that only incremental changes are maintained.
  • Discovery points can be deleted manually through a delete discovery point command, or automatically based on time or analysis in order to save storage space or for off-site migration. Deletion of discovery points is complicated by management of analytics metadata.
  • the analytics metadata stored within a discovery point contains information about data changed within a period of time. If the stored analytics are deleted they can be lost. To prevent this, the time period for analytics associated with one or more other discovery points can be adjusted, and relevant portions of analytics metadata from a discovery point being deleted extracted and merged with other analytics already stored within the other discovery points.
  • an adaptive parallel processing engine operates on the change catalog 600 to derive these more complex analytics, including tracking changes and use over time.
  • the Rule Engine applies rules 510 to analyze content on the underlying primary data, enabling deeper analytics on stored data.
  • a second level dictionary can provide sentiment attributes to an already indexed document. Regular expression processing may be applied to see if a document contains information such as social security or credit card numbers.
  • Each rule may have a filter 530 to match content, and an action 540 to take based on results. Rules can be nested, and used to answer user-specific questions.
  • Rules are configurable by administrators or system users, allowing dynamic rule creation and combination based on different applicable policies 520. Rules can be combined in multiple ways to discover more complex information. Rules may also be configured for actions based on results. For example, notifications may be set to trigger based on detected access or content, and different retention policies may be applied based on content or access patterns or other tracked metadata. Other actions may include, but are not limited to, data retention, quarantine, data extraction, deletion, and data distribution. Results of applied rules may be indexed or tracked for future analysis.
  • rules 510 identify results, such results may be indexed or tracked for other analytical use. This additional metadata may be added to the change catalogs for the relevant files or objects. The metadata may also be tracked as custom tags added to objects. Tags may be stored as extended attributes of files, or metadata tracked in a separate analytics index such as data in a directory or volume hidden from normal end user view, or in other data stores for analytics. Rules, and therefore analytics, may be applied both to data tracked and to the metadata generated by analytics. This enables analytics of both content and gathered intelligence, allowing point-in-time and over-time analysis. The rules results and actions may serve as feedback from one or more rules to one or more other rules (or even self-feedback to the same rule), enabling multi-stage analysis and workflow processing.
  • Recovery Intelligence is the set of analytics implemented by Intelligence Node 120 around data protection. The purpose is to protect data and associated analytics. When data reaches the Intelligence Node a mirrored copy is stored in the intelligence pool, creating redundancy with primary storage, and these changes are tracked for use in discovery point creation. Primary data, discovery points, and intelligence data are preferably separated on actual physical media at the spindle or disk pool level, such that a failure of a single individual physical device is always recoverable. As discovery points are created based on change catalogs tracked at the Intelligence Node, they can be created at any time without any impact on the performance of primary storage. This eliminates a need to schedule time- windows for discovery point creation.
  • Each discovery point includes incremental changes from the prior discovery point, including data object changes and the analytics gathered and associated with the data during such changes.
  • Intelligent rules can be applied to automate discovery point creation, such that, in addition to manual or time -based creation, discovery point creation may be triggered by content changes.
  • Such changes may be percentage based, specific to percentage change of certain identifiable subsets of the entire data pool, based on detected deviations from usage patterns such as increase in frequency of specific accesses, or based on real-time analysis of data content.
  • the change catalog accumulating real-time changes is closed.
  • the change catalog is then stored within the created discovery point, and a new change catalog created for changes to be associated with a next created discovery point.
  • the analytics and data stored within discovery points enable efficient restores, allowing search over multiple discovery points for specific object changes without requiring restoration of the data objects from each discovery point.
  • Such search can be based on any analytics performed, such as data tracked in the extended metadata and content-based analysis performed by application of the Rule Engine.
  • the tracking further enables indexing and partial restores - for example specific objects, or embedded objects within complex objects, can be restored from a discovery point without a complete restore of all data from that discovery point.
  • Data Intelligence is a set of analytics at the Intelligence Node analyzing content.
  • Data Intelligence operates through the Rule Engine, and can be applied to unstructured data, for example file metadata such as document properties of Microsoft Office documents or the actual content of such documents, semi-structured data such as log files or specific applications such as Mail programs, structured data such as databases or other formats for which schema may be known or discovered by the system, and recursive containers such as virtual machines, file systems on file systems, file systems on volumes, or archives.
  • file metadata such as document properties of Microsoft Office documents or the actual content of such documents
  • semi-structured data such as log files or specific applications such as Mail programs
  • structured data such as databases or other formats for which schema may be known or discovered by the system
  • recursive containers such as virtual machines, file systems on file systems, file systems on volumes, or archives.
  • File systems use internal data structures, called metadata, to manage files, directories and data in files.
  • a typical file system uses logging to guarantee crash consistency.
  • One of the popular techniques to guarantee crash consistency is a write-ahead log.
  • the file system logs the intent of modifications to the log, and then performs the metadata modifications on disk. In case of a panic, power failure, or crash, the log is then replayed to bring the file system back to a consistent state.
  • the PART 310 intercepts data access requests, forwards them to a primary node, mirrors them to a high availability restore node, and performs analytics to create intelligence data.
  • each of the primary node 100 and restore node 140 operate with their own independent file system 102, 142 (FS).
  • FS file system 102, 142
  • the file systems 102, 142 may be a ZFS -compatible file system or some other file system.
  • FS file systems
  • a logged transaction includes all the metadata modifications that will be done as part of an I/O. For example, if an I/O operation allocates on indirect block, the log entry in 101 or 141 consists of the new allocated indirect block, the parent indirect block where the new block will be inserted, an offset in the parent indirect block, the inode where associated with the indirect block, and so on.
  • the PART 310 maintains its own log 311 independent of the logs 101, 141, if any, as maintained by the file systems 102, 142 in primary 100 and restore 140 nodes.
  • This PART-level, "virtual file system" log 311 is implemented in a durable storage medium that can be written to in random order, such as nonvolatile memory. To achieve crash consistency, access requests can be replayed at the primary 100 and restore 140 nodes consistent with the original order in which they were received at the PART 310.
  • the PART log 310 may obviate the need for logs 101, 141, which then may be disabled or bypassed if the file systems 102, 142 allow this.
  • any metadata in the PART log 311 is stored with a corresponding transaction ID.
  • the transaction IDs are a unique number maintained by the PART 310 and incremented upon each access request received. Writes to the PART log 311 may therefore be multithreaded such that they can be written any time and in any order, with the order information retained in the transaction ID associated with each request.
  • Fig. 7 shows a typical PART log entry including a transaction ID, an operation type, a file handle, offset, length and data.
  • the entries in the PART log 311 are arranged in a number of chunks 301 typically with each chunk being of equal size to other chunks.
  • access requests received by the PART 310 may be multithreaded.
  • the various chunks 301 in the PART log 311 enable log entries to be written in any order and also concurrently.
  • writes to the random access, high speed PART log 310 do not have to observe any ordering dependencies, yet the ordering can be regenerated when the PART log 310 is replayed to the primary and restore nodes.
  • the PART 310 is executing five (5) concurrent threads labeled Al, A2, A3 and B.
  • some of the threads are issuing access requests for a data tree structure that is to be populated with engineering data concerning the configuration of a manufactured component.
  • Other threads executing in the PART 310 are concerned with processing customer orders for the component.
  • a first thread Al may be responsible for creating the tree while threads A2 and A3 are responsible for writing data to the tree.
  • thread B is handling an entirely different operation such as supporting database accesses concerning the customer orders for the component.
  • each access request is written to the PART log 311 , it is forwarded in parallel to each of the primary 100 and restore 140 nodes.
  • the primary 100 and restore 140 nodes then copy the request data to a respective local cache 104, 144, but do not yet actually issue the request to their underlying file systems 102, 142 to access permanent storage such as respective disks 103, 143. Without the PART log 311, such behavior by primary 100 and restore 140 nodes would not provide crash consistency locally within file systems 102, 142, and also does not provide consistency between file systems 102, 142.
  • the PART 310 issues a synchronization ("sync") request to the primary 100 and restore 140.
  • sync synchronization
  • the primary 100 and restore 140 flushes their respective cached data to disks 103, 143.
  • the primary 100 and restore 140 then acknowledge the sync back to the PART 310.
  • the PART 310 can now free the corresponding chunks 301 in PART log 311. In other words, it is not until the sync command is complete that data related to the requests is known to be correctly persisted to respective disks in the primary and restore nodes.
  • Fig. 8 shows a typical process flow among the PART 310 and primary node 100. It should be understood that the corresponding operation between the PART 310 and the restore node 140 is similar.
  • the PART 310 receives an access request from a host.
  • the PART assigns a next available transaction ID to the request.
  • the access request is then written to any available chunk in the PART log 311.
  • the request is then sent to both the primary 100 and restore 140 nodes.
  • the primary 100 receives the request from the PART 310.
  • a next step 842 if the primary 100 and restore nodes maintain a local log 101, 141, they determine a place for the transaction in their respective log order from the transaction ID (which is known to have been assigned in the same order in which multithreaded PART receive it).
  • the primary 100 and restore nodes maintain a local log 101, 141, they determine a place for the transaction in their respective log order from the transaction ID (which is known to have been assigned in the same order in which multithreaded PART receive it).
  • data associated with the request is stored in the primary's local cache memory 104,144.
  • the primary can send an access complete acknowledgment back to the PART 310.
  • the PART 310 can then report that fact that the access is logically complete even though the data has not yet been flushed to disk at the primary 100.
  • Fig. 9 illustrates the process flow between the PART 310 and the primary 100 and restore 140 when the multithreaded log 311 is full or nearly full.
  • state 910 the PART 311 log is recognized as no longer being able (or soon to become unable) to store additional requests.
  • step 911 a sync command is sent from the PART 310 to both the primary 100 and restore nodes 140.
  • state 920 the primary 100 (or restore node 140) receive the sync command and in state 922 they flush their local cache to permanent file system (FS) storage such as one or more disk(s). Once the flush operation is complete in state 923, an FS file system (FS) storage such as one or more disk(s).
  • FS permanent file system
  • the PART receives the acknowledgment from the primary 100, and at some point (either prior to, at the same time, or subsequent to state 930) the PART 310 also receives an acknowledgment from the restore node 140.
  • state 933 having received flush acknowledgments from both the primary 100 and the restore 140, the PART 310 can finally release the associated chunks 301 in PART log 311.
  • the durable storage used for the PART log 311 is a fast access storage device, such as a solid state device, so that the log file can be sorted in transaction ID order as quickly as possible when it needs to be read back, such as when a fault occurs before data is flushed to disk by both the primary 100 and restore nodes 140.
  • a storage system It is typical for a storage system to aggregate write operations in a cache before being flushed to main storage.
  • a data intelligent storage system is implemented with a primary node 100 and high availability / intelligence data stored at restore node 140.
  • a write access request may come into the PART 310, and recorded in a PART log 311 before being forwarded to primary 100 file system and restore 140 file system.
  • the primary and restore file systems may maintain their own logs 101, 141 as previously described.
  • those file system level logs 101, 141 are also copied to a remote disk such that it another remote copy 151 is made of the primary log 101 and another remote copy 105 is made of the restore log 141.
  • each single I/O transaction may result in many different write operations to different primary data stores and logs.
  • a PART level cache which we refer to as a write gathering cache 333, is implemented to store data associated with write requests.
  • write gathering cache 333 is implemented to store data associated with write requests.
  • the associated data is immediately copied to the write gathering cache 333, and the I/O request is also acknowledged.
  • Certain other operations that involve metadata such as a make directory (mkdir) operation, are first logged in the PART log 310 and then issued to the primary 100 and restore 140.
  • FIG. 11 shows the data intelligence storage system being accessed by one or more applications 1010.
  • the application 1010 is a hypervisor environment such as an ESX or ESXi server (ESX and ESXi are trademarks of VMware, Inc. of Palo Alto, California).
  • the application 1010 creates and maintains various virtual machine (VM) files in such an environment on a subdirectory basis.
  • VM0 virtual machine
  • /vmO first subdirectory
  • FIG. 11 shows the data intelligence storage system being accessed by one or more applications 1010.
  • VM0 virtual machine
  • /vmO first subdirectory
  • the files associated with a second virtual machine are to be disposed within second directory (/vml), and the files associated with an (n-l)'th virtual machine in subdirectory (/vmn).
  • the files (kO.vmdk) associated with yet another virtual machine are to be stored in a directory (/vml 1) that is subordinate to directory /vmO.
  • the ESX server application 1010 may therefore be hosting a number of virtual machines; the data associated with each virtual machine including its operating system image files, application files and associated data are stored in one or more files arranged in a directory tree 1011 within a single file system 1015 tree.
  • PART 310 issues access requests to the PART 310.
  • PART 310 not only sends the access request to one or more file systems on primary node 100, but also sends the access request to the file system(s) on restore node 140.
  • intelligence 145 in the restore node 140 it also becomes desirable to use intelligence 145 in the restore node 140 to perform certain tasks.
  • One such task creates intelligence data in the form of a change catalog entry with associated discovery points (1020-1, 1020-2).
  • these discovery points 1020 may include snapshots of the state of the virtual machine files and their associated data, metadata, other intelligence data, and change catalog.
  • snapshots become discovery points includes one or more snapshots of each VM.
  • Such existing snapshot technologies are directed to instead storing a snapshot of an entire file system. However it may be desirable in certain circumstances to enable the use of such snapshot technologies on a single VM.
  • PART 310 To identify particular applications such as ESX server 1010 that create subdirectories, such as those containing virtual machine files, and manage them in a distinct way. The PART 310 therefore can more efficiently enable certain actions by intelligence 145. As shown in Fig. 11, as it handles access requests, the PART 310 maintains an entire set of filesystems 1050 for each sub-directory on the primary 100 and an entire set of filesystems 1070 on the restore 140. What appears to the user application (ESX server 1010) to be an ordinary filesystem containing ordinary subdirectories is actually a virtual filesystem 1040 wherein any given subdirectory may actually be a link to a separate, associated file system that actually contains the .vmdk files for a given VM.
  • ESX server 1010 What appears to the user application (ESX server 1010) to be an ordinary filesystem containing ordinary subdirectories is actually a virtual filesystem 1040 wherein any given subdirectory may actually be a link to a separate, associated file system that actually contains the .vmd
  • the PART 310 When these subdirectories are accessed in the virtual file system 1015, the PART 310 thus transparently redirects those accesses to the associated file system(s) 1050, 1070 on the primary and restore.
  • a make directory (mkdir) command to create VM subdirectory /vml is intercepted by the PART 310, which then creates file system v.vml (1050-1) on the primary 100 and its mirror v.vml (1070-1) on the restore node 140.
  • the PART 310 then creates the new file system directory /vml in the primary filesystem 1040, which is a virtual "mount point" linking the subdirectory /vml in virtual file system 1015 with its associated actual file system v.vml (1050-1, 1070-1). This link is denoted by pointer 1042.
  • a write access directed to file /vmO/al.vmdk is intercepted by the PART 310, which, following link 1041, redirects that write access to the filesystem v.vmO (1050-0) on primary 100 which actually contain the file al.vmdk.
  • the PART 310 also mirrors write accesses to the restore node 140; in this case, the mirrored write access is directed to the filesystem v.vmO (1070-0) on the restore node 140 which actually contains the mirror of al.vmdk.
  • the PART 310 maintains the illusion of a subdirectory tree 1011 but actually creates a number of file systems 1050-0, 1050-1, 1050-2, 1050-11, 1050-n on primary 100 and a number of file systems 1070-0, 1070-1, 1070-2., 1070-11, 1070-n on restore 140.
  • the snapshot processes running as part of intelligence 145 can be executed using the standard file system oriented snapshot process but using the virtual mount point information to locate the underlying filesystems 1070 associated with a given subdirectory.
  • the virtual filesystem (VFS) 1090 hides the existence of multiple independent, "container file systems" from user application 1010.
  • Subdirectories in the virtual file system (VFS) 1090 are accessible as subdirectories, but at the same time the underlying container file systems 1070 are accessible to the snapshot processes.
  • Associated file system snapshot technology in the restore node 140 can now be relied upon to obtain snapshots of a given VM independently of snapshots of other VMs, and the restore node 140 can thus treat them as it treats any file system.
  • the PART 310 layer Upon receiving a mkdir command, the PART 310 layer then issues requests to primary 100 and restore 140 to create the file system constructs v.vmO (1050-0, 1070-0), v.vml (1050-1, 1070-1), v.vml l (1050-k, 1070-k), v.vmn (1050-n, 1070-n) and associates them with links (1041, 1042, 1043) to the actual virtual machine files and subdirectories as expected by the ESX server 1010.
  • the corresponding v.vm structure(s) can be removed from or edited within the VFS 1090 data structure.
  • This shadow virtual file system 1090 is in effect, created and maintained via processes internal to the PART 310.
  • the virtual mount points in VFS 1090 are thus accessed by the snapshot processes 1020-1, 1020-2, but the subdirectory structures remain in place for other I/O requests as received from the ESX 1010.
  • Fig. 12 illustrates a more general case where restore node 140 applies a storage related action 1120 (such as snapshot, de-duplicate, compression, or some other storage related action) to portions of a file system (such as a subdirectory 1016) rather than an entire file system 1015 created by application 1010.
  • a virtual file system (VFS) layer 1090 in the PART 310 again used to provide links between file systems and the subdirectories 1011, 1012 associated with these file systems.
  • VFS virtual file system
  • the application 1010 may wish to apply a de-duplicate action only to a particular sub-sub-directory 1016.
  • a "no de- duplicate" property can be maintained in the VFS 1090 by PART 310 for that subdirectory, because it is a virtual mount point linking to an actual underlying file system 1070-k upon which the property can be applied.
  • the PART 310 may apply properties to the virtual machine subdirectories in consistency groups. So, for example, the VFS 1090 maintained by PART 310 may further indicate that the virtual subdirectories (1070-0, 1070-1) for two of the VMs (such as /vmO and /vml) are to always be treated together and subjected to the same snapshot policies.
  • the VFS 1090 may be exposed to analytics running in the restore node 140. The results of those analytics can then be used to determine how to further refine the directory structure(s) 1050 and/or 1070.
  • the intelligence process 145 in the restore node may detect the application 1010 accesses files in a particular way which indicates some need to apply a certain scope to a given action.
  • the intelligence 145 may determine that a certain virtual machine contains a sub-sub-directory /vm5 that is type of database (a.005.db) having contents that are known to change often and thus will not benefit from compression. Thus intelligence can maintain a compression property when accessing the virtual directory structures in VFS 1090 to exclude sub-subdirectory /vm5 from any compression action.
  • a user may wish to create a clone of a prior snapshot that contains prior obtained intelligence data, to perform still further analytics.
  • the clone may be needed to recover a failed file system.
  • the user may simply wish to copy an object from one file system to another without destroying the original object.
  • a primary node 100 stores primary data and restore node 140 stores associated intelligence data and other information.
  • the object 1215 to be cloned may reside within snapshot ("snap") 1210; this object 1215 may be a file, a directory or even the entire file system within the snapshot 1210.
  • a clone request 1205 is made to the PART 310, it does not simply forward the request to the file systems 102, 142 on primary 110 and restore 140 nodes. Instead, a new file system clone object 1220 is immediately thin provisioned on the primary node 100 and exposed by the PART 310 to the user (host), and a new file system clone object 1230 is also thin provisioned on the restore node 140 (which is consistent with the system's usual process of mirroring requests to the restore node 140, although that step is not critical to handling the clone request as described herein). The PART 310 therefore does not have to first actually populate all of the metadata and data to the new file system clone objects 1220, 1230.
  • the PART level 310 thus coordinates execution of a clone process rather than pushing the clone process down to an FS 102, 142 or other file system layer in the primary 100 and restore 140 nodes.
  • the file object 1300 to be cloned consists of a metadata portion 1310 and a data portion 1320.
  • the particular structure of metadata 1310 depends on the type of underlying file system, and will be different for a Linux file system, than say, a Windows or MAC OSX file system as is known in the art.
  • the data portion 1320 can be considered to be a collection of data bytes of a certain size, such as chunks 1308 each of 512 kilobytes (kB).
  • a clone file process executed in PART 310 maintains a clone bitmap 1350 for each such cloned file object 1300.
  • the clone bitmap 1350 includes at least a single bit 1355 for each chunk 1308 in the file object 1300.
  • the clone bitmap 1350 is used as a tool to coordinate the status of actually copying portions of the data 1320 from the original file to the cloned object.
  • Fig. 15 is an example flow for a clone process 1400 where the cloned object is a single file.
  • a request is received at the PART 310 to create the cloned object.
  • a new clone object 1220 is created on the primary node 100 and a new clone object 1230 is created on the restore node 140, but only thin provisioned, without actually copying any data yet.
  • the thin provisioned file objects at this point may contain some metadata in state 1406 depending upon whether or not the underlying file system maintains metadata within the file object itself (certain operating systems such as Windows and MAC OSX do this; other operating systems such as Linux maintain file metadata as part of a separate inode).
  • the bitmap 1355 is created for the file with all bits therein set to logical "false" value indicating that the corresponding data for chunk has not yet been populated to the clones.
  • state 1410 the new clone file objects 1220, 1230 are made accessible for subsequent input/output (I/O) requests by the user I/O even though no data portion 1320 has yet to be copied from the source snap 1210.
  • I/O input/output
  • a background live restore thread 1420 is started.
  • the live restore thread 1420 may typically be a background thread executing only when the PART 310 is otherwise not executing other tasks. In other instances, the live restore thread 1420 may be a thread with low priority (keeping in mind that the PART 310 is a multi-threaded processor as described above).
  • the purpose of the live restore thread 1420 is to perform the task of copying data from the source snap 1210 to the clones 1220, 1230.
  • a next chunk is located.
  • the next chunk is copied from the source snap 1210 to the clones 1220, 1230.
  • the bit in the bitmap associated with that chunk is then set to a logical "true" value.
  • the live restore process then continues as a background / low priority process as long as and until all chunks of the file have been copied to the clone.
  • Fig. 16 shows a typical process 1500 performed when the PART 310 receives an access request for a previously thin-provisioned clone object.
  • state 1502 the clone access request is received.
  • the clone bitmap 1350 associated with the clone object is consulted. If the bit or bits associated with the chunks accessed in the request are all set to "true", then that is an indication that the data has already been copied to clones 1220, 1230. Processing may proceed to step 1505 where the access request can be handled as per state 1510.
  • the access request refers to one or more chunks 1308 for which the bitmap 1350 indicate a "false” value and thus have not been previously processed
  • the bitmap is updated to set those bits to "true”.
  • data and possible metadata affecting chunks within the scope of the request are then populated to clones 1220, 1230.
  • the access request is then further handled.
  • the access request to the clone may be issued to both the primary 100 and restore 140 nodes.
  • the access request may also typically be issued to the both the primary 100 and restore 140 nodes by the PART 310 using the multithreaded log process described above. This then results in duplicate copies of the clone 1220, 1230 being eventually instantiated on the primary 100 and restore nodes 140 once data is flushed from the caches.
  • a different process is executed when the object 1210 to be cloned is an object that includes more than one file, such as a directory or even an entire file system.
  • Such an object as shown in Fig. 17, may be represented as a tree or graph structure consisting of nodes and various levels with edges connecting the nodes.
  • the node consist of values, such as metadata defining the content of an associated directory together with a list of references to child nodes and parent nodes that contain metadata for sub-directories and parent directories.
  • the snap to be cloned is a root directory with several subdirectories and files stored within those subdirectories.
  • the data structure 1600 thus consists of a root node 1610, and a first level 1620 consisting of four nodes representing four subdirectories 1621, 1622, 1623 and 1624. Each subdirectory contains pointers or other metadata concerning the files contained within, as is known in the art.
  • the structure also includes the subdirectory at second level 1630 with further subdirectories 1631 and 1632 and files. Still further subdirectories are located at third level 1630 with respective files.
  • the initial task when asked to clone such an object is to create a copy of the directory tree structure in the clone 1650 in a particular way. That process 1700 is shown in more detail in Fig. 18 and begins when a "clone directory" request is initially received in state 1702. In a step 1704 the PART 310 thin provisions the clone directory such as by only creating a copy 1660 of the root node 1610. In the next step 1706 metadata associated with the root node 1610 would also be copied as may be required by the particular type of file system (in the case of a Linux-compatible file system, that may include copying the inode for the directory). In state 1708 the clone object is then made available for user I/O.
  • a background and/or low priority live restore 1720 thread is kicked off for the directory object.
  • processing may continue with the clone appearing to be available for access by the user but without any data and without even the entire tree structure having actually been propagated to the clone yet.
  • the live restore process 1720 for a directory object begins in state 1722.
  • the directory tree for the original snap 1650 is walked in a depth first search.
  • the Depth First Search (DFS) from node 1610 would first located node 1621 for processing, then node 1631 and then node 1641 before returning to level two and node 1632 and so forth.
  • step 1723 the node that has been located in a depth first search is then added to the clone tree.
  • state 1724 another background thread is also started concurrently with thread 1720.
  • the background live restore process 1720 continues to determine if the depth first search locates any additional nodes, and if so, processing loops back to step 1723 to process the new node. If no new nodes are found in step 1725, then in state 1726 a background data restore thread (such as that described in connection with thread 1420 in Fig. 14) can then be triggered to restore data for the files referenced in the now cloned directory tree.
  • BFS Breadth First Search
  • the background live restore thread need not be executed for the temporary clone; similarly, the temporary clone need not necessarily recreate all of the data, metadata and directory/subdirectory trees for which the user only requests read access. Thus it is only when a user wishes to perform a read-modify-write, the corresponding chunk(s) need to be fetched from the original snap, modified, and then only those chunk(s) written to the clone.
  • Fig. 19 illustrates a typical temporary clone access process 1800 in more detail. It is understood that before this process 1800 is executed, a temporary clone structure has been created such as per the process 1500 in Fig. 16.
  • a temporary clone structure has been created such as per the process 1500 in Fig. 16.
  • an access request to the temporary clone is received.
  • a determination is made as to whether or not the access request is a read or a write.
  • the access request is a read, and if the corresponding bits in the bitmap are set to a logic false (indicating that there has been no prior write access to those chunks of the temporary clone), then the access request can be serviced in state 1806 from the original snap data structure 1210.
  • the access request is a read, and if the corresponding bits in the bitmap are set to a logic true (bits set) (indicating that there has been a prior write access to those chunks, then the access request can be services in state 1807 from the clone structure 1230.
  • process 1800 proceeds to state 1808 with bits now being set in the bitmap.
  • state 1810 data (and metadata if needed) are populated within the scope of the request to the clone 1230.
  • state 1812 the PART 310 finishes the write request. As before, this may be performed via the multithreaded log process in the PART 310.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Debugging And Monitoring (AREA)

Abstract

Dans cette invention, un système unifié assure une protection de données basée sur une mémoire primaire et une analyse intégrée. Les renseignements et analyses supplémentaires recueillis sur les données protégées et les analyses antérieures sont mémorisés dans des points de découverte. Le système ci-décrit met en œuvre : des écritures de journaux multifilières dans des nœuds primaires et des nœuds de rétablissement, avec collecte d'écritures dans des systèmes de fichiers ; des répertoires imbriqués qui peuvent servir à mémoriser des fichiers de machines virtuelles, chaque sous-répertoire ayant un système de fichiers associé pour les instantanés ; et le clonage d'objets sur demande avec des métadonnées d'arrière-plan et une migration de données.
PCT/US2015/045659 2014-08-18 2015-08-18 Journal de transactions multifilière pour le domaine primaire et le rétablissement/le renseignement WO2016028757A2 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201462038498P 2014-08-18 2014-08-18
US62/038,498 2014-08-18

Publications (2)

Publication Number Publication Date
WO2016028757A2 true WO2016028757A2 (fr) 2016-02-25
WO2016028757A3 WO2016028757A3 (fr) 2016-05-19

Family

ID=54072957

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2015/045659 WO2016028757A2 (fr) 2014-08-18 2015-08-18 Journal de transactions multifilière pour le domaine primaire et le rétablissement/le renseignement

Country Status (1)

Country Link
WO (1) WO2016028757A2 (fr)

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8046548B1 (en) * 2007-01-30 2011-10-25 American Megatrends, Inc. Maintaining data consistency in mirrored cluster storage systems using bitmap write-intent logging
EP2724279A4 (fr) * 2011-06-27 2015-07-01 Nokia Corp Système, procédé et appareil pour faciliter une sécurité de ressource
US8713170B2 (en) * 2011-11-18 2014-04-29 The Boeing Company Server-side web analytics system and method

Also Published As

Publication number Publication date
WO2016028757A3 (fr) 2016-05-19

Similar Documents

Publication Publication Date Title
US9785518B2 (en) Multi-threaded transaction log for primary and restore/intelligence
US10089192B2 (en) Live restore for a data intelligent storage system
US9213706B2 (en) Live restore for a data intelligent storage system
US10102079B2 (en) Triggering discovery points based on change
US10061658B2 (en) System and method of data intelligent storage
US9442952B2 (en) Metadata structures and related locking techniques to improve performance and scalability in a cluster file system
US7596713B2 (en) Fast backup storage and fast recovery of data (FBSRD)
US9501546B2 (en) System and method for quick-linking user interface jobs across services based on system implementation information
JP5918244B2 (ja) フォールトトレラントデータベース管理システムにおいてクエリ結果を統合するシステム及び方法
US20210064486A1 (en) Access arbitration to a shared cache storage area in a data storage management system for live browse, file indexing, backup and/or restore operations
JP2013545162A5 (fr)
KR20140051107A (ko) 데이터 관리 가상화를 위한 시스템들 및 방법들
US20050278382A1 (en) Method and apparatus for recovery of a current read-write unit of a file system
US20230259529A1 (en) Timestamp consistency for synchronous replication
US10289495B1 (en) Method and system for performing an item level restore from a backup
WO2016028757A2 (fr) Journal de transactions multifilière pour le domaine primaire et le rétablissement/le renseignement
WO2017112737A1 (fr) Déclenchement de points de découverte basé sur le changement
US20240111716A1 (en) Data analytics systems for file systems including examples of path generation
Wang et al. Fast off-site backup and recovery system for HDFS
Kamath et al. From Idea to Reality: Google File System

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15762812

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15762812

Country of ref document: EP

Kind code of ref document: A2

122 Ep: pct application non-entry in european phase

Ref document number: 15762812

Country of ref document: EP

Kind code of ref document: A2

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 12.12.2017)

122 Ep: pct application non-entry in european phase

Ref document number: 15762812

Country of ref document: EP

Kind code of ref document: A2