WO2015088918A1 - System and method for supporting persistent store versioning and integrity in a distributed data grid - Google Patents

System and method for supporting persistent store versioning and integrity in a distributed data grid Download PDF

Info

Publication number
WO2015088918A1
WO2015088918A1 PCT/US2014/068904 US2014068904W WO2015088918A1 WO 2015088918 A1 WO2015088918 A1 WO 2015088918A1 US 2014068904 W US2014068904 W US 2014068904W WO 2015088918 A1 WO2015088918 A1 WO 2015088918A1
Authority
WO
WIPO (PCT)
Prior art keywords
partition
data grid
distributed data
identifier
persistent storage
Prior art date
Application number
PCT/US2014/068904
Other languages
French (fr)
Inventor
Robert H. LEE
Jason John HOWES
Mark A. FALCO
Gene Gleyzer
Original Assignee
Oracle International Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oracle International Corporation filed Critical Oracle International Corporation
Priority to CN201480067510.9A priority Critical patent/CN105830033B/en
Priority to JP2016537499A priority patent/JP6483699B2/en
Priority to EP14819219.8A priority patent/EP3080698A1/en
Publication of WO2015088918A1 publication Critical patent/WO2015088918A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/182Distributed file systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1415Saving, restoring, recovering or retrying at system level
    • G06F11/142Reconfiguring to eliminate the error
    • G06F11/1425Reconfiguring to eliminate the error by reconfiguration of node membership
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1415Saving, restoring, recovering or retrying at system level
    • G06F11/1438Restarting or rejuvenating
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • G06F11/1448Management of the data involved in backup or backup restore
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • G06F11/1458Management of the backup or restore process
    • G06F11/1464Management of the backup or restore process for networked environments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • G06F11/1458Management of the backup or restore process
    • G06F11/1469Backup restoration techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1479Generic software techniques for error detection or fault masking
    • G06F11/1482Generic software techniques for error detection or fault masking by means of middleware or OS functionality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/11File system administration, e.g. details of archiving or snapshots
    • G06F16/128Details of file system snapshots on the file-level, e.g. snapshot creation, administration, deletion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/1858Parallel file systems, i.e. file systems supporting multiple processors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/1865Transactional file systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/21Design, administration or maintenance of databases
    • G06F16/219Managing data history or versioning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/27Replication, distribution or synchronisation of data between databases or within a distributed database system; Distributed database system architectures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/84Using snapshots, i.e. a logical point-in-time copy of the data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/505Clust

Definitions

  • the present invention is generally related to computer systems, and is particularly related to supporting persistence in a distributed data grid.
  • a resolver in the distributed data grid can receive a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid. Then, the resolver can select an identifierforeach partition, wherein each selected identifier is associated with a most recent valid version of a partition. Furthermore, the resolver can determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
  • Figure 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention.
  • Figure 2 shows an illustration of supporting persistence in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 3 shows an illustration of using a shared storage to support persistence in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 4 shows an illustration of using distributed local disks to support persistence in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 5 shows an illustration of supporting distributed persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 6 shows an illustration of coordinating persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 7 shows an illustration of supporting consistent partition recovery in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 8 illustrates an exemplary flow chart for supporting distributed persistent store recovery in a distributed data grid in accordance with an embodiment of the invention.
  • Figure 9 shows an illustration of supporting persistent store versioning in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 10 shows an illustration of supporting persistent store integrity in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 11 shows an illustration of restoring the persisted partitions in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 12 illustrates an exemplary flow chart for supporting persistent store versioning and integrity and in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 13 shows an illustration of providing a persistent snapshot of a running system in a distributed data grid, in accordance with an embodiment of the invention.
  • Figure 14 illustrates an exemplary flow chart for providing a persistent snapshot of a running system in a distributed data grid in accordance with an embodiment of the invention.
  • Figure 15 shows an exemplary block graph illustrating the resolver in accordance with an embodiment of the invention.
  • Figure 16 is an illustration of a functional configuration of an embodiment of the invention.
  • Figure 17 is an illustration of a computer system for implementing an embodiment of the invention.
  • a “data grid cluster”, or “data grid” is a system comprising a plurality of computer servers which work together to manage information and related operations, such as computations, within a distributed or clustered environment.
  • the data grid cluster can be used to manage application objects and data that are shared across the servers.
  • a data grid cluster should have low response time, high throughput, predictable scalability, continuous availability and information reliability. As a result of these capabilities, data grid clusters are well suited for use in computational intensive, stateful middle-tier applications.
  • Some examples of data grid clusters can store the information in-memoryto achieve higher performance, and can employ redundancy in keeping copies of that information synchronized across multiple servers, thus ensuring resiliency of the system and the availability of the data in the event of server failure.
  • Coherence provides replicated and distributed (partitioned) data management and caching services on top of a reliable, highly scalable peer-to-peer clustering protocol.
  • An in-memory data grid can provide the data storage and management capabilities by distributing data over a number of servers working together.
  • the data grid can be middleware that runs in the same tier as an application server or within an application server. It can provide management and processing of data and can also push the processing to where the data is located in the grid.
  • the in-memory data grid can eliminate single points of failure by automatically and transparently failing over and redistributing its clustered data management services when a server becomes inoperative or is disconnected from the network. When a new server is added, or when a failed server is restarted, it can automatically join the cluster and services can be failed back over to it, transparently redistributing the cluster load.
  • the data grid can also include network-level fault tolerance features and transparent soft re-start capability.
  • the functionality of a data grid cluster is based on using different cluster services.
  • the cluster services can include root cluster services, partitioned cache services, and proxy services.
  • each cluster node can participate in a number of cluster services, both in terms of providing and consuming the cluster services.
  • Each cluster service has a service name that uniquely identifies the service within the data grid cluster, and a service type, which defines what the cluster service can do.
  • the services can be either configured by the user, or provided by the data grid cluster as a default set of services.
  • FIG. 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention.
  • a data grid cluster 100 e.g. an Oracle Coherence data grid, includes a plurality of cluster members (or server nodes) such as cluster nodes 101-106, having various cluster services 1 1 1-1 16 running thereon. Additionally, a cache configuration file 1 10 can be used to configure the data grid cluster 100.
  • cluster members or server nodes
  • cluster services 1 1 1-1 16 running thereon.
  • a cache configuration file 1 10 can be used to configure the data grid cluster 100.
  • the distributed data grid can provide recoverable persistent storage for different types of cache content and can prevent data loss after the distributed data grid is shut down.
  • FIG. 2 shows an illustration of supporting persistence in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 200 can include various types of cache content 21 1 -213 in an in-memory data store 202.
  • the distributed data grid 200 can use a persistence layer 201 to persist the cache content 21 1 -213 in a persistent storage 203.
  • the persistence layer 201 allows the persistent storage 203 to use different physical topologies.
  • the persistence layer 201 can store the cache content in a central location, such as a storage area network (SAN) 221 , where all members in the distributed data grid 200 can share the same visibility.
  • the persistence Iayer201 can store the cache content into different local disks 222, where members of the distributed data grid 200 may have only local visibility.
  • the persistence layer 201 can be agnostic to the choice of the physical topology (e.g. a SAN 221 or distributed local disks 222).
  • the distributed data grid 200 can take advantage of multiple SANs or multiple SAN mount points.
  • the distributed data grid 200 can take advantage of a physical topology that includes multiple SANs that are not shared by the plurality of members.
  • the physical topology may include multiple SANs exporting storage locations, or may include hybrid deployments of local disks and SANs.
  • the persistence layer 201 can support partition-wide atomicity of persisted data/metadata, and can provide transaction guarantee after a restart of the distributed data grid 200. Also, the persistence layer 201 can minimize performance impact and reduce recovery time needed to restart the distributed data grid 200.
  • FIG. 3 shows an illustration of using a shared storage to support persistence in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 300 which includes a plurality of members (e.g. the members 301-305 on the machines A-C 31 1 -313), can support various cache services 320.
  • the distributed data grid 300 can use a shared persistent storage, such as a storage area network (SAN) 310, to store the cache content for the cache services 320 in a central location.
  • SAN storage area network
  • the different members 301 -305 on the machines A-C 31 1- 313 can share the same visibility, and can all have access to the persisted partitions 322 in the SAN 310.
  • the system can recover the persisted cache content and prevent data loss, when the distributed data grid 300 is restarted after a shutdown.
  • FIG. 4 shows an illustration of using distributed local disks to support persistence in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 400 which includes a plurality of members (e.g. the members 401-405 on the machines A-C 41 1 -413), can support various cache services 420.
  • the distributed data grid 400 can store the cache content for the cache services 420 into the local disks on different machines.
  • the members 401-402 can store the related cache content into the local disk A 431 on machine A 41 1 (e.g. the persisted partitions 421 ).
  • the members 403-404 can store the related cache content into the local disk B 432 on the machine B 412 (e.g. the persisted partitions 422), and the machine C 413 can store the related cache content into the local disk C 433 on the machine C 413 (e.g. the persisted partitions 423).
  • the distributed data grid 400 can support the automatic recovery of various types of cache content in a distributed fashion, and prevent data loss during the restart of the distributed data grid 400.
  • the distributed data grid can support persistent store recovery in a distributed fashion.
  • FIG. 5 shows an illustration of supporting distributed persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 500 can include a plurality of members, e.g. members 501-505, and can persist the cache content using the distributed local disks, e.g. local disks A-C 51 1-513.
  • each member in the distributed data grid 500 may only have visibility to the partitions persisted in the local disk.
  • the member 501 and the member 502 may only be aware of the persisted partitions 521 in the local disk A 511
  • the member 503 and the member 504 may only be aware of the persisted partitions 522 in the local disk B 512
  • the member 505 may only be aware of the persisted partitions 523 in the local disk C 513.
  • the distributed data grid 500 can use an internal protocol to discover the persisted partitions 521 -523 on different local disks A-C 51 1-513.
  • the discovery protocol supports the persistent store recovery during both the cluster cold-start/restart scenario and the multiple-node failure scenario (e.g. with a loss of a primary owner of a partition and/or one or more backup owners of the partition).
  • the distributed data grid 500 can use, a coordinator such as a coordinator member 510, to coordinate the recovery of various persisted partitions 521-523 in the distributed data grid 500.
  • the coordinator member 510 can send a distributed query to other members 501 -505 in the distributed data grid 500 in order to obtain a complete list of persisted partitions 521 -523.
  • a member (501 - 505, 510) is an element of a distributed data grid.
  • One member 510 of a plurality of members of the distributed data grid can work as a coordinator, which handles information for a recovery of various persisted partitions in a distributed data grid.
  • the handling of information may include synchronizing a view of partition ownership among the plurality of members in the distributed data grid, and receiving information on the plurality of persisted partitions from the plurality of members in the distributed data grid.
  • the coordinator may be referred to as a "coordinator member 510".
  • the coordinator member 510 can use a pluggable partition assignment strategy component 520 to determine the partition recovery assignment 540. For example, the system can go down the list of the partitions to examine which member can see a version of the partition. Then, the system can determine which member should be used to recover which partition based on a synchronized partition ownership view 530.
  • the system can minimize the performance impact caused by adding persistence support to the distributed data grid 500.
  • the system can use an asynchronous messaging process in the distributed data grid 500 for implementing the write operation to a persistent store.
  • the system allows the performing of multiple input/output (I/O) operations concurrently.
  • the coordinator member 510 can avoid using only one or a few members in the distributed data grid 500 for performing the recovery, which may be prone to create performance bottleneck.
  • the system can use a recovery quorum to ensure that all persisted partitions are visible prior to the recovery in order to prevent data loss due to recovery.
  • the distributed data grid 500 can automatically carry out a recovery of persisted cache contents in a distributed fashion during a restart of the distributed data grid 500.
  • FIG. 6 shows an illustration of coordinating persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention.
  • a coordinator member 610 in a distributed data grid 600 can coordinate the recovery of the persisted partitions from the distributed local disks.
  • the coordinator member 610 can direct a member 620 to recover persisted partitions from a local disk 630.
  • the coordinator 610 can instruct the member 620 (and all other members in the distributed data grid 600 concurrently) to prepare for restoring persisted partitions. Then, at step 602, the member 620 (possibly along with each other member in the distributed data grid 600) can provide a local partition ownership back to the coordinator member 610.
  • the coordinator member 610 can synchronize a view of the overall partition ownership, after obtaining the partition ownership information from the different members in the distributed data grid 600.
  • the coordinator 610 can instruct the member 620 to prepare for recovering the persisted partitions based on the view of the overall partition ownership.
  • the member 620 can check for the persisted partitions in the local disk 630. Then, at step 606, the member 620 can report the persisted partitions (e.g. the persisted partition IDs) in the local disk 630 to the coordinator member 610.
  • the coordinator member 610 can make decision on how to configure a recovery process, such as determining a recovery assignment.
  • the coordinator 610 can provide the partition recovery assignment (e.g. the recover partition IDs) to each member in the distributed data grid 600.
  • the different members in the distributed data grid 600 can carry out the recovery of the persisted partitions based on the received partition recovery assignment.
  • FIG. 7 shows an illustration of supporting consistent partition recovery in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 700 can include a plurality of members, e.g. members 701-705, each of which may only have visibility to the partitions persisted in the local disk.
  • a coordinator member 710 can coordinate the recovery of various persisted partitions 721 -723 from the distributed local disks A-C 711 -713. Also, the coordinator member 710 can use a pluggable partition assignment strategy component 720 to determine which member should be used to recover which partition.
  • the system can promote in-memory backups to in-memory primaries.
  • the system can create a new persisted partition on disk and can also create one or more in-memory backups on other members from the data in memory.
  • the system can recover a new in- memory primary from the persisted version on disk, when there is a member having visibility to the disk.
  • the distributed data grid 700 can rebalance itself.
  • the distributed data grid 700 can promote a back-up partition which is persisted in either the local disk B 712 or the local disk C 713 as the primary partition.
  • the distributed data grid 700 can ensure that the system always restores the most recent valid partition.
  • the persisted partitions 722 in the local disk B 712 may contain a newer version of the partition, since the persisted partitions 721 in the local disk A 71 1 may not be updated correctly or an older version of the partition exists due to the death of the prior owner of the partition.
  • the distributed data grid 700 can use a recovery quorum for supporting the discovery and/or the recovery of the persisted partitions 721-723.
  • the recovery quorum By using the recovery quorum, the recovery from persistence can be gated or protected.
  • the distributed data grid 700 can ensure that no data is lost, even when the number of members that are lost exceeds the in-memory redundancy target.
  • the distributed data grid 700 can ensure that all persisted partitions are visible prior to recovery.
  • the recovery quorum can be configured such that it guarantees visibility to all of the possible storage locations (such as local disks and/or SANs within the cluster).
  • the distributed data grid 700 can recover orphaned partitions from the persistent store and assign them as empty partitions
  • the distributed data grid 700 can establish different recovery policies based on the recovery quorum.
  • the distributed data grid 700 can establish SAN/shared-storage policies that focus on capacity.
  • the distributed data grid 700 can establish distributed/shared-nothing storage policies that ensure all storage locations are reachable.
  • the distributed data grid 700 can establish various policies based on the configured membership size and the host-list.
  • the system allows various members 701-705 in the distributed data grid 700 to be shut down (and/or restarted) in an orderly fashion, and allows for a graceful suspend/resume of an service or the entire cluster. Additionally, the system can prevent partition transfers and persistent store movements, during the shutdown of the distributed data grid. For example, a quiesced service/cluster may not join new members, may not restore partitions from backup, may not recover orphaned partitions from persistent store, may not assign empty orphaned partitions, and may not perform partition distribution.
  • FIG. 8 illustrates an exemplary flow chart for supporting distributed persistent store recovery in a distributed data grid in accordance with an embodiment of the invention.
  • the system allowing a plurality of members in the distributed data grid to persist a plurality of partitions associated with one or more cache services in a persistent storage.
  • a coordinator can synchronize a view of partition ownership among the plurality of members in the distributed data grid.
  • the distributed data grid can form, based on the synchronized view, a distributed consensus on which partition can be recovered from which member in the distributed data grid.
  • the distributed consensus is an agreement reached among a plurality of cluster members (or server nodes) such as cluster nodes 101 -106, which cooperate to solve a problem.
  • FIG. 9 shows an illustration of supporting persistent store versioning in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 900 can use various partitions (e.g. a partition 901 ) in an in-memory data store 920 to support different cache services.
  • the distributed data grid 900 can use a persistent store (e.g. a persisted partition 91 1 ) to persist the partition 901 in the distributed local disks 910.
  • a persistent store e.g. a persisted partition 91 1
  • the system can provide a unique identifier (ID), or a unique version number 906, for each persisted partition in the distributed local disks 910.
  • ID unique identifier
  • a member 902 in the distributed data grid 900 can generate a globally unique identifier (GUID) 921 for the persistent partition 91 1.
  • GUID 921 can contain various types of information using a special naming format.
  • the GUID 921 can include at least a partition number (or a partition ID 903) and a partition version number 91 1 associated with the partition 901. Additionally, the GUID 921 can contain a member ID 904, which indicates that the member 902 generates the GUID 921.
  • the GUID 921 can include other information, such as a time stamp 905 that indicates the time when the partition 901 is first persisted.
  • the time stamp 905 is a stamp of logical time (e.g. a stamp of a vector clock per partition), instead of a global wall clock.
  • the system can guarantee that the GUID stamps move monotonically forward in the face of any kind of failure or transfer scenario.
  • the distributed data grid 900 can maintain the version number 910 for each persisted partition in a monotonically increasing order.
  • the system can account for the data mutation at any member or ownership changes in the distributed data grid 900.
  • FIG 10 shows an illustration of supporting persistent store integrity in a distributed data grid, in accordance with an embodiment of the invention.
  • a persistent store 1001 in a distributed data grid 1000 can contain cache content from different caches A-C 101 1 -1013, each of which is associated with a cache ID 1021 -1 123.
  • the system can apply a seal operation 1002 on the persistent store 1001 .
  • the seal operation 1002 can ensure that the persistent store 1001 is fully initialized and is eligible to be recovered.
  • the system can apply a validation operation 1003 on the persistent store 1001 .
  • the validation operation 1003 can check whether the persistent store 1001 has been sealed. For example, the system may decide that the cache content in the persistent store 1001 is not valid if the persistent store 1001 is not sealed.
  • the system can ensure that the distributed data grid 1000 always restores a valid persisted partition and avoids recovering a partial copy that may be caused by cascading cluster failures.
  • FIG 11 shows an illustration of restoring the persisted partitions in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 1 100 can store various persisted partitions 1 1 1 1-1 1 13 in distributed local disks 1 1 10.
  • Each persisted partition 1 1 1 1 1 -1 1 13 stored in the distributed local disks 1 1 10 can be associated with a globally unique identifier (GUID), e.g. GUID 1 141-1 143.
  • GUIDs 1 141-1 143 can contain different types of information that includes at least a partition number (i.e. a partition - id) and a version number.
  • the members 1 101-1 102 in the distributed data grid 1 100 may have different visibility to the persisted partitions 101 1 -1013 in the distributed local disks 1 1 10.
  • the system can configure the GUIDs 1 141-1 143 to contain information on which member may have visibility to a particular persisted partition 1 1 1 1 -1 1 13.
  • each of the members 1 101 -1 102 in the distributed data grid 1 100 can report the GUIDs 1 141 -1 143 (which can include the partition numbers and other information) for each of the persisted partitions that are found. In accordance with an embodiment of the invention, only members reporting the presence of the most recent GUID for a partition can be considered for recovery.
  • each member 1 101-1 102 in the distributed data grid 1 100 can collect a list of available GUIDs 1 121 -1 122 from the distributed local disks 1 1 10 based on local visibility. Then, each member 1 101 -1 102 can provide (or register) the list of available GUIDs 1 121 -1 122 to a resolver 1 103 in the distributed data grid 1 100, and the resolver 1 103 can determine the newest GUIDs 1 130 for different partitions based on the partition number and version number information encoded in the GUIDs 1 141 -1 143.
  • the distributed local disks 1 1 10 may contain multiple different versions of the same partition.
  • the resolver 1 103 may receive multiple GUIDs that contain the same partition number and different version numbers.
  • the resolver 1 103 can obtain the version number from each GUID associated with the same partition, and determine which GUID has the most recent version number. Also, the distributed data grid 1 100 can ensure that the persisted partition with the most recent version number is valid based on performing the seal operation and validation operation.
  • the resolver 1 103 can determine which member 1 101-1 102 in the distributed data grid 1 100 is responsible for recovering a particular persisted partition 1 1 1 1-1 1 13, based on the member ID information encoded in the GUIDs 1 141 -1 143.
  • the resolver 1 103 can provide the partition recovery assignment, which may include a list of the newest GUIDs 1 131 -1 132, to each different member 1 101-1 102. Accordingly, the members 1 101 -1 102 can carry out the actual operation that restores the persisted partitions 1 1 1 1 -1 1 13.
  • the system can ensure that the distributed data grid 1 100 always restores the newest valid version of any persisted partition, and can avoid recovering a partial copy that may be caused by cascading cluster failures.
  • Figure 12 illustrates an exemplary flow chart for supporting persistent store versioning and integrity and in a distributed data grid, in accordance with an embodiment of the invention.
  • the system can receive a plurality of identifiers (e.g. the GUIDs) from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid.
  • the system can select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition.
  • the system can determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
  • FIG. 13 shows an illustration of providing a persistent snapshot of a running system in a distributed data grid, in accordance with an embodiment of the invention.
  • a distributed data grid 1300 can support various cache services 1320 using an in-memory data store 1302.
  • the system allows a user to use a management tool 1310 to take a snapshot 1301 of the running system on the in-memory data store 1302 that supports the cache services 1320 on-demand, at any particular time.
  • the snapshot 1301 can be used to make a backup of the running system overnight.
  • the system can suspend the cache services 1320, prior to taking the snapshot 1301.
  • the system can provide a consistent point in time for taking the snapshot 1301.
  • the cache service 1320 can be resumed after the snapshot 1301 is taken.
  • the snapshot 1301 can provide a consistent view of each partitioned cache service 1320.
  • the snapshot 1301 can provide a catalogue of state information of the running system, including metadata 1311 and cache data 1312 for the partitioned cache services 1320.
  • the system can store the snapshot 1301 either in a central location (e.g. a SAN 1321 ) or in distributed local disks 1322.
  • the system can use a pluggable (or portable) archiver 1303 to retrieve the persisted state information of the snapshot 1301 from the distributed local disks 1322, and can create a single archive unit 1330, which can be used for auditing or other purposes.
  • the system allows a user to take a snapshot on the state of a partitioned cache service in a distributed data grid 1300, instead of persisting the cache content in the distributed data grid 1300 in a continuing fashion.
  • Figure 14 illustrates an exemplary flow chart for providing a persistent snapshot of a running system in a distributed data grid in accordance with an embodiment of the invention.
  • the system allows one or more cache services to run on a plurality of cluster members in the distributed data grid.
  • the system can collect a catalogue of state information associated with said one or more cache services from the plurality of cluster members in the distributed data grid.
  • the system can create a snapshot for said one or more cache services running on the distributed data grid.
  • Figure 15 shows an exemplary block graph illustrating the resolver in accordance with an embodiment of the invention.
  • the blocks of the resolver 1500 may be implemented by hardware, software, or a combination of hardware and software to carry out the principles of the invention. It is understood by persons of skill in the art that the blocks described in Figure 15 may be combined or separated into sub-blocks to implement the principles of the invention as described above. Therefore, the description herein may support any possible combination or separation orfurther definition of the functional blocks described herein.
  • Figure 15 there is shown a resolver (1500) that may be used for a distributed data grid, such as any distributed data grid described previously, in particular, distributed data grid shown in Figure 1 1. Therefore, the resolver 1500 may be any resolver mentioned in the embodiments described previously. Moreover, the resolver 1500 and the component therein described below may perform various operations described previously according to the principle of the invention, but not limited to the operations and functions described below.
  • the resolver 1500 may comprise a receiving unit 1501 which may be configured to receive a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid.
  • the resolver 1500 may further comprise a selecting unit 1502 which may be configured to select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition.
  • the resolver 1500 may further comprise a determining unit 1503 which may be configured to determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
  • the resolver 1500 may further comprise a resolving unit 1504 which may be configured to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
  • the system for supporting persistence in a distributed data grid may comprise a persistent storage, configured to store one or more persisted partitions, each of which is associated with an identifier; one or more members, configured to collect a list of available identifiers from the persistent storage; and a resolver, configured to receive the list of available identifiers from one or more members, select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition, and determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
  • the resolver herein may be embodied as the resolver 1500 as shown in Figure 15.
  • the persistent storage may comprise a plurality of distributed local disk, wherein each member in the distributed data grid only has visibility to one or more distribute local disks.
  • an identifier is assigned to each persisted partition stored in the persistent storage, wherein each said identifier is associated with a partition number and a version number of a partition.
  • the resolver may be further configured to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
  • the persistent storage conprises a storage area network (SAN), wherein the SAN is visible to a plurality of members in the distributed data grid.
  • SAN storage area network
  • different members in the distributed data grid operate to persist multiple versions of a partition in the persistent storage.
  • the distributed data grid operates to seal a persistent store in the persistent storage, and indicate that the sealed persistent store is fully initialized and eligible to be recovered, and validate a persistent store in the persistent storage to determine whether said persistent store is sealed.
  • said determined member operate to recover said partition using a persisted partition with a selected identifier.
  • system 1600 is illustrated in accordance with an embodiment of the invention.
  • Figure 16 shows an illustration of a functional configuration realized by system 1600.
  • system 1600 includes a persistent storage 1610, one or more members 1620, a receiver module 1630, a selector 1640, and a determination module 1650.
  • Receiver module receives a plurality of identifiers from one or more members 1620 of a distributed data grid. Each identifier is associated with a persisted partition in persistent storage 1610 for the distributed data grid. Selector 1640 selects an identifier for each partition. Each selected identifier is associated with a most recent valid version of a partition. Determination module 1650 determines a member in the distributed data grid that is responsible for recovering the partition from a persisted partition associated with the selected identifier.
  • FIG. 17 shows an illustration of a computer system 1700 which includes well- known hardware elements.
  • computer system 1700 includes a central processing unit (CPU) 1710, a mouse 1720, a key board 1730, a random access memory (RAM) 1740, a hard disc 1750, a disc drive 1760, a communication interface (l/F) 1770, and a monitor 1780.
  • Computer system 1700 may function as a server node constituting system 1600.
  • persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 are provided by one or more computer systems 1700.
  • Persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 are implemented by CPU 1710.
  • more than one processors can be used so that persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 are implemented. Namely, any of persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 can be physically remote from each other.
  • system 1600 can be realized by using a plurality of hardwired circuits which function as persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650.
  • the present invention may be conveniently implemented using one or more conventional general purpose or specialized digital computer, computing device, machine, or microprocessor, including one or more processors, memory and/or computer readable storage media programmed according to the teachings of the present disclosure.
  • Appropriate software coding can readily be prepared by skilled programmers based on the teachings of the present disclosure, as will be apparent to those skilled in the software art.
  • the present invention includes a computer program product which is a storage medium or computer readable medium (media) having instructions stored thereon/in which can be used to program a computer to perform any of the processes of the present invention.
  • the storage medium can include, but is not limited to, any type of disk including floppy disks, optical discs, DVD, CD-ROMs, microdrive, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, DRAMs, VRAMs, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Quality & Reliability (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Retry When Errors Occur (AREA)

Abstract

A system and method can support persistence in a distributed data grid, such as persistent store versioning and integrity. A resolver in the distributed data grid can receive a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid. Then, the resolver can select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition. Furthermore, the resolver can determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.

Description

SYSTEM AND METHOD FOR SUPPORTING PERSISTENT STORE VERSIONING AND INTEGRITY IN A DISTRIBUTED DATA GRID
Copyright Notice:
[0001] A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever. Field of Invention:
[0002] The present invention is generally related to computer systems, and is particularly related to supporting persistence in a distributed data grid.
Background:
[0003] Modern computing systems, particularly those employed by larger organizations and enterprises, continue to increase in size and complexity. Particularly, in areas such as Internet applications, there is an expectation that millions of users should be able to simultaneously access that application, which effectively leads to an exponential increase in the amount of content generated and consumed by users, and transactions involving that content. Such activity also results in a corresponding increase in the number of transaction calls to databases and metadata stores, which have a limited capacity to accommodate that demand. This is the general area that embodiments of the invention are intended to address.
Summary:
[0004] Described herein are systems and methods that can support persistence in a distributed data grid, such as persistent store versioning and integrity. A resolver in the distributed data grid can receive a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid. Then, the resolver can select an identifierforeach partition, wherein each selected identifier is associated with a most recent valid version of a partition. Furthermore, the resolver can determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier. Brief Description of the Figures:
[0005] Figure 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention. [0006] Figure 2 shows an illustration of supporting persistence in a distributed data grid, in accordance with an embodiment of the invention.
[0007] Figure 3 shows an illustration of using a shared storage to support persistence in a distributed data grid, in accordance with an embodiment of the invention.
[0008] Figure 4 shows an illustration of using distributed local disks to support persistence in a distributed data grid, in accordance with an embodiment of the invention.
[0009] Figure 5 shows an illustration of supporting distributed persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention.
[0010] Figure 6 shows an illustration of coordinating persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention.
[0011] Figure 7 shows an illustration of supporting consistent partition recovery in a distributed data grid, in accordance with an embodiment of the invention.
[0012] Figure 8 illustrates an exemplary flow chart for supporting distributed persistent store recovery in a distributed data grid in accordance with an embodiment of the invention.
[0013] Figure 9 shows an illustration of supporting persistent store versioning in a distributed data grid, in accordance with an embodiment of the invention.
[0014] Figure 10 shows an illustration of supporting persistent store integrity in a distributed data grid, in accordance with an embodiment of the invention.
[0015] Figure 11 shows an illustration of restoring the persisted partitions in a distributed data grid, in accordance with an embodiment of the invention.
[0016] Figure 12 illustrates an exemplary flow chart for supporting persistent store versioning and integrity and in a distributed data grid, in accordance with an embodiment of the invention.
[0017] Figure 13 shows an illustration of providing a persistent snapshot of a running system in a distributed data grid, in accordance with an embodiment of the invention.
[0018] Figure 14 illustrates an exemplary flow chart for providing a persistent snapshot of a running system in a distributed data grid in accordance with an embodiment of the invention.
[0019] Figure 15 shows an exemplary block graph illustrating the resolver in accordance with an embodiment of the invention.
[0020] Figure 16 is an illustration of a functional configuration of an embodiment of the invention.
[0021 ] Figure 17 is an illustration of a computer system for implementing an embodiment of the invention.
Detailed Description:
[0022] Described herein are systems and methods that can support persistence in a distributed data grid. Distributed Data Grid
[0023] In accordance with an embodiment, as referred to herein a "data grid cluster", or "data grid", is a system comprising a plurality of computer servers which work together to manage information and related operations, such as computations, within a distributed or clustered environment. The data grid cluster can be used to manage application objects and data that are shared across the servers. Preferably, a data grid cluster should have low response time, high throughput, predictable scalability, continuous availability and information reliability. As a result of these capabilities, data grid clusters are well suited for use in computational intensive, stateful middle-tier applications. Some examples of data grid clusters, e.g., the Oracle Coherence data grid cluster, can store the information in-memoryto achieve higher performance, and can employ redundancy in keeping copies of that information synchronized across multiple servers, thus ensuring resiliency of the system and the availability of the data in the event of server failure. For example, Coherence provides replicated and distributed (partitioned) data management and caching services on top of a reliable, highly scalable peer-to-peer clustering protocol.
[0024] An in-memory data grid can provide the data storage and management capabilities by distributing data over a number of servers working together. The data grid can be middleware that runs in the same tier as an application server or within an application server. It can provide management and processing of data and can also push the processing to where the data is located in the grid. In addition, the in-memory data grid can eliminate single points of failure by automatically and transparently failing over and redistributing its clustered data management services when a server becomes inoperative or is disconnected from the network. When a new server is added, or when a failed server is restarted, it can automatically join the cluster and services can be failed back over to it, transparently redistributing the cluster load. The data grid can also include network-level fault tolerance features and transparent soft re-start capability.
[0025] In accordance with an embodiment, the functionality of a data grid cluster is based on using different cluster services. The cluster services can include root cluster services, partitioned cache services, and proxy services. Within the data grid cluster, each cluster node can participate in a number of cluster services, both in terms of providing and consuming the cluster services. Each cluster service has a service name that uniquely identifies the service within the data grid cluster, and a service type, which defines what the cluster service can do. Other than the root cluster service running on each cluster node in the data grid cluster, there may be multiple named instances of each service type. The services can be either configured by the user, or provided by the data grid cluster as a default set of services.
[0026] Figure 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention. As shown in Figure 1 , a data grid cluster 100, e.g. an Oracle Coherence data grid, includes a plurality of cluster members (or server nodes) such as cluster nodes 101-106, having various cluster services 1 1 1-1 16 running thereon. Additionally, a cache configuration file 1 10 can be used to configure the data grid cluster 100.
Persistent Storage of Cache Contents
[0027] In accordance with an embodiment of the invention, the distributed data grid can provide recoverable persistent storage for different types of cache content and can prevent data loss after the distributed data grid is shut down.
[0028] Figure 2 shows an illustration of supporting persistence in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 2, a distributed data grid 200 can include various types of cache content 21 1 -213 in an in-memory data store 202. Furthermore, the distributed data grid 200 can use a persistence layer 201 to persist the cache content 21 1 -213 in a persistent storage 203.
[0029] The persistence layer 201 allows the persistent storage 203 to use different physical topologies. For example, the persistence layer 201 can store the cache content in a central location, such as a storage area network (SAN) 221 , where all members in the distributed data grid 200 can share the same visibility. Alternatively, the persistence Iayer201 can store the cache content into different local disks 222, where members of the distributed data grid 200 may have only local visibility.
[0030] Furthermore, the persistence layer 201 can be agnostic to the choice of the physical topology (e.g. a SAN 221 or distributed local disks 222). For example, the distributed data grid 200 can take advantage of multiple SANs or multiple SAN mount points. Also, the distributed data grid 200 can take advantage of a physical topology that includes multiple SANs that are not shared by the plurality of members. Alternatively, the physical topology may include multiple SANs exporting storage locations, or may include hybrid deployments of local disks and SANs.
[0031] Additionally, the persistence layer 201 can support partition-wide atomicity of persisted data/metadata, and can provide transaction guarantee after a restart of the distributed data grid 200. Also, the persistence layer 201 can minimize performance impact and reduce recovery time needed to restart the distributed data grid 200.
[0032] Figure 3 shows an illustration of using a shared storage to support persistence in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 3, a distributed data grid 300, which includes a plurality of members (e.g. the members 301-305 on the machines A-C 31 1 -313), can support various cache services 320.
[0033] Furthermore, the distributed data grid 300 can use a shared persistent storage, such as a storage area network (SAN) 310, to store the cache content for the cache services 320 in a central location. As shown in Figure 3, the different members 301 -305 on the machines A-C 31 1- 313 can share the same visibility, and can all have access to the persisted partitions 322 in the SAN 310.
[0034] Thus, the system can recover the persisted cache content and prevent data loss, when the distributed data grid 300 is restarted after a shutdown.
[0035] Figure 4 shows an illustration of using distributed local disks to support persistence in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 4, a distributed data grid 400, which includes a plurality of members (e.g. the members 401-405 on the machines A-C 41 1 -413), can support various cache services 420.
[0036] Furthermore, the distributed data grid 400 can store the cache content for the cache services 420 into the local disks on different machines. For example, the members 401-402 can store the related cache content into the local disk A 431 on machine A 41 1 (e.g. the persisted partitions 421 ). Also, the members 403-404 can store the related cache content into the local disk B 432 on the machine B 412 (e.g. the persisted partitions 422), and the machine C 413 can store the related cache content into the local disk C 433 on the machine C 413 (e.g. the persisted partitions 423).
[0037] Thus, the distributed data grid 400 can support the automatic recovery of various types of cache content in a distributed fashion, and prevent data loss during the restart of the distributed data grid 400.
Distributed Persistent Store Recovery
[0038] In accordance with an embodiment of the invention, the distributed data grid can support persistent store recovery in a distributed fashion.
[0039] Figure 5 shows an illustration of supporting distributed persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 5, a distributed data grid 500 can include a plurality of members, e.g. members 501-505, and can persist the cache content using the distributed local disks, e.g. local disks A-C 51 1-513.
[0040] Furthermore, each member in the distributed data grid 500 may only have visibility to the partitions persisted in the local disk. For example, the member 501 and the member 502 may only be aware of the persisted partitions 521 in the local disk A 511 , while the member 503 and the member 504 may only be aware of the persisted partitions 522 in the local disk B 512 and the member 505 may only be aware of the persisted partitions 523 in the local disk C 513.
[0041] In accordance with an embodiment of the invention, the distributed data grid 500 can use an internal protocol to discover the persisted partitions 521 -523 on different local disks A-C 51 1-513. For example, the discovery protocol supports the persistent store recovery during both the cluster cold-start/restart scenario and the multiple-node failure scenario (e.g. with a loss of a primary owner of a partition and/or one or more backup owners of the partition).
[0042] As shown in Figure 5, the distributed data grid 500 can use, a coordinator such as a coordinator member 510, to coordinate the recovery of various persisted partitions 521-523 in the distributed data grid 500. The coordinator member 510 can send a distributed query to other members 501 -505 in the distributed data grid 500 in order to obtain a complete list of persisted partitions 521 -523.
[0043] Thus, in accordance with the embodiment of the present disclosure, a member (501 - 505, 510) is an element of a distributed data grid. One member 510 of a plurality of members of the distributed data grid can work as a coordinator, which handles information for a recovery of various persisted partitions in a distributed data grid. The handling of information may include synchronizing a view of partition ownership among the plurality of members in the distributed data grid, and receiving information on the plurality of persisted partitions from the plurality of members in the distributed data grid. In an aspect, the coordinator may be referred to as a "coordinator member 510".
[0044] In accordance with an embodiment of the invention, the coordinator member 510 can use a pluggable partition assignment strategy component 520 to determine the partition recovery assignment 540. For example, the system can go down the list of the partitions to examine which member can see a version of the partition. Then, the system can determine which member should be used to recover which partition based on a synchronized partition ownership view 530.
[0045] Furthermore, the system can minimize the performance impact caused by adding persistence support to the distributed data grid 500. For example, the system can use an asynchronous messaging process in the distributed data grid 500 for implementing the write operation to a persistent store. Also, the system allows the performing of multiple input/output (I/O) operations concurrently.
[0046] Additionally, the coordinator member 510 can avoid using only one or a few members in the distributed data grid 500 for performing the recovery, which may be prone to create performance bottleneck.
[0047] Also, the system can use a recovery quorum to ensure that all persisted partitions are visible prior to the recovery in order to prevent data loss due to recovery.
[0048] Additional descriptions of various embodiments of supporting service level quorum in a distributed data grid 500 are provided in U.S. Patent Application titled "SYSTEM AND METHOD FOR SUPPORTING SERVICE LEVEL QUORUM IN A DATA GRID CLUSTER", Application No. 13/352,203, filed on January 17, 2012 (Attorney Docket No. ORACL-05131 US2), which application is herein incorporated by reference.
[0049] Thus, the distributed data grid 500 can automatically carry out a recovery of persisted cache contents in a distributed fashion during a restart of the distributed data grid 500.
[0050] Figure 6 shows an illustration of coordinating persistent store recovery in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 6, a coordinator member 610 in a distributed data grid 600 can coordinate the recovery of the persisted partitions from the distributed local disks. For example, the coordinator member 610 can direct a member 620 to recover persisted partitions from a local disk 630.
[0051] At step 601 , the coordinator 610 can instruct the member 620 (and all other members in the distributed data grid 600 concurrently) to prepare for restoring persisted partitions. Then, at step 602, the member 620 (possibly along with each other member in the distributed data grid 600) can provide a local partition ownership back to the coordinator member 610.
[0052] At step 603, the coordinator member 610 can synchronize a view of the overall partition ownership, after obtaining the partition ownership information from the different members in the distributed data grid 600.
[0053] Furthermore, at step 604, the coordinator 610 can instruct the member 620 to prepare for recovering the persisted partitions based on the view of the overall partition ownership. At step 605, the member 620 can check for the persisted partitions in the local disk 630. Then, at step 606, the member 620 can report the persisted partitions (e.g. the persisted partition IDs) in the local disk 630 to the coordinator member 610.
[0054] At step 607, after obtaining information about the persisted partitions from the different members in the distributed data grid 600, the coordinator member 610 can make decision on how to configure a recovery process, such as determining a recovery assignment.
[0055] Then, at step 608, the coordinator 610 can provide the partition recovery assignment (e.g. the recover partition IDs) to each member in the distributed data grid 600. Finally, at step 609, the different members in the distributed data grid 600 (including the member 620) can carry out the recovery of the persisted partitions based on the received partition recovery assignment.
[0056] Figure 7 shows an illustration of supporting consistent partition recovery in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 7, a distributed data grid 700 can include a plurality of members, e.g. members 701-705, each of which may only have visibility to the partitions persisted in the local disk.
[0057] Furthermore, a coordinator member 710 can coordinate the recovery of various persisted partitions 721 -723 from the distributed local disks A-C 711 -713. Also, the coordinator member 710 can use a pluggable partition assignment strategy component 720 to determine which member should be used to recover which partition.
[0058] In accordance with an embodiment of the invention, when a machine in the distributed data grid 700 is lost, the system can promote in-memory backups to in-memory primaries. As part of this process, the system can create a new persisted partition on disk and can also create one or more in-memory backups on other members from the data in memory.
[0059] Additionally, when in-memory data loss occurs due to two or more (depending on the backup count) member processes dying simultaneously, the system can recover a new in- memory primary from the persisted version on disk, when there is a member having visibility to the disk.
[0060] As shown in Figure 7, when a machine that is associated with the local disk A 71 1 is lost, the persisted partitions 721 may become unavailable. In such a case, the distributed data grid 700 can rebalance itself. For example, the distributed data grid 700 can promote a back-up partition which is persisted in either the local disk B 712 or the local disk C 713 as the primary partition.
[0061] In accordance with an embodiment of the invention, the distributed data grid 700 can ensure that the system always restores the most recent valid partition. For example, the persisted partitions 722 in the local disk B 712 may contain a newer version of the partition, since the persisted partitions 721 in the local disk A 71 1 may not be updated correctly or an older version of the partition exists due to the death of the prior owner of the partition.
[0062] In accordance with an embodiment of the invention, the distributed data grid 700 can use a recovery quorum for supporting the discovery and/or the recovery of the persisted partitions 721-723. By using the recovery quorum, the recovery from persistence can be gated or protected. Thus, the distributed data grid 700 can ensure that no data is lost, even when the number of members that are lost exceeds the in-memory redundancy target.
[0063] Also, the distributed data grid 700 can ensure that all persisted partitions are visible prior to recovery. For example, the recovery quorum can be configured such that it guarantees visibility to all of the possible storage locations (such as local disks and/or SANs within the cluster). Additionally, the distributed data grid 700 can recover orphaned partitions from the persistent store and assign them as empty partitions
[0064] Furthermore, the distributed data grid 700 can establish different recovery policies based on the recovery quorum. For example, the distributed data grid 700 can establish SAN/shared-storage policies that focus on capacity. Also, the distributed data grid 700 can establish distributed/shared-nothing storage policies that ensure all storage locations are reachable. Also, the distributed data grid 700 can establish various policies based on the configured membership size and the host-list.
[0065] In accordance with an embodiment of the invention, the system allows various members 701-705 in the distributed data grid 700 to be shut down (and/or restarted) in an orderly fashion, and allows for a graceful suspend/resume of an service or the entire cluster. Additionally, the system can prevent partition transfers and persistent store movements, during the shutdown of the distributed data grid. For example, a quiesced service/cluster may not join new members, may not restore partitions from backup, may not recover orphaned partitions from persistent store, may not assign empty orphaned partitions, and may not perform partition distribution.
[0066] Figure 8 illustrates an exemplary flow chart for supporting distributed persistent store recovery in a distributed data grid in accordance with an embodiment of the invention. As shown in Figure 8, at step 801 , the system allowing a plurality of members in the distributed data grid to persist a plurality of partitions associated with one or more cache services in a persistent storage. Then, at step 802, a coordinator can synchronize a view of partition ownership among the plurality of members in the distributed data grid. Furthermore, at step 803, the distributed data grid can form, based on the synchronized view, a distributed consensus on which partition can be recovered from which member in the distributed data grid. The distributed consensus is an agreement reached among a plurality of cluster members (or server nodes) such as cluster nodes 101 -106, which cooperate to solve a problem.
Persistent Store Versioning and Integrity
[0067] Figure 9 shows an illustration of supporting persistent store versioning in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 9, a distributed data grid 900 can use various partitions (e.g. a partition 901 ) in an in-memory data store 920 to support different cache services.
[0068] Furthermore, the distributed data grid 900 can use a persistent store (e.g. a persisted partition 91 1 ) to persist the partition 901 in the distributed local disks 910.
[0069] The system can provide a unique identifier (ID), or a unique version number 906, for each persisted partition in the distributed local disks 910. As shown in Figure 9, a member 902 in the distributed data grid 900 can generate a globally unique identifier (GUID) 921 for the persistent partition 91 1. The GUID 921 can contain various types of information using a special naming format.
[0070] For example, the GUID 921 can include at least a partition number (or a partition ID 903) and a partition version number 91 1 associated with the partition 901. Additionally, the GUID 921 can contain a member ID 904, which indicates that the member 902 generates the GUID 921.
[0071] Additionally, the GUID 921 can include other information, such as a time stamp 905 that indicates the time when the partition 901 is first persisted. The time stamp 905 is a stamp of logical time (e.g. a stamp of a vector clock per partition), instead of a global wall clock. Thus, the system can guarantee that the GUID stamps move monotonically forward in the face of any kind of failure or transfer scenario.
[0072] In accordance with an embodiment of the invention, the distributed data grid 900 can maintain the version number 910 for each persisted partition in a monotonically increasing order. Thus, the system can account for the data mutation at any member or ownership changes in the distributed data grid 900.
[0073] Figure 10 shows an illustration of supporting persistent store integrity in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 10, a persistent store 1001 in a distributed data grid 1000 can contain cache content from different caches A-C 101 1 -1013, each of which is associated with a cache ID 1021 -1 123.
[0074] Furthermore, the system can apply a seal operation 1002 on the persistent store 1001 . The seal operation 1002 can ensure that the persistent store 1001 is fully initialized and is eligible to be recovered.
[0075] Additionally, the system can apply a validation operation 1003 on the persistent store 1001 . The validation operation 1003 can check whether the persistent store 1001 has been sealed. For example, the system may decide that the cache content in the persistent store 1001 is not valid if the persistent store 1001 is not sealed.
[0076] Thus, the system can ensure that the distributed data grid 1000 always restores a valid persisted partition and avoids recovering a partial copy that may be caused by cascading cluster failures.
[0077] Figure 11 shows an illustration of restoring the persisted partitions in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 1 1 , a distributed data grid 1 100 can store various persisted partitions 1 1 1 1-1 1 13 in distributed local disks 1 1 10.
[0078] Each persisted partition 1 1 1 1 -1 1 13 stored in the distributed local disks 1 1 10 can be associated with a globally unique identifier (GUID), e.g. GUID 1 141-1 143. The GUIDs 1 141-1 143 can contain different types of information that includes at least a partition number (i.e. a partition - id) and a version number.
[0079] In accordance with an embodiment of the invention, the members 1 101-1 102 in the distributed data grid 1 100 may have different visibility to the persisted partitions 101 1 -1013 in the distributed local disks 1 1 10. The system can configure the GUIDs 1 141-1 143 to contain information on which member may have visibility to a particular persisted partition 1 1 1 1 -1 1 13.
[0080] Additionally, as a result of a cascading failure in the distributed local disks 1 1 10, multiple versions of the same persisted partitions 101 1-1013 may present on the different members 1 101 -1 102 of the distributed data grid 1 100. In order to disambiguate these different versions, each of the members 1 101 -1 102 in the distributed data grid 1 100 can report the GUIDs 1 141 -1 143 (which can include the partition numbers and other information) for each of the persisted partitions that are found. In accordance with an embodiment of the invention, only members reporting the presence of the most recent GUID for a partition can be considered for recovery.
[0081] As shown in Figure 1 1 , each member 1 101-1 102 in the distributed data grid 1 100 can collect a list of available GUIDs 1 121 -1 122 from the distributed local disks 1 1 10 based on local visibility. Then, each member 1 101 -1 102 can provide (or register) the list of available GUIDs 1 121 -1 122 to a resolver 1 103 in the distributed data grid 1 100, and the resolver 1 103 can determine the newest GUIDs 1 130 for different partitions based on the partition number and version number information encoded in the GUIDs 1 141 -1 143.
[0082] Furthermore, due to the distributed nature of the system, the distributed local disks 1 1 10 may contain multiple different versions of the same partition. In other words, the resolver 1 103 may receive multiple GUIDs that contain the same partition number and different version numbers.
[0083] In such a case, the resolver 1 103 can obtain the version number from each GUID associated with the same partition, and determine which GUID has the most recent version number. Also, the distributed data grid 1 100 can ensure that the persisted partition with the most recent version number is valid based on performing the seal operation and validation operation.
[0084] Additionally, the resolver 1 103 can determine which member 1 101-1 102 in the distributed data grid 1 100 is responsible for recovering a particular persisted partition 1 1 1 1-1 1 13, based on the member ID information encoded in the GUIDs 1 141 -1 143.
[0085] Then, the resolver 1 103 can provide the partition recovery assignment, which may include a list of the newest GUIDs 1 131 -1 132, to each different member 1 101-1 102. Accordingly, the members 1 101 -1 102 can carry out the actual operation that restores the persisted partitions 1 1 1 1 -1 1 13.
[0086] Thus, the system can ensure that the distributed data grid 1 100 always restores the newest valid version of any persisted partition, and can avoid recovering a partial copy that may be caused by cascading cluster failures.
[0087] Figure 12 illustrates an exemplary flow chart for supporting persistent store versioning and integrity and in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 12, at step 1201 , the system can receive a plurality of identifiers (e.g. the GUIDs) from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid. Then, at step 1202, the system can select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition. Furthermore, at step 1203, the system can determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
Persistent Snapshot of a Running System
[0088] Figure 13 shows an illustration of providing a persistent snapshot of a running system in a distributed data grid, in accordance with an embodiment of the invention. As shown in Figure 13, a distributed data grid 1300 can support various cache services 1320 using an in-memory data store 1302. [0089] Furthermore, the system allows a user to use a management tool 1310 to take a snapshot 1301 of the running system on the in-memory data store 1302 that supports the cache services 1320 on-demand, at any particular time. For example, the snapshot 1301 can be used to make a backup of the running system overnight.
[0090] In accordance with an embodiment of the invention, the system can suspend the cache services 1320, prior to taking the snapshot 1301. Thus, the system can provide a consistent point in time for taking the snapshot 1301. Then, the cache service 1320 can be resumed after the snapshot 1301 is taken.
[0091] Additionally, the snapshot 1301 can provide a consistent view of each partitioned cache service 1320. For example, the snapshot 1301 can provide a catalogue of state information of the running system, including metadata 1311 and cache data 1312 for the partitioned cache services 1320. Additionally, the system can store the snapshot 1301 either in a central location (e.g. a SAN 1321 ) or in distributed local disks 1322.
[0092] Furthermore, when various artifacts in a snapshot 1301 are created and stored in the distributed local disks 1322, the system can use a pluggable (or portable) archiver 1303 to retrieve the persisted state information of the snapshot 1301 from the distributed local disks 1322, and can create a single archive unit 1330, which can be used for auditing or other purposes.
[0093] Thus, the system allows a user to take a snapshot on the state of a partitioned cache service in a distributed data grid 1300, instead of persisting the cache content in the distributed data grid 1300 in a continuing fashion.
[0094] Figure 14 illustrates an exemplary flow chart for providing a persistent snapshot of a running system in a distributed data grid in accordance with an embodiment of the invention. As shown in Figure 14, at step 1401 , the system allows one or more cache services to run on a plurality of cluster members in the distributed data grid. Then, at step 1402, the system can collect a catalogue of state information associated with said one or more cache services from the plurality of cluster members in the distributed data grid. Furthermore, at step 1403, the system can create a snapshot for said one or more cache services running on the distributed data grid. Resolver
[0095] Figure 15 shows an exemplary block graph illustrating the resolver in accordance with an embodiment of the invention.
[0096] The blocks of the resolver 1500 may be implemented by hardware, software, or a combination of hardware and software to carry out the principles of the invention. It is understood by persons of skill in the art that the blocks described in Figure 15 may be combined or separated into sub-blocks to implement the principles of the invention as described above. Therefore, the description herein may support any possible combination or separation orfurther definition of the functional blocks described herein.
[0097] In Figure 15, there is shown a resolver (1500) that may be used for a distributed data grid, such as any distributed data grid described previously, in particular, distributed data grid shown in Figure 1 1. Therefore, the resolver 1500 may be any resolver mentioned in the embodiments described previously. Moreover, the resolver 1500 and the component therein described below may perform various operations described previously according to the principle of the invention, but not limited to the operations and functions described below.
[0098] As shown, the resolver 1500 may comprise a receiving unit 1501 which may be configured to receive a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid. The resolver 1500 may further comprise a selecting unit 1502 which may be configured to select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition. The resolver 1500 may further comprise a determining unit 1503 which may be configured to determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
[0099] The resolver 1500 may further comprise a resolving unit 1504 which may be configured to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
[00100] As described previously, the system for supporting persistence in a distributed data grid according to the present invention may comprise a persistent storage, configured to store one or more persisted partitions, each of which is associated with an identifier; one or more members, configured to collect a list of available identifiers from the persistent storage; and a resolver, configured to receive the list of available identifiers from one or more members, select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition, and determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier. The resolver herein may be embodied as the resolver 1500 as shown in Figure 15.
[00101] In one embodiment, the persistent storage may comprise a plurality of distributed local disk, wherein each member in the distributed data grid only has visibility to one or more distribute local disks.
[00102] In one embodiment, an identifier is assigned to each persisted partition stored in the persistent storage, wherein each said identifier is associated with a partition number and a version number of a partition.
[00103] In one embodiment, the resolver may be further configured to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
[00104] In one embodiment, the persistent storage conprises a storage area network (SAN), wherein the SAN is visible to a plurality of members in the distributed data grid.
[00105] In one embodiment, different members in the distributed data grid operate to persist multiple versions of a partition in the persistent storage.
[00106] In one embodiment, the distributed data grid operates to seal a persistent store in the persistent storage, and indicate that the sealed persistent store is fully initialized and eligible to be recovered, and validate a persistent store in the persistent storage to determine whether said persistent store is sealed.
[00107] In one embodiment, said determined member operate to recover said partition using a persisted partition with a selected identifier.
[00108] Referring to Figure 16, a system 1600 is illustrated in accordance with an embodiment of the invention. Figure 16 shows an illustration of a functional configuration realized by system 1600. In accordance with an embodiment of the disclosure, system 1600 includes a persistent storage 1610, one or more members 1620, a receiver module 1630, a selector 1640, and a determination module 1650.
[00109] Receiver module receives a plurality of identifiers from one or more members 1620 of a distributed data grid. Each identifier is associated with a persisted partition in persistent storage 1610 for the distributed data grid. Selector 1640 selects an identifier for each partition. Each selected identifier is associated with a most recent valid version of a partition. Determination module 1650 determines a member in the distributed data grid that is responsible for recovering the partition from a persisted partition associated with the selected identifier.
[00110] Figure 17 shows an illustration of a computer system 1700 which includes well- known hardware elements. Namely, computer system 1700 includes a central processing unit (CPU) 1710, a mouse 1720, a key board 1730, a random access memory (RAM) 1740, a hard disc 1750, a disc drive 1760, a communication interface (l/F) 1770, and a monitor 1780. Computer system 1700 may function as a server node constituting system 1600.
[00111] In accordance with an embodiment of the invention, persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 are provided by one or more computer systems 1700. Persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 are implemented by CPU 1710. In a further aspect, more than one processors can be used so that persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 are implemented. Namely, any of persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650 can be physically remote from each other.
[00112] In yet another aspect, system 1600 can be realized by using a plurality of hardwired circuits which function as persistent storage 1610, one or more members 1620, receiver module 1630, selector 1640 and determination module 1650.
[00113] The present invention may be conveniently implemented using one or more conventional general purpose or specialized digital computer, computing device, machine, or microprocessor, including one or more processors, memory and/or computer readable storage media programmed according to the teachings of the present disclosure. Appropriate software coding can readily be prepared by skilled programmers based on the teachings of the present disclosure, as will be apparent to those skilled in the software art.
[00114] In some embodiments, the present invention includes a computer program product which is a storage medium or computer readable medium (media) having instructions stored thereon/in which can be used to program a computer to perform any of the processes of the present invention. The storage medium can include, but is not limited to, any type of disk including floppy disks, optical discs, DVD, CD-ROMs, microdrive, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, DRAMs, VRAMs, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.
[00115] The foregoing description of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to the practitioner skilled in the art. The modification and variation include any relevant combination of the described features. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalence.

Claims

Claims:
What is claimed is: 1. A method for supporting persistence in a distributed data grid, comprising:
receiving a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid;
selecting an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition; and
determining a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
2. The method according to Claim 1 , further comprising:
including a plurality of distributed local disks in the persistent storage, wherein each member in the distributed data grid only has visibility to one or more of the distributed local disks.
3. The method according to Claim 1 or 2, further comprising:
assigning an identifier to each persisted partition in the persistent storage, wherein each said identifier is associated with a partition number and a version number of a partition.
4. The method according to Claim 3, further comprising:
associating a timestamp with each said identifier. 5. The method according to any preceding Claim, further comprising:
resolving each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
6. The method according to any preceding Claim, further comprising:
including a storage area network (SAN) in the persistent storage, wherein the SAN is visible to a plurality of members in the distributed data grid.
7. The method according to any preceding Claim, further comprising:
allowing different members in the distributed data grid to persist multiple versions of a partition in the persistent storage.
8. The method according to any preceding Claim, further comprising:
sealing a persistent store in the persistent storage, and indicating that the sealed persistent store is fully initialized and eligible to be recovered. 9. The method according to any preceding Claim, further comprising:
validating a persistent store in the persistent storage to determine whether said persistent store is sealed.
10. The method according to any preceding Claim, further comprising:
recovering, via said determined member, said partition using a persisted partition with a selected identifier.
1 1 . A computer program comprising program instructions in machine-readable form that when executed by a computer system cause the computer system to perform the method of any preceding Claim.
12. A computer program product comprising the computer program according to Claim 1 1 stored on a machine readable storage medium. 13. A non-transitory machine readable storage medium having instructions stored thereon that when executed cause a system to perform the steps of the method according to any of Claims 1 to 10.
14. A system for supporting asynchronous message processing in a distributed data grid, comprising:
one or more microprocessors;
a distributed data grid running on the one or more microprocessors, wherein the distributed data grid includes a plurality of server nodes that are interconnected with one or more communication channels, and wherein the distributed data grid operates to perform the steps comprising:
receiving a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid;
selecting an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition; and
determining a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
15. The system according to Claim 14, wherein:
a plurality of distributed local disks are included in the persistent storage, wherein each member in the distributed data grid only has visibility to one or more of the distributed local disks.
16. The system according to Claim 14 or 15, wherein:
an identifier is assigned to each persisted partition in the persistent storage, wherein each said identifier is associated with a partition number and a version number of a partition.
17. The system according to Claim 16, wherein:
a timestamp is associated with each said identifier.
18. The system according to any of Claims 14 to 17, wherein:
a resolver operates to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
19. The system according to any of Claims 14 to 18, wherein:
a storage area network (SAN) is included in the persistent storage, wherein the SAN is visible to a plurality of members in the distributed data grid.
20. The system according to any of Claims 14 to 19, wherein:
different members in the distributed data grid operate to persist multiple versions of a partition in the persistent storage.
21 . The system according to any of Claims 14 to 20, wherein:
the distributed data grid operates to:
seal a persistent store in the persistent storage, and indicate that the sealed persistent store is fully initialized and eligible to be recovered, and
validate a persistent store in the persistent storage to determine whether said persistent store is sealed.
22. The system according to any of Claims 14 to 21 , wherein:
said determined member operate to recover said partition using a persisted partition with a selected identifier.
23. A resolver for supporting persistence in a distributed data grid, comprising:
receiving unit configured to receive a plurality of identifiers from one or more members of the distributed data grid, wherein each said identifier is associated with a persisted partition in a persistent storage for the distributed data grid;
selecting unit configured to select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition; and
determining unit configured to determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
24. The resolver according to Claim 23, further comprising:
resolving unit configured to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage. 25. A system for supporting persistence in a distributed data grid, comprising:
a persistent storage, configured to store one or more persisted partitions, each of which is associated with an identifier;
one or more members, configured to collect a list of available identifiers from the persistent storage;
a resolver, configured to receive the list of available identifiers from one or more members, select an identifier for each partition, wherein each selected identifier is associated with a most recent valid version of a partition, and determine a member in the distributed data grid that is responsible for recovering said partition from a persisted partition associated with the selected identifier.
26. The system according to Claim 25, wherein:
the persistent storage comprises a plurality of distributed local disk, wherein each member in the distributed data grid only has visibility to one or more distribute local disks. 27. The system according to Claim 25, wherein:
an identifier is assigned to each persisted partition stored in the persistent storage, wherein each said identifier is associated with a partition number and a version number of a partition.
The system according to Claim 27, wherein:
a timestamp is associated with each said identifier.
29. The system according to Claim 25, wherein:
the resolver is further configured to resolve each received identifier to obtain a partition number and a version number associated with each persisted partition in the persistent storage.
30. The system according to Claim 25, wherein:
the persistent storage conprises a storage area network (SAN), wherein the SAN is visible to a plurality of members in the distributed data grid.
31 . The system according to Claim 25, wherein:
different members in the distributed data grid operate to persist multiple versions of a partition in the persistent storage.
32. The system according to Claim 25, wherein:
the distributed data grid operates to
seal a persistent store in the persistent storage, and indicate that the sealed persistent store is fully initialized and eligible to be recovered, and
validate a persistent store in the persistent storage to determine whether said persistent store is sealed.
33. The system according to Claim 25, wherein:
said determined member operate to recover said partition using a persisted partition with a selected identifier.
PCT/US2014/068904 2013-12-13 2014-12-05 System and method for supporting persistent store versioning and integrity in a distributed data grid WO2015088918A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN201480067510.9A CN105830033B (en) 2013-12-13 2014-12-05 System and method for supporting persistent storage versioning and integrity in a distributed data grid
JP2016537499A JP6483699B2 (en) 2013-12-13 2014-12-05 System and method for supporting persistent store versioning and integrity in a distributed data grid
EP14819219.8A EP3080698A1 (en) 2013-12-13 2014-12-05 System and method for supporting persistent store versioning and integrity in a distributed data grid

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201361915912P 2013-12-13 2013-12-13
US61/915,912 2013-12-13
US14/271,150 2014-05-06
US14/271,150 US10176184B2 (en) 2012-01-17 2014-05-06 System and method for supporting persistent store versioning and integrity in a distributed data grid

Publications (1)

Publication Number Publication Date
WO2015088918A1 true WO2015088918A1 (en) 2015-06-18

Family

ID=53368673

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/US2014/068904 WO2015088918A1 (en) 2013-12-13 2014-12-05 System and method for supporting persistent store versioning and integrity in a distributed data grid
PCT/US2014/068898 WO2015088916A1 (en) 2013-12-13 2014-12-05 System and method for supporting persistence partition recovery in a distributed data grid

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/US2014/068898 WO2015088916A1 (en) 2013-12-13 2014-12-05 System and method for supporting persistence partition recovery in a distributed data grid

Country Status (5)

Country Link
US (4) US10176184B2 (en)
EP (2) EP3080698A1 (en)
JP (2) JP6483699B2 (en)
CN (2) CN105830033B (en)
WO (2) WO2015088918A1 (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10320703B2 (en) 2015-09-30 2019-06-11 Veritas Technologies Llc Preventing data corruption due to pre-existing split brain
US10191817B2 (en) * 2015-12-28 2019-01-29 Veritas Technologies Llc Systems and methods for backing up large distributed scale-out data systems
JP2019518065A (en) 2016-06-14 2019-06-27 ピュアサークル ユーエスエー インコーポレイテッド Steviol glycoside composition, method of preparation and use
CN107885671B (en) 2016-09-30 2021-09-14 华为技术有限公司 Nonvolatile memory persistence method and computing device
US11550820B2 (en) * 2017-04-28 2023-01-10 Oracle International Corporation System and method for partition-scoped snapshot creation in a distributed data computing environment
US10769019B2 (en) * 2017-07-19 2020-09-08 Oracle International Corporation System and method for data recovery in a distributed data computing environment implementing active persistence
CN110764940A (en) * 2018-07-26 2020-02-07 北京国双科技有限公司 Processing method and device for service exception of distributed system
US11100086B2 (en) * 2018-09-25 2021-08-24 Wandisco, Inc. Methods, devices and systems for real-time checking of data consistency in a distributed heterogenous storage system
CN111352878B (en) * 2018-12-21 2021-08-27 达发科技(苏州)有限公司 Digital signal processing system and method
CN109947375B (en) * 2019-04-04 2021-05-14 江南大学 Distributed storage system optimization method based on partition processing consensus algorithm
CN110309128B (en) * 2019-07-05 2020-07-17 广东铭太信息科技有限公司 Oracle backup file automatic importing device, implementation method thereof and method for importing backup file by using device
CN110795605B (en) * 2020-01-03 2020-05-12 北京东方通科技股份有限公司 Data storage system based on distributed memory grid
US11438224B1 (en) 2022-01-14 2022-09-06 Bank Of America Corporation Systems and methods for synchronizing configurations across multiple computing clusters
US20240152429A1 (en) * 2022-11-04 2024-05-09 Microsoft Technology Licensing, Llc Recoverable Processes
CN116361389B (en) * 2023-03-17 2024-03-08 国网江苏省电力有限公司营销服务中心 Data synchronization link method and system based on national network marketing acquisition system
CN117827699B (en) * 2023-12-27 2024-09-24 中电云计算技术有限公司 Parallel read cache persistence method, device, equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120254118A1 (en) * 2011-03-31 2012-10-04 Microsoft Corporation Recovery of tenant data across tenant moves
US20130047165A1 (en) * 2011-08-15 2013-02-21 Sap Ag Context-Aware Request Dispatching in Clustered Environments

Family Cites Families (115)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5819272A (en) 1996-07-12 1998-10-06 Microsoft Corporation Record tracking in database replication
US5784569A (en) 1996-09-23 1998-07-21 Silicon Graphics, Inc. Guaranteed bandwidth allocation method in a computer system for input/output data transfers
US5940367A (en) 1996-11-06 1999-08-17 Pluris, Inc. Fault-tolerant butterfly switch
US6233601B1 (en) 1996-11-14 2001-05-15 Mitsubishi Electric Research Laboratories, Inc. Itinerary based agent mobility including mobility of executable code
US6125368A (en) 1997-02-28 2000-09-26 Oracle Corporation Fault-tolerant timestamp generation for multi-node parallel databases
US5933818A (en) 1997-06-02 1999-08-03 Electronic Data Systems Corporation Autonomous knowledge discovery system and method
US5991894A (en) 1997-06-06 1999-11-23 The Chinese University Of Hong Kong Progressive redundancy transmission
US5999712A (en) 1997-10-21 1999-12-07 Sun Microsystems, Inc. Determining cluster membership in a distributed computer system
US6605120B1 (en) 1998-12-10 2003-08-12 International Business Machines Corporation Filter definition for distribution mechanism for filtering, formatting and reuse of web based content
US6453426B1 (en) 1999-03-26 2002-09-17 Microsoft Corporation Separately storing core boot data and cluster configuration data in a server cluster
US6693874B1 (en) 1999-05-26 2004-02-17 Siemens Information & Communication Networks, Inc. System and method for enabling fault tolerant H.323 systems
US6871222B1 (en) 1999-05-28 2005-03-22 Oracle International Corporation Quorumless cluster using disk-based messaging
US7020695B1 (en) 1999-05-28 2006-03-28 Oracle International Corporation Using a cluster-wide shared repository to provide the latest consistent definition of the cluster (avoiding the partition-in time problem)
US6553389B1 (en) 1999-09-13 2003-04-22 Hewlett-Packard Company Resource availability determination mechanism for distributed data storage system
US6487622B1 (en) 1999-10-28 2002-11-26 Ncr Corporation Quorum arbitrator for a high availability system
WO2001082678A2 (en) 2000-05-02 2001-11-08 Sun Microsystems, Inc. Cluster membership monitor
US20020035559A1 (en) 2000-06-26 2002-03-21 Crowe William L. System and method for a decision engine and architecture for providing high-performance data querying operations
US6915391B2 (en) 2000-12-15 2005-07-05 International Business Machines Corporation Support for single-node quorum in a two-node nodeset for a shared disk parallel file system
JP4637382B2 (en) * 2001-02-13 2011-02-23 サイボウズ株式会社 Data backup system
US7792977B1 (en) 2001-02-28 2010-09-07 Oracle International Corporation Method for fencing shared resources from cluster nodes
US20040179471A1 (en) 2001-03-07 2004-09-16 Adisak Mekkittikul Bi-directional flow-switched ring
US20020169738A1 (en) 2001-05-10 2002-11-14 Giel Peter Van Method and system for auditing an enterprise configuration
US7113980B2 (en) 2001-09-06 2006-09-26 Bea Systems, Inc. Exactly once JMS communication
US7376953B2 (en) 2001-10-29 2008-05-20 Hewlett-Packard Development Company, L.P. Apparatus and method for routing a transaction to a server
US6904448B2 (en) 2001-12-20 2005-06-07 International Business Machines Corporation Dynamic quorum adjustment
US20030187927A1 (en) 2002-02-22 2003-10-02 Winchell David F. Clustering infrastructure system and method
US7139925B2 (en) 2002-04-29 2006-11-21 Sun Microsystems, Inc. System and method for dynamic cluster adjustment to node failures in a distributed data system
US6952758B2 (en) 2002-07-31 2005-10-04 International Business Machines Corporation Method and system for providing consistent data modification information to clients in a storage system
US7206836B2 (en) 2002-09-23 2007-04-17 Sun Microsystems, Inc. System and method for reforming a distributed data system cluster after temporary node failures or restarts
US20040153558A1 (en) 2002-10-31 2004-08-05 Mesut Gunduc System and method for providing java based high availability clustering framework
US7451359B1 (en) 2002-11-27 2008-11-11 Oracle International Corp. Heartbeat mechanism for cluster systems
US7127577B2 (en) * 2003-01-21 2006-10-24 Equallogic Inc. Distributed snapshot process
KR100553920B1 (en) 2003-02-13 2006-02-24 인터내셔널 비지네스 머신즈 코포레이션 Method for operating a computer cluster
US7376754B2 (en) 2003-02-27 2008-05-20 Bea Systems, Inc. System and method for communications between servers in a cluster
US20040176968A1 (en) 2003-03-07 2004-09-09 Microsoft Corporation Systems and methods for dynamically configuring business processes
US7958026B2 (en) 2003-04-29 2011-06-07 Oracle International Corporation Hierarchical transaction filtering
US20050021737A1 (en) 2003-05-01 2005-01-27 Ellison Carl M. Liveness protocol
US20040267897A1 (en) 2003-06-24 2004-12-30 Sychron Inc. Distributed System Providing Scalable Methodology for Real-Time Control of Server Pools and Data Centers
JP5068000B2 (en) 2003-07-31 2012-11-07 富士通株式会社 Information processing method and program in XML driven architecture
US8234517B2 (en) * 2003-08-01 2012-07-31 Oracle International Corporation Parallel recovery by non-failed nodes
US7551552B2 (en) 2003-10-17 2009-06-23 Microsoft Corporation Method for providing guaranteed distributed failure notification
US7260698B2 (en) 2003-10-30 2007-08-21 International Business Machines Corporation Method and system for page initialization using off-level worker thread
US7464378B1 (en) 2003-12-04 2008-12-09 Symantec Operating Corporation System and method for allowing multiple sub-clusters to survive a cluster partition
US7779386B2 (en) 2003-12-08 2010-08-17 Ebay Inc. Method and system to automatically regenerate software code
US7299378B2 (en) 2004-01-15 2007-11-20 Oracle International Corporation Geographically distributed clusters
US7712077B2 (en) 2004-02-27 2010-05-04 International Business Machines Corporation Method and system for instantiating components conforming to the “COM” specification in custom contexts
US7428733B2 (en) 2004-05-13 2008-09-23 Bea Systems, Inc. System and method for custom module creation and deployment
US7386753B2 (en) 2004-09-02 2008-06-10 International Business Machines Corporation Subscription-based management and distribution of member-specific state data in a distributed computing system
US7640339B1 (en) 2005-02-14 2009-12-29 Sun Microsystems, Inc. Method and apparatus for monitoring a node in a distributed system
US7530059B2 (en) 2005-02-18 2009-05-05 International Business Machines Corporation Method for inlining native functions into compiled java code
US7613774B1 (en) 2005-03-01 2009-11-03 Sun Microsystems, Inc. Chaperones in a distributed system
US7979457B1 (en) 2005-03-02 2011-07-12 Kayak Software Corporation Efficient search of supplier servers based on stored search results
US7698390B1 (en) 2005-03-29 2010-04-13 Oracle America, Inc. Pluggable device specific components and interfaces supported by cluster devices and systems and methods for implementing the same
US7739677B1 (en) 2005-05-27 2010-06-15 Symantec Operating Corporation System and method to prevent data corruption due to split brain in shared data clusters
US7870230B2 (en) 2005-07-15 2011-01-11 International Business Machines Corporation Policy-based cluster quorum determination
US7720971B2 (en) 2005-09-12 2010-05-18 Microsoft Corporation Arbitrating an appropriate back-end server to receive channels of a client session
US20070118693A1 (en) 2005-11-19 2007-05-24 International Business Machines Cor Method, apparatus and computer program product for cache restoration in a storage system
US7627584B2 (en) 2005-11-30 2009-12-01 Oracle International Corporation Database system configured for automatic failover with no data loss
US7882079B2 (en) 2005-11-30 2011-02-01 Oracle International Corporation Database system configured for automatic failover with user-limited data loss
US7756924B2 (en) 2005-12-21 2010-07-13 Microsoft Corporation Peer communities
JP2007219609A (en) 2006-02-14 2007-08-30 Hitachi Ltd Snapshot management device and method
WO2007110094A1 (en) 2006-03-27 2007-10-04 Telecom Italia S.P.A. System for enforcing security policies on mobile communications devices
US7676628B1 (en) * 2006-03-31 2010-03-09 Emc Corporation Methods, systems, and computer program products for providing access to shared storage by computing grids and clusters with large numbers of nodes
US8570857B2 (en) 2006-04-07 2013-10-29 At&T Intellectual Property I, Lp Resilient IP ring protocol and architecture
US7975288B2 (en) 2006-05-02 2011-07-05 Oracle International Corporation Method and apparatus for imposing quorum-based access control in a computer system
US20070271584A1 (en) 2006-05-16 2007-11-22 Microsoft Corporation System for submitting and processing content including content for on-line media console
US7953861B2 (en) 2006-08-10 2011-05-31 International Business Machines Corporation Managing session state for web applications
US8775402B2 (en) 2006-08-15 2014-07-08 Georgia State University Research Foundation, Inc. Trusted query network systems and methods
US20080077622A1 (en) 2006-09-22 2008-03-27 Keith Robert O Method of and apparatus for managing data utilizing configurable policies and schedules
US7814248B2 (en) 2006-12-07 2010-10-12 Integrated Device Technology, Inc. Common access ring/sub-ring system
US9111276B2 (en) 2006-12-08 2015-08-18 Sap Se Secure execution environments for process models
US8104080B2 (en) 2007-01-26 2012-01-24 Microsoft Corporation Universal schema for representing management policy
US9026655B2 (en) 2007-01-31 2015-05-05 Oracle America, Inc. Method and system for load balancing
JP5036041B2 (en) 2007-04-25 2012-09-26 アズビル株式会社 RSTP processing method
US8745584B2 (en) 2007-05-03 2014-06-03 International Business Machines Corporation Dependency injection by static code generation
US20080281959A1 (en) 2007-05-10 2008-11-13 Alan Robertson Managing addition and removal of nodes in a network
WO2009068952A2 (en) 2007-11-30 2009-06-04 Telefonaktiebolaget L M Ericsson (Publ) Method, network, and node for distributing electronic content in a content distribution network
US8397227B2 (en) 2007-12-04 2013-03-12 International Business Machines Corporation Automatic deployment of Java classes using byte code instrumentation
US8306951B2 (en) * 2009-09-18 2012-11-06 Oracle International Corporation Automated integrated high availability of the in-memory database cache and the backend enterprise database
US8401994B2 (en) 2009-09-18 2013-03-19 Oracle International Corporation Distributed consistent grid of in-memory database caches
US20090228321A1 (en) 2008-03-04 2009-09-10 Oracle International Corporation Accessing an Enterprise Calendar and Scheduling Group Meetings Using a Mobile Device
US7990850B2 (en) 2008-04-11 2011-08-02 Extreme Networks, Inc. Redundant Ethernet automatic protection switching access to virtual private LAN services
US20090265449A1 (en) 2008-04-22 2009-10-22 Hewlett-Packard Development Company, L.P. Method of Computer Clustering
US7543046B1 (en) 2008-05-30 2009-06-02 International Business Machines Corporation Method for managing cluster node-specific quorum roles
US8719803B2 (en) 2008-06-04 2014-05-06 Microsoft Corporation Controlling parallelization of recursion using pluggable policies
US8375001B2 (en) * 2008-10-03 2013-02-12 Telefonaktiebolaget Lm Ericsson (Publ) Master monitoring mechanism for a geographical distributed database
JP5425448B2 (en) * 2008-11-27 2014-02-26 インターナショナル・ビジネス・マシーンズ・コーポレーション Database system, server, update method and program
US8402464B2 (en) 2008-12-01 2013-03-19 Oracle America, Inc. System and method for managing contention in transactional memory using global execution data
US7917596B2 (en) 2009-01-07 2011-03-29 Oracle International Corporation Super master
US8595714B1 (en) 2009-03-04 2013-11-26 Amazon Technologies, Inc. User controlled environment updates in server cluster
US8626552B2 (en) 2009-03-26 2014-01-07 International Business Machines Corporation Quorum management of appointment scheduling
US8209307B2 (en) 2009-03-31 2012-06-26 Commvault Systems, Inc. Systems and methods for data migration in a clustered file system
US20100268571A1 (en) 2009-04-16 2010-10-21 Mitel Networks Corporation System and method for determining availibility of a group to communicate with a user
GB2472620B (en) * 2009-08-12 2016-05-18 Cloudtran Inc Distributed transaction processing
CN101997823B (en) * 2009-08-17 2013-10-02 联想(北京)有限公司 Distributed file system and data access method thereof
US8108734B2 (en) 2009-11-02 2012-01-31 International Business Machines Corporation Intelligent rolling upgrade for data storage systems
US8578038B2 (en) 2009-11-30 2013-11-05 Nokia Corporation Method and apparatus for providing access to social content
US9286369B2 (en) 2009-12-30 2016-03-15 Symantec Corporation Data replication across enterprise boundaries
US9135268B2 (en) * 2009-12-30 2015-09-15 Symantec Corporation Locating the latest version of replicated data files
US8417899B2 (en) 2010-01-21 2013-04-09 Oracle America, Inc. System and method for controlling access to shared storage device
US8725951B2 (en) 2010-04-12 2014-05-13 Sandisk Enterprise Ip Llc Efficient flash memory-based object store
JP5691306B2 (en) * 2010-09-03 2015-04-01 日本電気株式会社 Information processing system
US8600944B2 (en) * 2010-09-24 2013-12-03 Hitachi Data Systems Corporation System and method for managing integrity in a distributed database
US8639758B2 (en) 2010-11-09 2014-01-28 Genesys Telecommunications Laboratories, Inc. System for determining presence of and authorizing a quorum to transact business over a network
US9558256B2 (en) 2010-11-16 2017-01-31 Linkedin Corporation Middleware data log system
US20120158650A1 (en) 2010-12-16 2012-06-21 Sybase, Inc. Distributed data cache database architecture
US9355145B2 (en) 2011-01-25 2016-05-31 Hewlett Packard Enterprise Development Lp User defined function classification in analytical data processing systems
US9262229B2 (en) 2011-01-28 2016-02-16 Oracle International Corporation System and method for supporting service level quorum in a data grid cluster
US9703610B2 (en) 2011-05-16 2017-07-11 Oracle International Corporation Extensible centralized dynamic resource distribution in a clustered data grid
WO2013018808A1 (en) 2011-08-02 2013-02-07 日本電気株式会社 Distributed storage system and method
US8868546B2 (en) * 2011-09-15 2014-10-21 Oracle International Corporation Query explain plan in a distributed data management system
US9621409B2 (en) * 2011-09-15 2017-04-11 Oracle International Corporation System and method for handling storage events in a distributed data grid
US20150058293A1 (en) 2012-03-22 2015-02-26 Nec Corporation Distributed storage system, storage control method and program
US9311014B2 (en) 2012-11-29 2016-04-12 Infinidat Ltd. Storage system and methods of mapping addresses of snapshot families
US20140278573A1 (en) * 2013-03-15 2014-09-18 State Farm Mutual Automobile Insurance Company Systems and methods for initiating insurance processing using ingested data

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120254118A1 (en) * 2011-03-31 2012-10-04 Microsoft Corporation Recovery of tenant data across tenant moves
US20130047165A1 (en) * 2011-08-15 2013-02-21 Sap Ag Context-Aware Request Dispatching in Clustered Environments

Also Published As

Publication number Publication date
US20150169718A1 (en) 2015-06-18
US20150169598A1 (en) 2015-06-18
JP2017504880A (en) 2017-02-09
EP3080698A1 (en) 2016-10-19
US20150169653A1 (en) 2015-06-18
CN105830033B (en) 2020-03-24
JP2016540312A (en) 2016-12-22
CN105814544B (en) 2020-03-24
US10817478B2 (en) 2020-10-27
EP3080697A1 (en) 2016-10-19
US10706021B2 (en) 2020-07-07
WO2015088916A1 (en) 2015-06-18
JP6483699B2 (en) 2019-03-13
JP6491210B2 (en) 2019-03-27
CN105814544A (en) 2016-07-27
US10176184B2 (en) 2019-01-08
CN105830033A (en) 2016-08-03
US20190121790A1 (en) 2019-04-25

Similar Documents

Publication Publication Date Title
US10817478B2 (en) System and method for supporting persistent store versioning and integrity in a distributed data grid
US20220147495A1 (en) Virtualized file server
US8954391B2 (en) System and method for supporting transient partition consistency in a distributed data grid
US9703853B2 (en) System and method for supporting partition level journaling for synchronizing data in a distributed data grid
US8856091B2 (en) Method and apparatus for sequencing transactions globally in distributed database cluster
US20070061379A1 (en) Method and apparatus for sequencing transactions globally in a distributed database cluster
EP2643771B1 (en) Real time database system
US20170003899A1 (en) System and method for distributed persistent store archival and retrieval in a distributed computing environment
US11550820B2 (en) System and method for partition-scoped snapshot creation in a distributed data computing environment
US9396076B2 (en) Centralized version control system having high availability
WO2007028249A1 (en) Method and apparatus for sequencing transactions globally in a distributed database cluster with collision monitoring

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14819219

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2016537499

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2014819219

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2014819219

Country of ref document: EP