IE83771B1 - System and method for managing storage resources in a clustered computing environment - Google Patents
System and method for managing storage resources in a clustered computing environment Download PDFInfo
- Publication number
- IE83771B1 IE83771B1 IE2001/0108A IE20010108A IE83771B1 IE 83771 B1 IE83771 B1 IE 83771B1 IE 2001/0108 A IE2001/0108 A IE 2001/0108A IE 20010108 A IE20010108 A IE 20010108A IE 83771 B1 IE83771 B1 IE 83771B1
- Authority
- IE
- Ireland
- Prior art keywords
- node
- computing environment
- clustered computing
- log out
- process log
- Prior art date
Links
- 238000000034 method Methods 0.000 claims description 26
- 239000000835 fiber Substances 0.000 claims description 19
- 230000000977 initiatory Effects 0.000 claims description 2
- 239000003795 chemical substances by application Substances 0.000 claims 1
- 239000004744 fabric Substances 0.000 description 9
- 230000001627 detrimental Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000000873 masking Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000000605 extraction Methods 0.000 description 2
- 238000000638 solvent extraction Methods 0.000 description 2
- 241000136406 Comones Species 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000004059 degradation Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 230000000737 periodic Effects 0.000 description 1
- 230000002085 persistent Effects 0.000 description 1
- 230000003334 potential Effects 0.000 description 1
- 238000004642 transportation engineering Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/16—Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/0614—Improving the reliability of storage systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0638—Organizing or formatting or addressing of data
- G06F3/0644—Management of space entities, e.g. partitions, extents, pools
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/067—Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]
Description
PATENTS ACT, 1992
2001/0108
SYSTEM AND METHOD FOR MANAGING STORAGE RESOURCES IN
A CLUSTERED COMPUTING ENVIRONMENT
DELL PRODUCTS L.P.
SYSTEM AND METHOD FOR MANAGING STORAGE RESOURCES IN A
CLUSTERED COMPUTING ENVIRONMENT
The present disclosure relates in general to the
field of data storage systems and, more particularly, to
a system and method for managing storage resources in a
clustered comuting environment.
Storage area networks (SANs) often include a
collection of data storage resources communicatively
coupled to a plurality of nodes such as workstations and
servers. In the present disclosure, the term "node" and
"server" are used interchangeably, with the understanding
that a "server" is one type of "node".
Within a SAN, a server may access a data storage
resource across a fabric using the Fibre Channel
protocol. The Fibre Channel protocol may act as a comon
physical layer that allows for the transportation of
multiple upper layer protocols, such as the small
computer system interconnect (SCSI) protocol. In a SAN
environment, the SCSI protocol may assign logical unit
numbers (LUNs) to the collection of data storage
resources. The LUNs may allow a server within a SAN to
access specific data storage resources by referencing a '
SCSI LUN for a specific data storage resource.
Though a Fibre Channel storage system can offer a
great deal of storage capacity, the system can also be
very expensive to implement. As a result, users often
seek to share the available storage provided by the
system among multiple servers. Unfortunately, if a
server coupled to a given SAN uses the MICROSOFT WINDOWS
NT*operating system, the server may attempt to take
ownership of any LUN visible to the server. For example,
if a particular server detects several LUNs when the
server boots, it may assume each LUN is available for its
use. Therefore, if multiple WINDOWS NT*servers are
attached to a storage pool or a collection of data
* Trade Mark
storage resources, each server may attempt to take
control of each LUN in the storage pool. This situation
can lead to conflicts when more than one server attempts
to access the same LUN.
A user seeking to solve this problem may partition
or zone the available storage through filtering or
through the use of miniport drivers that have LUN masking
capabilities. In effect, this partitioning may prevent a
server running WINDOWS NT*from seeing storage capacity
that is not assigned to it. This approach may be
effective for stand-alone servers, but the approach has
several shortcomings in a clustered computing
environment.
Clustering involves the configuring of a group of
independent servers so that they appear on a network as a
single machine. often, clusters are managed as a single
system, share a common namespace, and are designed
specifically to tolerate component failures and to
support the addition or subtraction of components in a
transparent manner. Unfortunately, because a cluster may
have two or more servers that appear to be a single
machine, the partitioning techniques mentioned above may
prove an ineffective solution for avoiding conflicts when
the two or more servers attempt to access the same LUN.
MICROSOFT CLUSTER SERVER (MSCs)*embodies one
currently available technique for arbitrating conflicts
and managing ownership of storage devices in a clustered
computing environment. An MSCS system may operate within
a cluster that has two servers, server A, which may be in
charge, and server B.
In operation, server A may pass a
* Trade Mark
periodic heartbeat signal to server B to let server B
know that server A is "alive". If server B does not
receive a timely heartbeat from server A, server B may
seek to determine whether server A is operable and/or
whether server B may take ownership of any LUNs reserved
for server A. Unfortunately, these MSCS system may
utilize SCSI target resets during this process, and the
SCSI resets may create several problems. For example, a
typical SCSI reset in the MSCS system may cause all
servers within a given Fibre Channel system to abort
These
aborted I/O processes may eventually be completed but not
until the bus settles.
their pending input/output "I/0' processes.
This abort/wait/retry approach
can have a detrimental effect on overall system
performance.
In addition to this potential effect on performance,
the MSCS system and its use of SCSI resets may have a
detrimental effect on overall system reliability. In
operation, the MSCS system may only account for one SCSI
reset at a time. The inability to account for subsequent
SCSI resets may lead to unexpected behavior and decrease
system reliability.
In accordance with the present disclosure, a system
and method for managing storage resources in a clustered
computing environment are disclosed that provide
significant advantages over prior developed techniques.
The disclosed system and method may allow for storage
resource management and conflict arbitration with a
reduced reliance on SCSI resets.
According to one aspect of the present disclosure, a
method for managing storage resources in a clustered
computing environment includes holding a reservation
on a storage resource for a first node of the clustered
computing environment. The node may be, for example, a
server, a workstation, or any other computing device
_ included within the cluster.
A third party process log out for the first node
is performed and the reservation held for the first
node is released. In one embodiment, the third party
process log out may occur in response to a log out
command sent on behalf of the first node. The third
party process log out command may be sent, for example,
by a second node or a Fibre Channel switch. The third
party process log out command may include identification
information that identifies the first node as the sender
of the log out command even though the first node was not
the actual sender. The identification information may
include, for example, a world wide name and a source
identifier assigned to the first node.
Managing storage resources in a clustered computing
environment may additionally involve the zoning of a
Fibre Channel storage system. A zone may group a first
node with a second node and a plurality of storage
resources such as hard drives and other data storage
devices. In the zoned system, a second node may log
itself out after a third party process log out command
has been issued for a first node. After the two nodes
are logged out, a loop initialization protocol (LIP) link
reset may be initiated, a state change notification may
be generated, and any functioning nodes may re-login.
According to another aspect of the present disclosure, a computer
system operable to manage storage resources in a clustered computing
environment includes a first node for which a reservation is held on a
storage resource, and a resource management engine operable to
perform a third party process log out for the first node and release the
reservation held for the first node. The third party process log out may be
performed in response to a failure by a second node to receive a
timely heartbeat signal from the first node. The system
may also include a computer readable medium storing the
resource management engine and a central processing unit
communicatively coupled to the computer readable medium
and operable to execute the resource management engine.
In one embodiment, the system may also include a
plurality of computing platforms communicatively coupled
to the first node. These computing platforms may be, for
example, a collection of networked personal computers.
The system may also include a Fibre Channel switch
communicatively coupled to the first node and to a
plurality of storage resources. The Fibre Channel switch
may, in some embodiments, include a central processing
unit operable to execute a resource management engine.
A system and method incorporating teachings of the
present disclosure may provide significant improvements
over conventional cluster resource management solutions.
For example, the disclosed techniques may be operable to
better manage and arbitrate storage resource conflicts.
As discussed above, a-SCSI reset in a clustered computing
environment can result in the initiation of an
abort/wait/retry approach to several I/O processes, which
can have a detrimental effect on overall system
performance. The teachings of the present disclosure may
help reduce reliance on SCSI resets and the resulting
performance degradations.
In addition, the teachings of the present disclosure
may facilitate the avoidance of system reliability
problems associated with SCSI resets in a clusteredi
computing environment. A conventional cluster resource
management system, such as MSCS, may be unable to account
for SCSI resets initiated during the bus disturbance of
an earlier SCSI reset. This limitation may lead to
unexpected behavior and decrease system reliability.
Because the teachings of the present disclosure may
facilitate the avoidance of at least some SCSI resets,
system reliability may be improved.
other technical advantages should be apparent to one
of ordinary skill in the art in view of the
specification, claims, and drawings,
The present invention will be described, by way of example, with
reference to the accompanying drawings, in which:
FIGURE 1 depicts a component diagram of a storage area network
including one embodiment of a resource management engine that
incorporates teachings of the present disclosure; and
FIGURE 2 shows a flow diagram for one embodiment of a method for
managing storage resources in a clustered computing environment in
accordance with teachings of the present disclosure.
FIGURE 1 depicts a general block diagram of a
storage area network (SAN), indicated generally at 10.
SAN 10 includes two clustered computing systems, clusters
12 and 14. As depicted, cluster 12 includes node 16 and
node 18, and cluster 14 includes node 20 and 22. Nodes
16, 18, 20, and 22 may be, for example, servers,
workstations, or other network computing devices. As
depicted in FIGURE 1,
number of client devices such as the client personal
cluster 12 may be supporting a
computers representatively depicted at 24.
SAN 10 may also include a storage pool 26, which may
include, for example, a plurality of physical storage
devices such as hard disk drives under the control of and
coupled to one or more storage controllers. The physical
storage devices of storage pool 26 may be assigned LUNB.
some physical storage devices may be grouped into RAID
volumes with each volume assigned a single SCSI LUN
address. Other physical storage devices may be
However the LUNs
are assigned, the LUNs of FIGURE 1 may map the available
individually assigned one or more LUNs.
physical storage of storage pool 26 into a plurality of
logical storage devices and allow these logical storage
devices to be identified and addressed.
In operation, nodes 16, 18, 20, and 22 may
communicate with and transfer data to and from storage
pool 26 through fabric 28 using fibre channel protocol.
As depicted in FIGURE 1, nodes 16 and 18 may be grouped
into zone 30 with LUN_1 and LUN_2. Similarly, nodes 20
and 22 may be grouped into zone 32 with LUN_3, LUN;4, and
LUN_5. Using switch zoning to create zone 30 may prevent
nodes 16 and 18 from seeing nodes 20 and 22. Similarly,
using switch zoning to create zone 32 may prevent nodes
and 22 from seeing nodes 16 and 18. In addition to
zoning, the embodiment of FIGURE 1 may employ LUN
masking. LUN masking may blind a specific node or
cluster from seeing certain LUNs. For example, LUN
masking may prevent nodes 16 and 18 from seeing LUN_3,
LUN_4, and LUN_5 .
In the embodiment of FIGURE 1, nodes 16, 18, 20, and
22 may be assigned a unique world wide name (WWN), which
may be an eight byte identifier. The Institute of
Electronics Engineers (IEEB) assigns blocks of WWNs to
manufacturers so manufacturers can build fiber channel
devices with unique WWNs. For illustrative purposes, in
the embodiment of FIGURE 1, node 16 may have a WWN of
"AAA", node 18 may have a WWN of "BBB", node 20 may haye
a WWN of "CCC", and node 22 may have a WWN of "DDD". As
such, nodes 16, 18, 20, and 22 may be uniquely
identifiable by other devices coupled to fabric 28.
Nodes 16, 18, 20, and 22 may have identification
information in addition to their respective WWNs. For
example, according to the fibre channel protocol, when a
node such as node 16 is initialized and logs into fabric
28, the node is assigned a fibre channel ID. This ID may
be subject to change each time some initialization event
occurs, for example, when another node or device logs
into fabric 28. As depicted in FIGURE 1, fabric 28 has
assigned fibre channel IDs as follows: node 16 is S_ID_1,
node 18 is S_ID_2, node 20 is S_ID_3, and node 22 is
S_ID_4.
In the embodiment of FIGURE 1, the various WWNs and
fibre channel IDs may be stored in a computer readable
medium 34, which may be accessible to devices of SAN 10.
As shown in FIGURE 1, SAN 10 may include a computing
device 38 for establishing fabric 28. Such a computing
device may include a CPU communicatively coupled to
computer readable medium 34. Switch 36 may also have at
least one port 40 for interfacing with other devices to
form an overall fibre channel network.
In one embodiment of a system incorporating
teachings of the present disclosure, computing device 38
may be operable to execute a resource management engine,
The
resource management engine may be operable to perform
which may be stored in computer readable medium 34.
several functions. For example, the resource management
engine may be operable to access a maintained list of the
WWNs and the fibre channel IDs of SAN 10 devices. In
addition, the resource management engine may be operable
to recognize a SCSI reset comand issued by a node and to
convert the command into a storage resource releasing
command. The storage resource releasing command may be,
for example, a third party process log out or a SCSI
persistent reserve out command with a clear action.
In a typical MSCS cluster, a SCSI reset command may
be issued when a node like node 18 or 20 fails to
acknowledge receipt of a timely heartbeat 42 or 44 from a
respective cluster mate. Heartbeats 42 and 44 may allow
.313
nodes 18 and 22 respectively to Fsee" if their cluster
mates are still functioning.
If, for example, node 18 can no longer "see" node
16, node 18 may seek to have any LUN reservations held
for node 16 released. To accomlish this release, node
18 may send a SCSI reset command to initiate a low-level
bus reset of the SCSI buses associated with nodes 16 and
18. In some systems, for example a MSCS system, node 18
may wait some specified amount of time before trying to
reserve the LUNs that had been reserved by node 16. The
waiting allows node 16 to regain control of the LUNs
As such,
is "alive" despite node 18's failure to receive
reserved to it before the SCSI reset. if node
heartbeat 42, node 16 may be able to re-establish its
resource reservations and in so doing let node 18 know
that it is "alive".
Unfortunately, as mentioned above, a SCSI reset in a
clustered computing environment can have a detrimental
effect on overall system performance and system
reliability. The disclosed system and resource
management engine may help lindt a clustered computing
environment‘s reliance on SCSI resets in several
different ways. Example techniques for avoiding SCSI
resets may be better understood through consideration of
FIGURES 2 and 3.
FIGURE 2 depicts a flow diagram of one embodiment
of a method 100 for managing storage resources in a
clustered computing environment. The method of FIGURE 2
may be implemented by a resource management engine
executing on a storage controller attached to a SAN
Mfabric. In some embodiments, the resource management
engine may be executing on a CPU associated with a switch
like switch 36 of FIGURE 1. In other embodiments, the
CPU may be associated with a SAN device other than the
switch. For example, a resource management engine may be
executing on one or more nodes of a SAN.
During the operation of a SAN, a port login (PLOGI)
comand may be received. As is known in the art, a PLOGI
command is a fibre channel comand wherein a node logs
into a storage device attached to a SAN. A node may
execute a PLOGI command after the fabric has assigned a
fibre channel ID (S_ID) to the node. As is also
conventionally known, the S_ID of a node may be assigned
when a node executes a fabric login (FLOGI) command.
At step 102, the S_ID and the WWN of a cluster node
may be extracted. The extraction may occur at different
times. For example, the extraction may occur when a node
issues a PLOGI command. Once extracted, the S_ID and the
WWN may be updated and may be stored in a computer
readable medium. In some embodiments, this computer
readable medium may be part of a SAN and may be
accessible to several devices of the SAN.
At step 104, a LUN reservation may be held for a
given node. In effect, the given node may have the
exclusive right to use the reserved LUN. As is mentioned
above, cluster nodes often communicate with one another
using a heartbeat signal. At step 106, a SAN device may
detect a failure to receive a timely heartbeat signal.
Though the failure to receive a heartbeat signal may only
indicate a failed comunication link between the
heartbeat sender and the heartbeat receiver, the failure
may result, as shown at step 108, in the determination
that a cluster node is inoperable.
In the embodiment of FIGURE 2, the determination
that a node is inoperable, may cause another node to
issue a SCSI reset. As shown at step 110, a SCSI reset
command may be sent to release LUN reservations held for
the node believed to be inoperable (the "dead" node). At
step 112, the SCSI reset command may be converted into a
third party process log out. This.conversion may, for
example, be performed by an executing resource management
engine.
At step 114 a log out comand for the “dead” node
may be sent on the "dead" node's behalf by a third party.
For example, a resource management engine may access a
computer readable medium storing the "dead" node's S_ID
and WWN.
and the WWN of the "dead" node to log out the "dead"
The resource management engine may use the S_ID
node. This third party process log out may result in the
releasing of LUN reservations held for the logged out
node.
As shown at step 116 of FIGURE 2, other nodes of a
cluster may also log out or be logged out and a loop
initialization protocol (LIP) link reset may be
initiated. The LIP link reset of step 118 may be
followed by step 120's generation of a state change
notification. In the embodiment of FIGURE 2, the state
change notification may cause active cluster nodes, nodes
that are not dead, to perform a port login and to seek
LUN reservations. The port login of active cluster nodes
may be seen at step 122. If the "dead" node was not
dead, it may be able to regain its LUN reservations. If
the "dead" node was dead, other cluster nodes may now be
able to capture the LUN reservations held by the "dead"
node. In effect, the storage resources held by the dead
node will be made available to "live" nodes —— resulting
in a better utilization of storage resources —— without a
SCSI reset.
Various changes to the above embodiments are
contemplated by the present disclosure. For example,
embodiments of the present disclosure may be implemented
in SANs having any number of topologies. There may be,
for example, numerous storage controllers, there may be a
resource management engine executing on each node of a
cluster, or there may be a single resource management
engine executing within each zone of a clustered
computing environment.
Claims (22)
1. A method for managing storage resources in a clustered computing environment, the method comprising: holding a reservation on a storage resource for a first node of the clustered computing environment; performing a third party process log out for the first node; and, releasing the reservation held for the first node.
2. The method of Claim 1, further comprising determining that the first node is not functioning.
3. The method of Claim 1 or Claim 2, further comprising: failing to receive a heartbeat signal from the first node; and, determining that the first node is not functioning as a result of the x failure to receive the heartbeat signal.
4. The method of any one of the preceding claims, in which the third party process log out is initiated from a second node of the clustered computing environment.
5. The method of any one of Claims 1 to 3, in which the third party process log out is initiated from a switch associated with the clustered computing environment.
6. The method of any one of the preceding claims, further comprising creating a zone within a Fibre Channel storage system, the zone comprising the first node of the clustered computing environment, a second node of the clustered computing environment, and at least one storage resource.
7. The method of any one of the preceding claims, wherein the storage resource comprises a hard drive.
8. The method of any one of the preceding claims, further comprising: storing identification information about the first node; and, using the identification information to issue a third party process log out request.
9. The method of any one of the preceding claims, wherein the first node comprises a server.
10. The method of any one of Claims 1 to 8, wherein the first node comprises a workstation.
11. The method of any one of the preceding claims, further comprising: assigning a worldwide name and a source identifier to the first node; and, using the worldwide name and the source identifier to issue a third party process log out request on behalf of the first node.
12. The method of Claim 11, further comprising: assigning a second worldwide name and a second source identifier to a second node; logging out the first node; and, logging out the second node.
13. The method of any one of Claims 1 to 11, further comprising logging out a second node of the clustered computing environment.
14. The method of Claim 13, further comprising generating a state change notification with a switch associated with the clustered computing environment.
15. The method of Claim 13 or Claim 14, further comprising: initiating a loop initialisation protocol link reset; and, generating a state change notification in response to the loop initialisation protocol link reset.
16. The method of Claim 13 or Claim 14, further comprising: in response to the state change notification, performing a responsive port login with the second node.
17. A computer system operable to manage storage resources in a clustered computing environment, the computer system including a first node for which a reservation is held on a storage resource, and a resource management engine operable to perform a third party process log out for the first node and release the reservation held for the first node.
18. A computer system according to Claim 17, including a second node, in which the third party process log out for the first node and release the reservation held for the first node is performed in response to a failure by the second node to receive a timely heartbeat signal from the first node.
19. A computer system according to Claim 17 or Claim 18. further comprising a computer readable medium storing the resource management engine and a central processing unit communicatively coupled to the computer readable medium ad operable to execute the resource management engine.
20. A computer system according to any one of Claims 17 to 19, further comprising a plurality of computer platforms communicatively coupled to the first node.
21. A method for managing storage resources in a clustered computing environment, substantially as described with respect to the accompanying drawings.
22. A computer system substantially as shown in or as described with respect to any of the accompanying drawings. F. R. KELLY & CO., AGENTS FOR THE APPLICANTS
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
USUNITEDSTATESOFAMERICA09/03/20000 | |||
US09/524,401 US6622163B1 (en) | 2000-03-09 | 2000-03-09 | System and method for managing storage resources in a clustered computing environment |
Publications (2)
Publication Number | Publication Date |
---|---|
IE20010108A1 IE20010108A1 (en) | 2001-09-19 |
IE83771B1 true IE83771B1 (en) | 2005-01-12 |
Family
ID=
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6622163B1 (en) | System and method for managing storage resources in a clustered computing environment | |
IE84170B1 (en) | System and method for managing storage resources in clustered computing environment | |
US7272674B1 (en) | System and method for storage device active path coordination among hosts | |
JP4567125B2 (en) | Method and apparatus for transferring write cache data in data storage and data processing system | |
US7711979B2 (en) | Method and apparatus for flexible access to storage facilities | |
US7921431B2 (en) | N-port virtualization driver-based application programming interface and split driver implementation | |
US7739541B1 (en) | System and method for resolving cluster partitions in out-of-band storage virtualization environments | |
US6594698B1 (en) | Protocol for dynamic binding of shared resources | |
US6247077B1 (en) | Highly-scalable parallel processing computer system architecture | |
US6256740B1 (en) | Name service for multinode system segmented into I/O and compute nodes, generating guid at I/O node and exporting guid to compute nodes via interconnect fabric | |
US7631066B1 (en) | System and method for preventing data corruption in computer system clusters | |
US20070022314A1 (en) | Architecture and method for configuring a simplified cluster over a network with fencing and quorum | |
US20090089609A1 (en) | Cluster system wherein failover reset signals are sent from nodes according to their priority | |
KR20040015223A (en) | Resource action in clustered computer system incorporating prepare operation | |
JP2010113707A (en) | Method, device, system, and program for dynamically managing physical and virtual multipath input/output | |
JP2005216151A (en) | Resource operation management system and resource operation management method | |
TW200925878A (en) | System and method for management of an IOV adapter through a virtual intermediary in an IOV management partition | |
US20060156055A1 (en) | Storage network that includes an arbiter for managing access to storage resources | |
US10708140B2 (en) | Automatically updating zone information in a storage area network | |
KR20100106460A (en) | A method, apparatus and system to dynamically manage logical path resources | |
US8707082B1 (en) | Method and system for enhanced granularity in fencing operations | |
US10782889B2 (en) | Fibre channel scale-out with physical path discovery and volume move | |
EP2382541B1 (en) | Computer-implemented multi-resource shared lock | |
AU2005200529B2 (en) | System and method for managing storage resources in a clustered computing environment | |
IE83771B1 (en) | System and method for managing storage resources in a clustered computing environment |