WO1999017204A1 - Redundant controller diagnosis using a private lun - Google Patents
Redundant controller diagnosis using a private lun Download PDFInfo
- Publication number
- WO1999017204A1 WO1999017204A1 PCT/US1998/020347 US9820347W WO9917204A1 WO 1999017204 A1 WO1999017204 A1 WO 1999017204A1 US 9820347 W US9820347 W US 9820347W WO 9917204 A1 WO9917204 A1 WO 9917204A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- controller
- data
- comparing
- controllers
- private lun
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2053—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
- G06F11/2089—Redundant storage control functionality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/0703—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
- G06F11/0751—Error or fault detection not based on redundancy
- G06F11/0754—Error or fault detection not based on redundancy by exceeding limits
- G06F11/0757—Error or fault detection not based on redundancy by exceeding limits by exceeding a time limit, i.e. time-out, e.g. watchdogs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2053—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
- G06F11/2089—Redundant storage control functionality
- G06F11/2092—Techniques of failing over between control units
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/22—Detection or location of defective computer hardware by testing during standby operation or during idle time, e.g. start-up testing
- G06F11/2205—Detection or location of defective computer hardware by testing during standby operation or during idle time, e.g. start-up testing using arrangements specific to the hardware being tested
- G06F11/221—Detection or location of defective computer hardware by testing during standby operation or during idle time, e.g. start-up testing using arrangements specific to the hardware being tested to test buses, lines or interfaces, e.g. stuck-at or open line faults
Definitions
- the invention relates generally to failure detection methods operable within a computer data storage subsystem and in particular to detecting failure in redundant controllers using a shared portion of a data storage area as a private LUN for diagnostic purposes.
- a popular choice that meets demands for increased capacity and reliability is the use of multiple smaller storage modules. These modules may be configured in geometries that permit redundancy of stored data to ensure data integrity in case of various failures. In many such redundant subsystems, recovery from many common failures can be automated within the storage subsystem itself due to the use of data redundancy, error codes, and so-called "hot spares" (extra storage modules that may be activated to replace a failed, previously active storage module). These subsystems are typically referred to as redundant arrays of inexpensive (or independent) disks (or more commonly by the acronym RAID). The 1987 publication by David A.
- RAID level 1 The simplest array, a RAID level 1 system, comprises one or more disks for storing data and an equal number of additional "mirror" disks for storing copies of the information written to the data disks.
- the remaining RAID levels identified as RAID level 2,3,4 and 5 systems, segment the data into portions for storage across several data disks. One or more additional disks are used to store error check or parity information.
- RAID storage subsystems typically utilize one or more controllers that shield the user or host system from the details of managing the redundant array.
- the controller makes the subsystem appear to the host computer as a single, highly reliable, high capacity disk drive.
- the RAID controller may distribute the host computer system supplied data across a plurality of the small independent drives with redundancy and error checking information to improve subsystem reliability.
- RAID subsystems provide large cache memory structures to further improve the performance of the RAID subsystem.
- the cache memory is associated with the control module such that the storage blocks on the disk array are mapped to blocks in the cache. This mapping is also transparent to the host system. The host system simply requests blocks of data to be read or written and the RAID controller manipulates the disk array and cache memory as required.
- pairs of controllers are configured such that they control the same physical array of disk drives. Each controller is assigned a portion of the data storage area, which it controls. The controllers may operate simultaneously using their assigned area. Upon failure of one of the controllers, the other controller can takeover operation of the area assigned to the failed controller.
- the surviving controller logically assumes ownership of the data storage area assigned to the failed controller. These data storage areas are commonly referred to as logical units or LUNs. If the controllers share a common bus, the surviving controller may also assume the SCSI ID of the failed controller so that it responds to both its original ID and the failed controller's ID.
- the problem addressed by the present invention concerns how failure is detected in one or more of the controllers. It is generally known in the art to send a timed alert message from one controller to the other. This alert message may be some type of query that asks the receiving controller to send a message back indicating that it is operational. If no message is received by the controller sending the message, it assumes that the receiving controller has failed. It then proceeds to takeover operations of the receiving controller's logical units.
- failure detection methods are deficient in their ability to ascertain the type of failure or where a particular failure may have occurred (e.g., in what component or data path).
- known methods are deficient in testing processes operating within the controllers, data paths from controller to storage area, the storage area itself, and data paths between controller and host computer.
- the simple alert message test noted above is insufficient to determine which of the controllers actually failed or whether the data path over which the alert message and response are exchanged has failed.
- the present invention solves the above and other problems, and thereby advances the useful arts, by providing methods and associated apparatus for detecting failure and switching over between redundant controllers in data storage subsystems.
- the methods and apparatus provide for failure detection methods operable within data storage subsystems having redundant controllers. A portion of the data storage area of the subsystem is reserved for diagnostic purposes of the redundant controllers to detect failure in the redundant controllers or the data paths common to these controllers. This improves over known methods by more particularly locating where the failure has occurred and by checking additional areas of the system such as the data paths.
- This reserved portion of the data storage area is referred to as a private LUN in that it is not generally accessible by standard system or application processes in attached host systems. Rather, the private LUN is used by all redundant controllers commonly attached to the private LUN for diagnosing their redundant operations. The diagnosis capability of the controllers is enhanced by including tests of the read and write data paths between the controllers and the data storage devices as well as testing the data paths between the controllers.
- the private LUN is a portion of the data storage area that is known and accessible to controllers in the system and can be used by them for diagnostic purposes consistent the RAID management levels supported by the controllers (e.g., including RAID parity testing). In this respect it is unlike other portions of the data storage area that are assigned to each controller and used for persistent data storage.
- the private LUN may be of any size, but is preferably defined as a striped unit accessing at least one storage device on each storage device interface channel in the subsystem.
- the private LUN is dynamically mapped such that the number of drives in the private LUN is dynamically changed when a device fails.
- mapping of the private LUN is "down sized" from a 4+1 array device to a 3+1 LUN.
- the mapping of the private LUN dynamically grows to include these drives as they are made operational (i.e., included in a logical unit definition).
- the methods of the present invention involve sending a diagnostic write command to a first controller with instructions for it to write test data to the private LUN.
- This first controller writes this test data to the private LUN.
- a second controller then reads this test data from the private LUN. Meanwhile, an original copy of the test data is sent to the second controller by other means.
- the second controller compares the read back test data to the value of the original test data before being written to the private LUN. Using the results, it can be determined if the first controller failed, the second controller failed or neither controller failed.
- Each controller of the redundant controllers may be assigned primary responsibility for particular LUNs (data LUNs other than the private LUN) as well as redundant or secondary responsibility for certain data LUNs. If the first controller fails, the second controller takes over ownership of portions of the data LUNs assigned to the first controller. The takeover is automatic in the sense that the switch over occurs within the controllers in the data storage subsystem without external control.
- the diagnosis and takeover processing using the private LUN is initiated by the controllers. No host computer need be involved or aware that failure has occurred depending upon host and controller interconnection topologies.
- a diagnostic write command is sent from a first controller to a second controller via a shared communication path.
- the diagnostic write command includes instructions for the second controller to write test data to the private LUN.
- the second controller writes the test data to the private LUN.
- the first controller reads the test data from the private LUN. This read-back test data may or may not have the same value as the original test data.
- the first controller then compares the read-back test data to the original test data. The results of this comparing step are used to determine which controller, if any, has failed.
- the failure detection method is repeated after a certain time interval. If the second controller failed, then the first controller takes over ownership of the portion of the data storage area previously assigned to the second controller. If the first controller (which did the comparing) fails, then the method stops within the first controller due to its failure. The method is repeated and is symmetrical. That is, the above method initiated by the first controller, to test the second controller, may likewise be initiated by the second controller, to test the first controller. This may then result in the second controller taking over LUNs owned by the first controller. For each particular pair of diagnosing controllers, the private LUN may define a unique "storage area" within dedicated to that diagnosing, redundant pair. The symmetric diagnostic process within that pair or other pairs will use their respective designated unique areas of the private LUN.
- the comparing process noted above includes several steps.
- the test data as written to the private LUN comprises a main data portion and a redundancy data portion.
- the comparing step includes comparing the main data portion of the read-back test data to the original test data. If they are not the same, then it is presumed that the second controller has failed. If they are the same, then further steps are performed.
- a first means is used to calculate additional redundancy data from the main portion of the read-back data (e.g., a parity assist circuit on the controller). This additional redundancy data is then compared to the redundancy data portion of the test data read from the private LUN. If they are the same, it is concluded that neither controller failed.
- a second means of the first controller e.g., software means within the controller
- This further additional redundancy data is likewise compared to the redundancy data portion of the test data read from the private LUN. If they are not the same, then it is confirmed that the second controller failed. As noted above in case of such failure, the first controller takes over LUNs owned by the second controller. If they are the same, then it is presumed that the first controller has failed due to failure of its first means for computing redundancy data and the method terminates within the first controller.
- one or more host computers are involved in the diagnosis.
- This method has the further advantage that data paths between host computers and controllers are diagnosed.
- a host computer sends a diagnostic write command to a first controller.
- the diagnostic write command carries instructions for the first controller to write test data to the private LUN.
- the host also sends this test data to a second controller.
- the second controller stores the host supplied test data in a first buffer.
- the first controller writes the test data to the private LUN.
- the second controller then reads the test data from the private LUN and stores it in a second buffer.
- the second controller compares the test data received from the host computer to the test data read from the private LUN. This comparing step is substantially as described above.
- An enhancement to the second method can diagnose failures in the host/controller communication path.
- the first controller After the first controller receives a diagnostic write command from the host, it starts a timer. If the first controller has not received a new diagnostic write command from the host in a designated time period, it sends an alert message to the second controller indicating that the host has not been heard from. The second controller, upon receiving this alert message will determine if it has received a diagnostic write command from the host within the designated period. If it has, it will invoke a takeover of the first controller's LUNs because it assumes that the host and first controller have lost communication ability. If the second controller has not received a diagnostic write command from the host during the specified period, it will ignore the alert message.
- FIG. 1 is a block diagram of a typical RAID storage subsystem in which the structures and methods of the present invention may be applied;
- FIG. 2 is a block diagram of a split-bus configuration in which redundant controllers as in FIG. 1 do not share a common communication path connecting each to host computers;
- FIG. 3 is a block diagram of a shared-bus configuration in which redundant controllers as in FIG. 1 share a common communication path connecting each to host computers and to one another;
- FIG. 4 is a flowchart describing the operation of the first controller in accordance with a first embodiment of the invention
- FIG. 5 is a flowchart describing the operation of the second controller in accordance with a first embodiment of the invention.
- FIG. 6 is a flowchart describing additional detail of the comparing step of FIG. 4;
- FIG. 7 is a flowchart describing in greater detail an timing element in the operation of the first controller in FIG. 4;
- FIG. 8 is a flowchart describing the operation of a host computer in accordance with a second embodiment of the invention;
- FIG. 9 is a flowchart describing the operation of the first controller in accordance with a second embodiment of the invention.
- FIG. 10 is a flowchart describing the operation of the second controller in accordance with a second embodiment of the invention.
- FIG. 11 is a flowchart describing timing steps of the first controller in accordance with a second embodiment of the invention.
- FIG. 12 is a flowchart describing timing steps of the second controller in accordance with a second embodiment of the invention.
- FIG. 1 is a block diagram of a typical RAID storage subsystem 100, having redundant disk array controllers 118.1 and 118.2 (hereinafter referred to as RDACs), in which the methods and associated apparatus of the present invention may be applied.
- RAID storage subsystem 100 includes at least one pair of RDACs 118.1 and 118.2.
- Each RDAC 118.1 and 118.2 is in turn connected to disk array 108 via buses 150.1 and 150.2, respectively, and to host computers 120.1 and 120.2 via buses 154.1 and 154.2, respectively.
- Disk array 108 is composed of a plurality of disk drives 110.
- interface buses 150.1 and 150.2 between RDACs 118.1 and 118.2, respectively, and disk array 108 (including disk drives 110) may each be comprised of a plurality of buses and may be any of several industry standard interface buses including SCSI, IDE, EIDE, IPI, Fiber Channel, SSA, PCI, etc. Circuits (not shown) within RDACs 118.1 and 118.2 appropriate to controlling buses 150.1 and 150.2, respectively are well known to those of ordinary skill in the art.
- Interface buses 154.1 and 154.2 between RDACs 118.1 and 118.2 and host computers 120.1 and 120.2, respectively may be any of several standard industry interface buses including SCSI, Fibre Channel, Ethernet (LAN), Token Ring (LAN), etc.
- RDACs 118.1 and 118.2 appropriate to controlling bus 154.1 and 154.2, respectively, are well known to those of ordinary skill in the art. Further, those skilled in the art will readily recognize that the methods of the present invention may operate within any number of redundant controller.
- the pair of RDACs 118.1 and 118.2 shown in FIG. 1 are therefore intended as suggestive of any plurality of redundant controllers.
- RAID storage subsystem 100 may be utilized to implement any of the well-known RAID levels (e.g., levels 0-5).
- the various RAID levels are distinguished by the manner in which the associated RAID controllers logically subdivides or partitions the disk drives 110 in disk array 108.
- RAID level 1 when implementing RAID level 1 features, approximately half of the disk drives 110 of disk array 108 are used to store and retrieve data while the other half is operated by the RAID controller to mirror the data storage contents of the first half.
- the RAID controller when implementing RAID level 4 features, utilizes a portion of the disk drives 110 in disk array 108 for the storage of data and the remaining disk drives 110 are utilized for the storage of redundancy data such as error checking/correcting information (e.g., parity information).
- error checking/correcting information e.g., parity information
- RDAC 118.1 includes CPU 112.1 , program memory 114.1 (e.g., ROM/RAM devices for storing program instructions and variables for the operation of CPU 112.1), and cache memory 116.1 for storing data and control information related to the data stored in disk array 108.
- CPU 112.1 , program memory 114.1 , and cache memory 116.1 are connected via memory bus 152.1 to enable CPU 112.1 to store and retrieve information in the memory devices.
- RDAC 118.2 is identical to RDAC 118.1 and is comprised of CPU 112.2, program memory 114.2 and cache memory 116.2, all interconnected via memory bus 152.2.
- the RDACs 118.1 and 118.2 are preferably interchangeable devices within RAID subsystem 100 to permit easy replacement, including hot swap, of a defective RDAC.
- FIG. 1 is intended only as an exemplary design that may embody the methods and structures of the present invention. Many alternative controller and subsystem designs may embody the methods and associated apparatus and structures of the present invention.
- a RAID subsystem may subdivide or partition the disk drives 110 of disk array 108 to define a plurality of logical units (data LUNs) for storage of host system data.
- data LUNs logical units
- each data LUN is associated with a primary controller (e.g., RDAC 118.1 or 118.2).
- RDAC 118.1 or 118.2 Other controllers serve as a redundant or secondary controller for that data LUN in a passive operating or standby mode with respect to that data LUN.
- Those redundant controllers may be primary controllers for others of the plurality of data LUNs.
- the redundant controllers stand prepared to takeover operations on behalf of the primary controller when they sense a failure in the operation of the primary controller.
- the redundant controller effectuate a takeover by assuming primary control responsibilities for the data LUNs associated with the failed primary controller.
- the redundant controllers therefore include methods to sense a failure in operation of the primary controller so as to initiate takeover processing.
- the present invention addresses problems in this need to sense a failure in the operation of another controller.
- the present invention provides for methods and associated apparatus to reserve a portion of the storage area of disk array 108 as a private LUN.
- the private LUN is one, unlike the data LUNs, which is not accessible for storage and retrieval of data except diagnostic data in accordance with the present invention. Rather, the private LUN is reserved for access by the redundant controllers for diagnosis of the operation of other controllers commonly attached to the private LUN.
- the methods and apparatus of the present invention provide for writing of test data in the private LUN by a first of the redundant controllers and reading of that test data from the private LUN by a second of the redundant controllers. By performing various comparisons on the data read back from the private LUN, the second controller may more accurately determine the nature and source of a failure as compared to prior techniques.
- methods of the present invention are operable within the redundant controllers of the RAID subsystem, devoid of interaction with attached host computers.
- the redundant controllers communicate amongst one another to effectuate the diagnostic sequences involving the private LUN for purposes of identifying a failure of one of the redundant controllers.
- the diagnostic commands may be issued over paths between the controllers used for host communication (e.g., common host channels) or other paths used for disk drive communication (e.g., common disk channels).
- the methods are modified to include host interaction in the diagnostic processes.
- This second embodiment enables more thorough testing of the redundant controllers by permitting testing of the common communication path using the host computer.
- Those skilled in the art will recognize a wide variety of communication topologies wherein multiple controllers may communicate with common host computers.
- a particular exemplary embodiment is described herein where multiple controllers share a single common connection with a single common host system.
- a variety of equivalent configurations may be envisioned wherein a single host system may communicate with multiple controllers.
- the multiple controllers need not necessarily reside on a common communication path with one another for application of this second embodiment. Both embodiments may be further enhanced by adding timed messaging features which, as known in the art, allow for identification of still other failure modes.
- FIG. 2 is a simplified view of FIG. 1 illustrating a split-bus configuration.
- Buses 154.1 and 154.2 directly connect host computers 120.1 and 120.2 with controllers 118.1 and 118.2, respectively.
- a host computer 120.1 cannot directly communicate with other controllers in the system (e.g., 118.2). Rather, each host system 120.1 and 120.2 has direct access only to one of the two controllers, 1 18.1 and 1 18.2, respectively.
- FIG. 2 further shows separate buses 150.1 and 150.2 connected to each disk drive 1 10. As will become apparent, the failure detection methods of the invention can be used to test each of the plurality of these buses 150.1 and 150.2.
- Each bus 150.1 and 150.2 is herein also referred to as a channel.
- the redundant controllers 1 18.1 and 1 18.2 can communicate with one another via the disk channels (150.1 and 150.2).
- redundant controller may also share a dedicated communication channel dedicated to communication of information and diagnostics amongst the redundant controllers.
- a dedicated channel might include a serial communication channel, LAN communications, SCSI, Fibre Channel, etc.
- the methods of the present invention as applied in the split bus configuration require merely that the redundant controllers share some communication channel useful for exchange of diagnostic requests and data.
- FIG. 3 shows what is hereinafter referred to as a shared-bus configuration.
- FIG. 3 is the same as FIGS.
- Bus 154.3 interconnects host computers 120.1 and 120.2 and controllers 1 18.1 and 1 18.2. This interconnection, as will be seen, makes the shared-bus configuration appropriate for the second embodiment of the invention wherein a host computer initiates the failure detection method by sending a diagnostic write command to two controllers. This second embodiment therefore enables more thorough testing of the operation of the redundant controller. The controllers' communications with a host computer are also tested by the methods and structures of the second embodiment of the invention. Those skilled in the art will recognize a variety of interconnection topologies between host computers and RAID controllers in which the methods and associated apparatus of the present invention may be advantageously applied.
- one or more host systems may have redundant connections to one or more RAID controllers.
- one or more RAID controllers may provide redundant connection paths to one or more host systems.
- host systems may have multi-path management subsystems within (e.g., dual ported Fibre Channel connections between host systems and storage subsystems). Takeover processing and associated notification of detected failures may therefore include various combinations of host system and RAID controller involvement.
- FIG. 3 is intended to represent any such topologies or configurations where a one (or more) host systems may communicate with each of a redundant multiplicity of controllers.
- the host system and controllers may or may not physically share a single common communication path.
- the shared bus configuration represented in FIG. 3 is intended to represent any configuration where at least one host system may communicate with each of a redundant multiplicity of controllers.
- FIG. 4 is a flowchart describing a first method of the present invention operable within a first controller to initiate the failure detection method of the invention.
- the first controller may be RDAC 118.1 in FIGS. 1-3, for example. However the method may be simultaneously initiated by RDAC 118.2 as the first controller. In other words, the methods of the present invention are symmetric and operable among all redundant controllers to test other redundant controllers.
- Elements 201-206 verify operation of the first controller 118.1 prior to initiating the diagnostic sequence to the second controller 118.2.
- element 201 is operable to write a first test pattern of data on a private LUN common to both controllers.
- the private LUN is preferably striped across all disk drives 110 in disk array 108 so as to effectuate a test of all disk channels 150.1 and 150.2 associated with both controllers.
- elements 203 determines that there is an error in writing the test pattern data or in reading the information back, it is determined that the first controller 118.1 has failed and the method continues at element 260 (label C) by awaiting takeover by the second controller. Otherwise, elements 204-206 repeat the same sequence, writing test data, reading it back, and sensing any errors, but with a second test pattern of data.
- elements 201-206 also serve to assure that any old data previously written to the private LUN is overwritten. Thus, old data will not be confused in a manner to defeat the testing of the controllers via the private LUN. If elements 201-206 determine that the first controller 118.1 is properly functional with respect to access to the private LUN, processing continue with element 210.
- Element 210 is operable to send a diagnostic write command from the first controller 118.1 to a second controller 118.2.
- the diagnostic write command comprises instructions for the second controller to write first test data to the private LUN.
- Element 220 shows the first controller awaiting action by the second controller to write the first test data to the private LUN.
- element 230 shows reading of the test data from the private LUN by the first controller.
- the first controller reads the entire stripe of test data the second controller has written.
- the read test data is merely the value of the test data after it has been written to the private LUN by the second controller and read back by the first controller.
- the value of the read test data may or may not be the same as that of the original test data. Differences in comparing the read and original test data may reveal particular failures of one of the redundant controllers.
- Element 240 shows comparing of the read test data to the original test data within the first controller. The results are used to determine which controller, if any, has failed. Label A in element 240 indicates the condition in which neither controller failed. The method loops back to START at element 210 and is continuously repeated (preferably after a time delay).
- Label B out of element 240 indicates the condition in which the second controller has failed.
- the second controller is the controller that was sent the diagnostic write command by the first controller. This condition continues execution of the method by jumping to element 250 to initiate takeover processing.
- the first controller takes over ownership of those portions of the data storage area assigned to the second controller.
- Label C out of element 240 indicates the condition in which the first controller has failed.
- the first controller is the controller performing the comparing. If the first controller fails, the method terminates in the first controller and the first controller becomes passive until appropriate repairs are enacted.
- the details of the comparing step are further explained below in reference to FIG. 6.
- FIG. 5 shows a method of the first embodiment of the invention operable within the second controller.
- the second controller may be RDAC 1 18.2, for example, if the first controller is RDAC 1 18.1 and vice versa.
- Element 270 shows the second controller awaiting diagnostic command instructions from the first controller. Namely, the second controller is awaiting diagnostic write command 210 from the first controller. After a receiving a diagnostic write command, the method continues with element 280 representing the second controller writing test data to the private LUN.
- test data is supplied with the diagnostic write command from the first controller.
- the method continues with element 290 wherein the second controller returns status to the first controller indicating that the supplied test data has been written to the private LUN. After completing element 290, the method loops back to START at element 270 and again awaits instructions from the first controller.
- FIG. 6 shows the comparing process of element 240 of FIG. 4 in more detail. Similar comparison element is operable in a second embodiment of the invention discussed below. FIG. 6 is therefore a detailed view of element 240 of both FIG. 4 (portion of a first embodiment) and FIG. 10 (portion of a second embodiment).
- the labeled paths A, B and C in FIG. 6 correspond to those in FIG. 4 and FIG. 10.
- element 240 is operable in the controller that is performing the comparison of test data as originally provided and as read back from the private LUN after being written thereto.
- the test data written to the private LUN comprises a main data portion and a redundancy data portion.
- the redundancy data portion may be parity data calculated from the main data portion by an XOR operation as is common in RAID techniques.
- Element 242 in FIG. 6 shows comparing the main data portion of the test data that has been read from the private LUN to the value of the test data before it was sent to the private LUN. The method continues with decision element 244. If the read-back main test data is not the same as the as the original test data, then an error has occurred and processing continues at label B representing takeover by the controller performing the comparison of the data LUNs assigned to the other controller. It is presumed that there is some type of problem in the other controller's ability to accurately store data in the private LUN.
- Element 246 represents an operation within the controller doing the comparing of using a first means (e.g., parity assist circuits) to calculate redundancy data from the main portion of the test data which has been read from the private LUN. The redundancy values so calculated are compared to the redundancy data portion of the test data read from the private LUN. If they are the same, then it is presumed that the redundancy generating means in the controller is operating correctly and processing continues at label A.
- a first means e.g., parity assist circuits
- a second means e.g., software computation
- a second means e.g., software computation
- These second calculated redundancy values are compared to the redundancy data portion of the test data read from the private LUN. If again they are not the same, then it is confirmed that there is not a problem with the first redundancy data calculating means.
- the error is presumed to be with the other controller and processing continues at label B to initiate takeover processing from the other controller. If the values are the same as determined by element 248, then it is presumed that the problem is with the redundancy data generating means of the controller doing the comparison.
- the method terminates in the controller doing the comparison as indicated at label C.
- the controller performing the comparison is presumed to have a faulty parity computation means (e.g., a faulty parity assist circuit).
- FIG. 7 is a detailed description of the processing of element 220 in FIG. 4.
- Element 220 represents a timed failure test operable in the first controller. After sending a diagnostic write command to the second controller, the first controller starts a timer.
- Element 222 in FIG. 7 represents a decision wherein a response is due within a predetermined timeout from the second controller. If no response is received, then it is presumed that the second controller has failed. The first controller takes over data LUNs assigned to the second controller and processing continues at label B of FIG. 4. If time has not run out, the other possibility is that there may be some error in receiving a response from the second controller.
- processing continues with element 223 to determine whether an invalid or unexpected status is returned.
- the requisite response may have been received but not in the expected form (e.g., a corrupted or otherwise erroneous version thereof). If such an erroneous response was detected, then it is presumed that the second controller has failed.
- the first controller takes over data LUNs assigned to the second controller and processing continues at label B of FIG. 4. If there is no error in receiving a response from the second controller, then the method continues at element 224 to determine whether the second controller returned a failure status. The second controller may return such a status if it senses that it is unable to complete the requested diagnostic write command for any of several reasons.
- processing continues at label B of FIG. 4. If an appropriate response is received from the second controller and no error status is indicated, processing continues with element 230 shown in FIG. 4.
- a host computer is involved in the failure detection method.
- This second embodiment has the further advantage that the diagnostic processing tests the data paths between one or more host computers and the redundant controllers.
- This second embodiment is operable within the configuration shown in FIG. 3 wherein a common bus 154.3 is used between a host computer (e.g., 120.1 and/or 120.2) and controllers 118.1 and 118.2.
- FIG. 8 is a flowchart describing a method of the present invention operable within a host computer 120.1 of FIG. 3 to initiate and coordinate control of the diagnostic process in accordance with the second embodiment of the invention.
- a symmetrical method could be simultaneously initiated by another host computer 120.2 reversing the roles of the first and second controllers as described below.
- a single host computer 120.1 may initiate the diagnostic processes reversing the roles of the first and second controllers as described below.
- Element 310 of FIG. 8 is first operable to send a diagnostic write command from a host computer 120.1 to a first controller 118.1 of the redundant controllers.
- the diagnostic write command contains instructions for the first controller to write host supplied test data to the private LUN.
- the method continues with element 320 in which a status is received by the host computer from the first controller indicating whether the diagnostic test completed writing to the private LUN.
- a determination of the success or failure of the diagnostic test of the first controller will be performed by completing the test procedure with the second controller. In other words, the second controller will determine whether the writing operation of the first controller was a success or a failure. Processing then continues with element 330.
- Element 330 shows that a similar special diagnostic write command is sent to a second controller 118.2 of the redundant controllers.
- Element 330 provides the second controller with the same test data as sent to the first controller and presumed correctly written to the private LUN.
- the special diagnostic write command to the second controller 118.2 is special in that it includes an identification that the test data supplied is not to be written to the private LUN as with the first controller. Rather, the supplied test data in the special diagnostic write command to the second controller 118.2 is used for comparison with the test data to be read from the private LUN.
- Element 340 shows that a status is received by the host from the second controller indicating that it has performed the special diagnostic write command and determined whether the first controller failed, second controller failed, or neither controller failed.
- Element 350 determines whether the returned status from the second controller 118.2 indicates that the second controller 118.2 has failed. If so, processing continues with element 354 to await the expected takeover by the first controller 1 18.1 of the data LUNs controlled by the failed second controller118.2 . If the second controller 1 18.2 has not failed as determined by element 350, processing continue with element 352. Element 352 determines whether the returned status from the second controller 118.2 indicates that the first controller 118.1 has failed. If so indicated, processing continues with element 324, as above, to await the expected takeover by the second controller 118.2 of data LUNs controlled by the failed first controller 118.1.
- FIG. 9 shows a method of the second embodiment which is cooperable within a first controller 118.1.
- Element 370 indicates a diagnostic write command is received from the host computer.
- Element 380 indicates that a timer is restarted. Details of the timer option are discussed below with reference to FIGS. 11 and 12. This timer marks a specified time starting from the moment a diagnostic write command is received by the first controller 1 18.1. This enables the first controller 1 18.1 to determine if there is a problem in receiving diagnostic write commands from the host computer.
- Element 390 next indicates that the first controller 1 18.1 executes the diagnostic write command by writing the host supplied test data to the private LUN.
- the first controller 118.1 then sends a message to the host computer indicating status of completion of the diagnostic write command within the first controller 118.1.
- error conditions may include, for example, receipt of erroneous test data as indicated by an improper redundancy portion as compared to the supplied test data.
- the first controller 118.1 may, for example, check the redundancy portion of the test data by computing an expected redundancy value for the main data portion by any or all of multiple redundancy computation means.
- redundancy computations may be used to determine whether such a failure likely lies in the host/controller communication path or in the controller redundancy assist circuits (e.g., XOR parity assist circuits).
- FIG. 10 shows a method of the second embodiment cooperable within a second controller 118.2.
- Element 410 indicates that a special diagnostic write command is received by the second controller 118.2 from the host computer 120.1.
- the special diagnostic write commands provide the test data which should have been written to the private LUN by operation of the methods in the first controller 118.1 described above.
- Element 420 shows the second controller 118.2 reading test data from the private LUN that was written by the first controller 118.1.
- Element 240 indicates that the read test data is compared to the original test data supplied in the special diagnostic write command received from the host.
- Element 240 of FIG. 10 is similar to element 240 in FIG. 4 and as described in additional detail in FIG. 6.
- the labeled paths A, B and C correspond to the three outcomes in the comparison performed by element 240.
- the outcome directed to label A indicates that the comparisons resulted in no detected failure of either controller. Such a success status is returned to the host by element 460.
- the outcome directed to label B indicates a failure of the second controller 118.2, the controller performing the comparison and finding a probable failure in its redundancy computation means.
- the first controller 118.1 will eventually takeover control of the data LUNs normally controlled by this failed second controller 118.2.
- Element 450 returns to the host such a failed status for the second controller 118.2.
- the outcome directed to label C indicates a probable failure in the operation of the first controller 118.1.
- Element 430 therefore initiates takeover processing by the second controller 118.2 to assume control of the data LUNs previously controlled by the failed first controller 118.1. Such a status is then returned to the host by the second controller 118.2 by operation of element 440. The method of the second controller 118.2 thereby complete processing.
- the methods of both the first and second embodiments may be enhanced by addition of timed event alert messaging among the redundant controllers.
- a timer is restarted each time the first controller 118.1 receives a diagnostic write command from a host system. This time is used to enhance the methods of the second embodiment as described below with respect to FIGS. 11 and 12.
- Such timer and event messaging techniques often generally referred to as watchdog timers, are well known in redundant systems. Though depicted here only with respect to the methods applied to the shared bus configuration (second embodiment), such timer and event messaging techniques are equally applicable to the methods described above with respect to the first embodiment. Those skilled in the art will readily recognize the applicability of such techniques largely regardless of the particular communication topology selected.
- FIG. 11 shows a method of the second embodiment operable within the first controller 118.1 to test for failure of data paths between controller and host computer.
- Element 510 detects whether the first controller has already been determined to have failed by operation of other tests described above. If the first controller 118.1 is so sensed to be in a failed state, the method completes. If there has not been a failure of the first controller 118.1 , then the process continues with element 512. Element 512 determines if the timer, restarted by element 380 in FIG. 9, has reached its time limit and expired. If it has not, then the process continues looping back to element 510 to repeat its processing (after a suitable delay time). If the time has expired, then the process continues with element 514.
- An alert message is sent to the second controller 118.2 indicating that the host has not been heard from by the first controller 118.1. This may indicate that there is a problem in the path between the host computer and the first controller 118.1. However, an alternative is that no diagnostic message was sent by the host for reasons unknown to the first controller 118.1.
- the second controller 118.2, as described below will then determine if the possible failure in fact indicates a failure of the first controller 118.1. Element 516 is then operable to restart the timer again and the process continues by looping back to element 510 (following a suitable delay period).
- FIG. 12 a method of the second embodiment operable within the second controller 118.2 of a timed operation that checks data paths between controllers.
- Element 520 awaits receipt of an alert message by the second controller 118.2 from the first controller 118.1.
- element 522 is operable to determine if any special diagnostic write commands have been received from the host in a specified time (preferably the time period used to reset the timer of the first controller 118.1). If special diagnostic write commands have been received by the second controller 118.2 from the host computer during this period, then it is presumed that the first controller 118.1 has failed or that its connection with the host computer has failed. Therefore the process continues at label C of FIG.
- the second controller 118.2 If the second controller 118.2 has not received a special diagnostic write command from the host computer, then it presumes that there is no problem with the first controller 118.1 or its connection with the host (i.e., no diagnostic write command was sent by the host computer to either controller). In this case, the alert message is ignored by the second controller 118.2 and processing continue by looping back to element 520 to await another alert message from the first controller 118.1.
- the methods of the embodiments described above are advantageous in that by using a private LUN striped across all disk drive channels of the array, each of the channels can be individually tested.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Quality & Reliability (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Techniques For Improving Reliability Of Storages (AREA)
- Hardware Redundancy (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE69802833T DE69802833T2 (en) | 1997-09-30 | 1998-09-29 | DIAGNOSTIC OF REDUNDANT CONTROLLERS USING A PRIVATE LUN |
AU95114/98A AU9511498A (en) | 1997-09-30 | 1998-09-29 | Redundant controller diagnosis using a private lun |
EP98948569A EP1019823B1 (en) | 1997-09-30 | 1998-09-29 | Redundant controller diagnosis using a private lun |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/941,894 US5975738A (en) | 1997-09-30 | 1997-09-30 | Method for detecting failure in redundant controllers using a private LUN |
US08/941,894 | 1997-09-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1999017204A1 true WO1999017204A1 (en) | 1999-04-08 |
Family
ID=25477237
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US1998/020347 WO1999017204A1 (en) | 1997-09-30 | 1998-09-29 | Redundant controller diagnosis using a private lun |
Country Status (5)
Country | Link |
---|---|
US (1) | US5975738A (en) |
EP (1) | EP1019823B1 (en) |
AU (1) | AU9511498A (en) |
DE (1) | DE69802833T2 (en) |
WO (1) | WO1999017204A1 (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2366048A (en) * | 2000-05-02 | 2002-02-27 | Ibm | Selecting a preferred path to a storage device |
EP1707205A2 (en) | 2002-07-09 | 2006-10-04 | Boehringer Ingelheim Pharma GmbH & Co. KG | Pharmaceutical compositions of anticholinergics and p38 kinase inhibitors in the treatment of respiratory diseases |
US7130970B2 (en) | 2002-09-09 | 2006-10-31 | Commvault Systems, Inc. | Dynamic storage device pooling in a computer system |
EP1724660A1 (en) * | 2005-04-27 | 2006-11-22 | Hitachi, Ltd. | Disk array device |
WO2008142031A1 (en) | 2007-05-18 | 2008-11-27 | Institut Curie | P38alpha as a therapeutic target in bladder carcinoma |
EP2116245A2 (en) | 2004-08-07 | 2009-11-11 | Boehringer Ingelheim International GmbH | EGFR kinase inhibitor combinations for treating respiratory and gastrointestinal disorders |
EP2384751A1 (en) | 2004-12-24 | 2011-11-09 | Boehringer Ingelheim International Gmbh | Medicaments for the treatment or prevention of fibrotic diseases |
US8234520B2 (en) | 2009-09-16 | 2012-07-31 | International Business Machines Corporation | Wear leveling of solid state disks based on usage information of data and parity received from a raid controller |
US8510516B2 (en) | 2003-04-03 | 2013-08-13 | Commvault Systems, Inc. | Systems and methods for sharing media in a computer network |
US8639877B2 (en) | 2009-06-30 | 2014-01-28 | International Business Machines Corporation | Wear leveling of solid state disks distributed in a plurality of redundant array of independent disk ranks |
US9122600B2 (en) | 2006-12-22 | 2015-09-01 | Commvault Systems, Inc. | Systems and methods for remote monitoring in a computer network |
US9170890B2 (en) | 2002-09-16 | 2015-10-27 | Commvault Systems, Inc. | Combined stream auxiliary copy system and method |
US9507525B2 (en) | 2004-11-05 | 2016-11-29 | Commvault Systems, Inc. | Methods and system of pooling storage devices |
US9898213B2 (en) | 2015-01-23 | 2018-02-20 | Commvault Systems, Inc. | Scalable auxiliary copy processing using media agent resources |
US9904481B2 (en) | 2015-01-23 | 2018-02-27 | Commvault Systems, Inc. | Scalable auxiliary copy processing in a storage management system using media agent resources |
US10379988B2 (en) | 2012-12-21 | 2019-08-13 | Commvault Systems, Inc. | Systems and methods for performance monitoring |
US11010261B2 (en) | 2017-03-31 | 2021-05-18 | Commvault Systems, Inc. | Dynamically allocating streams during restoration of data |
US11593223B1 (en) | 2021-09-02 | 2023-02-28 | Commvault Systems, Inc. | Using resource pool administrative entities in a data storage management system to provide shared infrastructure to tenants |
Families Citing this family (81)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6073209A (en) * | 1997-03-31 | 2000-06-06 | Ark Research Corporation | Data storage controller providing multiple hosts with access to multiple storage subsystems |
US6665703B1 (en) * | 1998-01-02 | 2003-12-16 | Emc Corporation | Storage system with internal LAN |
JP3360719B2 (en) * | 1998-06-19 | 2002-12-24 | 日本電気株式会社 | Disk array clustering reporting method and system |
US6282584B1 (en) * | 1998-08-18 | 2001-08-28 | International Business Machines Corporation | Structure and method for reading/writing signature commands from/to a plurality of controller pairs |
JP2000089916A (en) * | 1998-09-11 | 2000-03-31 | Fujitsu Ltd | File controller |
US6351829B1 (en) * | 1998-10-28 | 2002-02-26 | Honeywell Inc | System and method for distinguishing a device failure from an inter-device communication failure |
US6366965B1 (en) * | 1998-12-31 | 2002-04-02 | Lsi Logic Corporation | System for updating an identity of a device in response to determining that a neighbor device has an identity different from one stored therein |
US6317839B1 (en) * | 1999-01-19 | 2001-11-13 | International Business Machines Corporation | Method of and apparatus for controlling supply of power to a peripheral device in a computer system |
US6442709B1 (en) * | 1999-02-09 | 2002-08-27 | International Business Machines Corporation | System and method for simulating disaster situations on peer to peer remote copy machines |
US6725392B1 (en) | 1999-03-03 | 2004-04-20 | Adaptec, Inc. | Controller fault recovery system for a distributed file system |
JP3837953B2 (en) * | 1999-03-12 | 2006-10-25 | 株式会社日立製作所 | Computer system |
EP1077400B1 (en) * | 1999-08-17 | 2007-07-18 | Hewlett-Packard Company, A Delaware Corporation | Transparent status interface |
US6487680B1 (en) * | 1999-12-03 | 2002-11-26 | International Business Machines Corporation | System, apparatus, and method for managing a data storage system in an n-way active controller configuration |
JP4054509B2 (en) * | 2000-04-19 | 2008-02-27 | 株式会社東芝 | Field device control system and computer-readable storage medium |
US6393539B1 (en) | 2000-05-04 | 2002-05-21 | Dell Products, L.P. | System and method for reliably assigning and protecting data in a centralizes storage system |
US6389559B1 (en) * | 2000-05-24 | 2002-05-14 | Mti Technology Corporation | Controller fail-over without device bring-up |
WO2002033551A1 (en) * | 2000-10-18 | 2002-04-25 | Tricord Systems, Inc. | Controller fault recovery system for a distributed file system |
US20020104008A1 (en) * | 2000-11-30 | 2002-08-01 | Cochran Robert A. | Method and system for securing control-device-lun-mediated access to luns provided by a mass storage device |
US6754853B1 (en) * | 2000-12-07 | 2004-06-22 | Lsi Logic Corporation | Testing components of a computerized storage network system having a storage unit with multiple controllers |
US7054927B2 (en) | 2001-01-29 | 2006-05-30 | Adaptec, Inc. | File system metadata describing server directory information |
US20020138559A1 (en) * | 2001-01-29 | 2002-09-26 | Ulrich Thomas R. | Dynamically distributed file system |
US20020124137A1 (en) * | 2001-01-29 | 2002-09-05 | Ulrich Thomas R. | Enhancing disk array performance via variable parity based load balancing |
US6862692B2 (en) | 2001-01-29 | 2005-03-01 | Adaptec, Inc. | Dynamic redistribution of parity groups |
US6990667B2 (en) | 2001-01-29 | 2006-01-24 | Adaptec, Inc. | Server-independent object positioning for load balancing drives and servers |
US6775792B2 (en) | 2001-01-29 | 2004-08-10 | Snap Appliance, Inc. | Discrete mapping of parity blocks |
US6990547B2 (en) * | 2001-01-29 | 2006-01-24 | Adaptec, Inc. | Replacing file system processors by hot swapping |
US6820212B2 (en) * | 2001-02-20 | 2004-11-16 | Digi-Data Corporation | RAID system having channel capacity unaffected by any single component failure |
US6802023B2 (en) * | 2001-03-15 | 2004-10-05 | Hewlett-Packard Development Company, L.P. | Redundant controller data storage system having hot insertion system and method |
US6513098B2 (en) * | 2001-05-25 | 2003-01-28 | Adaptec, Inc. | Method and apparatus for scalable error correction code generation performance |
US6874100B2 (en) * | 2001-07-12 | 2005-03-29 | Digi-Data Corporation | Raid system with multiple controllers and proof against any single point of failure |
US6996670B2 (en) * | 2001-10-05 | 2006-02-07 | International Business Machines Corporation | Storage area network methods and apparatus with file system extension |
US7506127B2 (en) * | 2001-12-21 | 2009-03-17 | Network Appliance, Inc. | Reconfiguration of storage system including multiple mass storage devices |
US7127798B1 (en) | 2003-04-04 | 2006-10-31 | Network Appliance Inc. | Method for converting disk drive storage enclosure into a standalone network storage system |
US7093043B2 (en) * | 2001-12-27 | 2006-08-15 | Hewlett-Packard Development Company, L.P. | Data array having redundancy messaging between array controllers over the host bus |
US7406039B2 (en) * | 2002-02-26 | 2008-07-29 | Dell Products L.P. | System and method for a failover protocol in storage area network controllers |
JP2003288279A (en) * | 2002-03-28 | 2003-10-10 | Hitachi Ltd | Fault detection method between storage device and storage device used for it |
US7318116B2 (en) * | 2002-11-08 | 2008-01-08 | International Business Machines Corporation | Control path failover in an automated data storage library |
US7383381B1 (en) | 2003-02-28 | 2008-06-03 | Sun Microsystems, Inc. | Systems and methods for configuring a storage virtualization environment |
US7236987B1 (en) | 2003-02-28 | 2007-06-26 | Sun Microsystems Inc. | Systems and methods for providing a storage virtualization environment |
US7430568B1 (en) | 2003-02-28 | 2008-09-30 | Sun Microsystems, Inc. | Systems and methods for providing snapshot capabilities in a storage virtualization environment |
US7290168B1 (en) | 2003-02-28 | 2007-10-30 | Sun Microsystems, Inc. | Systems and methods for providing a multi-path network switch system |
DE10313318A1 (en) * | 2003-03-25 | 2004-10-21 | Giesecke & Devrient Gmbh | Controlled execution of a program intended for a virtual machine on a portable data carrier |
US7646705B2 (en) * | 2003-04-11 | 2010-01-12 | International Business Machines Corporation | Minimizing data loss chances during controller switching |
US20050125538A1 (en) * | 2003-12-03 | 2005-06-09 | Dell Products L.P. | Assigning logical storage units to host computers |
US7340639B1 (en) | 2004-01-08 | 2008-03-04 | Network Appliance, Inc. | System and method for proxying data access commands in a clustered storage system |
US7290170B2 (en) * | 2004-04-07 | 2007-10-30 | International Business Machines Corporation | Arbitration method and system for redundant controllers, with output interlock and automatic switching capabilities |
TWI306241B (en) * | 2004-07-12 | 2009-02-11 | Infortrend Technology Inc | A controller capable of self-monitoring, a redundant storage system having the same, and its method |
JP4305336B2 (en) * | 2004-09-02 | 2009-07-29 | 日本電気株式会社 | Disk enclosure device and program thereof |
US8019842B1 (en) | 2005-01-27 | 2011-09-13 | Netapp, Inc. | System and method for distributing enclosure services data to coordinate shared storage |
US8180855B2 (en) | 2005-01-27 | 2012-05-15 | Netapp, Inc. | Coordinated shared storage architecture |
US20060174085A1 (en) * | 2005-01-28 | 2006-08-03 | Dell Products L.P. | Storage enclosure and method for the automated configuration of a storage enclosure |
US20090106461A1 (en) * | 2005-01-31 | 2009-04-23 | Yokogawa Electric Corporation | Information Processing Apparatus and Information Processing Method |
US8073899B2 (en) * | 2005-04-29 | 2011-12-06 | Netapp, Inc. | System and method for proxying data access commands in a storage system cluster |
JP4409483B2 (en) * | 2005-06-30 | 2010-02-03 | 富士通株式会社 | Storage system, storage control device, and storage control method |
US20070022314A1 (en) * | 2005-07-22 | 2007-01-25 | Pranoop Erasani | Architecture and method for configuring a simplified cluster over a network with fencing and quorum |
US20070088917A1 (en) * | 2005-10-14 | 2007-04-19 | Ranaweera Samantha L | System and method for creating and maintaining a logical serial attached SCSI communication channel among a plurality of storage systems |
US8484365B1 (en) | 2005-10-20 | 2013-07-09 | Netapp, Inc. | System and method for providing a unified iSCSI target with a plurality of loosely coupled iSCSI front ends |
US7797570B2 (en) | 2005-11-29 | 2010-09-14 | Netapp, Inc. | System and method for failover of iSCSI target portal groups in a cluster environment |
US8560503B1 (en) * | 2006-01-26 | 2013-10-15 | Netapp, Inc. | Content addressable storage system |
US7516352B2 (en) * | 2006-03-21 | 2009-04-07 | International Business Machines Corporation | Isolating a drive from disk array for diagnostic operations |
JP5057755B2 (en) * | 2006-11-28 | 2012-10-24 | 株式会社日立製作所 | Storage control device and control method thereof |
US7613947B1 (en) | 2006-11-30 | 2009-11-03 | Netapp, Inc. | System and method for storage takeover |
US7661017B2 (en) * | 2007-01-30 | 2010-02-09 | International Business Machines Corporaion | Diagnostic operations within a switched fibre channel arbitrated loop system |
JP5148236B2 (en) | 2007-10-01 | 2013-02-20 | ルネサスエレクトロニクス株式会社 | Semiconductor integrated circuit and method for controlling semiconductor integrated circuit |
TWI384471B (en) * | 2008-04-03 | 2013-02-01 | Inventec Corp | A disk level examining system for a dual-host redundant storage system and method thereof |
US7849167B2 (en) * | 2008-04-15 | 2010-12-07 | International Business Machines Corporation | Dynamic distributed adjustment of maximum use of a shared storage resource |
US8271706B2 (en) * | 2008-05-22 | 2012-09-18 | International Business Machines Corporation | Stabilization of host to storage subsystem ownership |
US20100050021A1 (en) * | 2008-08-25 | 2010-02-25 | International Business Machines Corporation | Error code handling in a storage subsystem |
US9377960B2 (en) * | 2009-07-29 | 2016-06-28 | Hgst Technologies Santa Ana, Inc. | System and method of using stripes for recovering data in a flash storage system |
JP2011170589A (en) * | 2010-02-18 | 2011-09-01 | Nec Corp | Storage control device, storage device, and storage control method |
US8381027B1 (en) * | 2010-09-28 | 2013-02-19 | Emc Corporation | Determining alternate paths in faulted systems |
JP2012247817A (en) * | 2011-05-25 | 2012-12-13 | Fujitsu Ltd | Storage controller, storage apparatus, and patrol diagnosis method |
US8984222B2 (en) * | 2011-09-09 | 2015-03-17 | Lsi Corporation | Methods and structure for task management in storage controllers of a clustered storage system |
WO2013115806A1 (en) * | 2012-01-31 | 2013-08-08 | Hewlett-Packard Development Company , L.P. | Drivers and controllers |
US9130971B2 (en) | 2012-05-15 | 2015-09-08 | Splunk, Inc. | Site-based search affinity |
US10387448B2 (en) | 2012-05-15 | 2019-08-20 | Splunk Inc. | Replication of summary data in a clustered computing environment |
US9124612B2 (en) | 2012-05-15 | 2015-09-01 | Splunk Inc. | Multi-site clustering |
US11003687B2 (en) | 2012-05-15 | 2021-05-11 | Splunk, Inc. | Executing data searches using generation identifiers |
US8788459B2 (en) * | 2012-05-15 | 2014-07-22 | Splunk Inc. | Clustering for high availability and disaster recovery |
US9367412B2 (en) * | 2012-06-25 | 2016-06-14 | Netapp, Inc. | Non-disruptive controller replacement in network storage systems |
KR101704787B1 (en) * | 2014-12-31 | 2017-02-22 | 주식회사 효성 | Dual apparatus for controller |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5166939A (en) * | 1990-03-02 | 1992-11-24 | Micro Technology, Inc. | Data storage apparatus and method |
WO1993018456A1 (en) * | 1992-03-13 | 1993-09-16 | Emc Corporation | Multiple controller sharing in a redundant storage array |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4007448A (en) * | 1974-08-15 | 1977-02-08 | Digital Equipment Corporation | Drive for connection to multiple controllers in a digital data secondary storage facility |
US5195100A (en) * | 1990-03-02 | 1993-03-16 | Micro Technology, Inc. | Non-volatile memory storage of write operation identifier in data sotrage device |
US5274799A (en) * | 1991-01-04 | 1993-12-28 | Array Technology Corporation | Storage device array architecture with copyback cache |
US5379417A (en) * | 1991-11-25 | 1995-01-03 | Tandem Computers Incorporated | System and method for ensuring write data integrity in a redundant array data storage system |
JP3160106B2 (en) * | 1991-12-23 | 2001-04-23 | ヒュンダイ エレクトロニクス アメリカ | How to sort disk arrays |
US5239632A (en) * | 1992-04-16 | 1993-08-24 | Hewlett-Packard Company | Device to translate logical unit number communications on one SCSI bus to ID communications on a subordinate SCSI bus |
US5459857A (en) * | 1992-05-15 | 1995-10-17 | Storage Technology Corporation | Fault tolerant disk array data storage subsystem |
US5546558A (en) * | 1994-06-07 | 1996-08-13 | Hewlett-Packard Company | Memory system with hierarchic disk array and memory map store for persistent storage of virtual mapping information |
US5504882A (en) * | 1994-06-20 | 1996-04-02 | International Business Machines Corporation | Fault tolerant data storage subsystem employing hierarchically arranged controllers |
US5790775A (en) * | 1995-10-23 | 1998-08-04 | Digital Equipment Corporation | Host transparent storage controller failover/failback of SCSI targets and associated units |
US5790773A (en) * | 1995-12-29 | 1998-08-04 | Symbios, Inc. | Method and apparatus for generating snapshot copies for data backup in a raid subsystem |
US5895493A (en) * | 1997-06-30 | 1999-04-20 | Lsi Logic Corporation | Method and apparatus for storage of multiple host storage management information on a storage subsystem |
-
1997
- 1997-09-30 US US08/941,894 patent/US5975738A/en not_active Expired - Lifetime
-
1998
- 1998-09-29 AU AU95114/98A patent/AU9511498A/en not_active Abandoned
- 1998-09-29 WO PCT/US1998/020347 patent/WO1999017204A1/en active IP Right Grant
- 1998-09-29 DE DE69802833T patent/DE69802833T2/en not_active Expired - Lifetime
- 1998-09-29 EP EP98948569A patent/EP1019823B1/en not_active Expired - Lifetime
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5166939A (en) * | 1990-03-02 | 1992-11-24 | Micro Technology, Inc. | Data storage apparatus and method |
WO1993018456A1 (en) * | 1992-03-13 | 1993-09-16 | Emc Corporation | Multiple controller sharing in a redundant storage array |
Cited By (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6601128B1 (en) | 2000-05-02 | 2003-07-29 | International Business Machines Corporation | Method, system, program, and data structure for selecting a preferred path to a storage device |
GB2366048A (en) * | 2000-05-02 | 2002-02-27 | Ibm | Selecting a preferred path to a storage device |
GB2366048B (en) * | 2000-05-02 | 2004-11-10 | Ibm | Selecting a preferred path to a storage device |
EP1707205A2 (en) | 2002-07-09 | 2006-10-04 | Boehringer Ingelheim Pharma GmbH & Co. KG | Pharmaceutical compositions of anticholinergics and p38 kinase inhibitors in the treatment of respiratory diseases |
US7130970B2 (en) | 2002-09-09 | 2006-10-31 | Commvault Systems, Inc. | Dynamic storage device pooling in a computer system |
US9170890B2 (en) | 2002-09-16 | 2015-10-27 | Commvault Systems, Inc. | Combined stream auxiliary copy system and method |
US9940043B2 (en) | 2003-04-03 | 2018-04-10 | Commvault Systems, Inc. | Systems and methods for performing storage operations in a computer network |
US9251190B2 (en) | 2003-04-03 | 2016-02-02 | Commvault Systems, Inc. | System and method for sharing media in a computer network |
US8510516B2 (en) | 2003-04-03 | 2013-08-13 | Commvault Systems, Inc. | Systems and methods for sharing media in a computer network |
US8892826B2 (en) | 2003-04-03 | 2014-11-18 | Commvault Systems, Inc. | Systems and methods for performing storage operations in a computer network |
US9021213B2 (en) | 2003-04-03 | 2015-04-28 | Commvault Systems, Inc. | System and method for sharing media in a computer network |
US9201917B2 (en) | 2003-04-03 | 2015-12-01 | Commvault Systems, Inc. | Systems and methods for performing storage operations in a computer network |
EP2116245A2 (en) | 2004-08-07 | 2009-11-11 | Boehringer Ingelheim International GmbH | EGFR kinase inhibitor combinations for treating respiratory and gastrointestinal disorders |
US9507525B2 (en) | 2004-11-05 | 2016-11-29 | Commvault Systems, Inc. | Methods and system of pooling storage devices |
US10191675B2 (en) | 2004-11-05 | 2019-01-29 | Commvault Systems, Inc. | Methods and system of pooling secondary storage devices |
EP2384751A1 (en) | 2004-12-24 | 2011-11-09 | Boehringer Ingelheim International Gmbh | Medicaments for the treatment or prevention of fibrotic diseases |
EP2878297A1 (en) | 2004-12-24 | 2015-06-03 | Boehringer Ingelheim International GmbH | Medicaments for the treatment or prevention of fibrotic diseases |
EP1724660A1 (en) * | 2005-04-27 | 2006-11-22 | Hitachi, Ltd. | Disk array device |
US9122600B2 (en) | 2006-12-22 | 2015-09-01 | Commvault Systems, Inc. | Systems and methods for remote monitoring in a computer network |
US11175982B2 (en) | 2006-12-22 | 2021-11-16 | Commvault Systems, Inc. | Remote monitoring and error correcting within a data storage system |
US11416328B2 (en) | 2006-12-22 | 2022-08-16 | Commvault Systems, Inc. | Remote monitoring and error correcting within a data storage system |
US10671472B2 (en) | 2006-12-22 | 2020-06-02 | Commvault Systems, Inc. | Systems and methods for remote monitoring in a computer network |
WO2008142031A1 (en) | 2007-05-18 | 2008-11-27 | Institut Curie | P38alpha as a therapeutic target in bladder carcinoma |
US8639877B2 (en) | 2009-06-30 | 2014-01-28 | International Business Machines Corporation | Wear leveling of solid state disks distributed in a plurality of redundant array of independent disk ranks |
US8234520B2 (en) | 2009-09-16 | 2012-07-31 | International Business Machines Corporation | Wear leveling of solid state disks based on usage information of data and parity received from a raid controller |
US8510595B2 (en) | 2009-09-16 | 2013-08-13 | International Business Machines Corporation | Wear leveling of solid state disks based on usage information of data and parity received from a raid controller |
US10379988B2 (en) | 2012-12-21 | 2019-08-13 | Commvault Systems, Inc. | Systems and methods for performance monitoring |
US10168931B2 (en) | 2015-01-23 | 2019-01-01 | Commvault Systems, Inc. | Scalable auxiliary copy processing in a data storage management system using media agent resources |
US10346069B2 (en) | 2015-01-23 | 2019-07-09 | Commvault Systems, Inc. | Scalable auxiliary copy processing in a data storage management system using media agent resources |
US9904481B2 (en) | 2015-01-23 | 2018-02-27 | Commvault Systems, Inc. | Scalable auxiliary copy processing in a storage management system using media agent resources |
US10996866B2 (en) | 2015-01-23 | 2021-05-04 | Commvault Systems, Inc. | Scalable auxiliary copy processing in a data storage management system using media agent resources |
US9898213B2 (en) | 2015-01-23 | 2018-02-20 | Commvault Systems, Inc. | Scalable auxiliary copy processing using media agent resources |
US11513696B2 (en) | 2015-01-23 | 2022-11-29 | Commvault Systems, Inc. | Scalable auxiliary copy processing in a data storage management system using media agent resources |
US11010261B2 (en) | 2017-03-31 | 2021-05-18 | Commvault Systems, Inc. | Dynamically allocating streams during restoration of data |
US11615002B2 (en) | 2017-03-31 | 2023-03-28 | Commvault Systems, Inc. | Dynamically allocating streams during restoration of data |
US11593223B1 (en) | 2021-09-02 | 2023-02-28 | Commvault Systems, Inc. | Using resource pool administrative entities in a data storage management system to provide shared infrastructure to tenants |
US11928031B2 (en) | 2021-09-02 | 2024-03-12 | Commvault Systems, Inc. | Using resource pool administrative entities to provide shared infrastructure to tenants |
Also Published As
Publication number | Publication date |
---|---|
DE69802833D1 (en) | 2002-01-17 |
DE69802833T2 (en) | 2002-06-27 |
EP1019823A1 (en) | 2000-07-19 |
EP1019823B1 (en) | 2001-12-05 |
US5975738A (en) | 1999-11-02 |
AU9511498A (en) | 1999-04-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5975738A (en) | Method for detecting failure in redundant controllers using a private LUN | |
US5822782A (en) | Methods and structure to maintain raid configuration information on disks of the array | |
US5761705A (en) | Methods and structure for maintaining cache consistency in a RAID controller having redundant caches | |
US6802023B2 (en) | Redundant controller data storage system having hot insertion system and method | |
US6715101B2 (en) | Redundant controller data storage system having an on-line controller removal system and method | |
JP3732440B2 (en) | Method and apparatus for maintaining consistency of data stored in mirroring devices | |
US6708285B2 (en) | Redundant controller data storage system having system and method for handling controller resets | |
US5267242A (en) | Method and apparatus for substituting spare memory chip for malfunctioning memory chip with scrubbing | |
US5790775A (en) | Host transparent storage controller failover/failback of SCSI targets and associated units | |
US6052795A (en) | Recovery method and system for continued I/O processing upon a controller failure | |
US6754853B1 (en) | Testing components of a computerized storage network system having a storage unit with multiple controllers | |
US6854071B2 (en) | Method and apparatus for providing write recovery of faulty data in a non-redundant raid system | |
US7853831B2 (en) | Methods and structure for detection and handling of catastrophic SCSI errors | |
CN101105780A (en) | Storage control system, control method for storage control system, port selector, and controller | |
JPS5833575B2 (en) | How to recover data automatically | |
JP4132322B2 (en) | Storage control device and control method thereof | |
US6643734B2 (en) | Control device and control method for a disk array | |
US7076686B2 (en) | Hot swapping memory method and system | |
JP4076174B2 (en) | Handling node address failures in distributed node systems with processors | |
JPS63184146A (en) | Information processor | |
US7398416B2 (en) | Proactive driver response to an operating system if a drive is removed from a RAID configuration | |
JPH08171463A (en) | Data read method in disk array device, and disk array device | |
JPH0816484A (en) | Doubled memory system | |
KR100363221B1 (en) | A disk array system and method of providing fault tolerance for it | |
JPH10187355A (en) | Disk control system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GE GH GM HR HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 1998948569 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 1998948569 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
NENP | Non-entry into the national phase |
Ref country code: CA |
|
WWG | Wipo information: grant in national office |
Ref document number: 1998948569 Country of ref document: EP |