WO1999030233A1 - Computer bus data transfer reduction with parity circuit - Google Patents

Computer bus data transfer reduction with parity circuit Download PDF

Info

Publication number
WO1999030233A1
WO1999030233A1 PCT/US1997/022807 US9722807W WO9930233A1 WO 1999030233 A1 WO1999030233 A1 WO 1999030233A1 US 9722807 W US9722807 W US 9722807W WO 9930233 A1 WO9930233 A1 WO 9930233A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
parity
memory
disk
controller
Prior art date
Application number
PCT/US1997/022807
Other languages
French (fr)
Inventor
Rodney A. Dekoning
Dennis E. Gates
Charles D. Binford
Original Assignee
Symbios Logic Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Symbios Logic Inc. filed Critical Symbios Logic Inc.
Priority to PCT/US1997/022807 priority Critical patent/WO1999030233A1/en
Priority to AU55228/98A priority patent/AU5522898A/en
Publication of WO1999030233A1 publication Critical patent/WO1999030233A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1076Parity data used in redundant arrays of independent storages, e.g. in RAID systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2211/00Indexing scheme relating to details of data-processing equipment not covered by groups G06F3/00 - G06F13/00
    • G06F2211/10Indexing scheme relating to G06F11/10
    • G06F2211/1002Indexing scheme relating to G06F11/1076
    • G06F2211/1054Parity-fast hardware, i.e. dedicated fast hardware for RAID systems with parity

Definitions

  • the present invention relates generally to disk array controllers, and more particularly to a method and apparatus for reducing data transfers across a memory bus of a disk array controller.
  • RAID Redundant Array of Inexpensive Disks
  • a .RAID storage device typically includes an array of small, inexpensive hard disk drives, such as the 5 or Z ⁇ inch disk drives currently used in personal computers and workstations.
  • disk array devices have been available for several years, significant improvements in the reliability and performance of small disk drives, and a decline in the cost of such disk drives have resulted in an enhanced interest in RAID systems.
  • FIG. 1 there is shown a simplified block diagram of a known disk array controller 10 which is operatively coupled between a host device 31 and a RAID disk array 33.
  • the known controller 10 includes processor 12, host interface circuit 16, one or more input/output processors (IOP) 24, RAID Parity Assist (RPA) circuit 20, RPA memory 22, system bus 28 and memory bus 66.
  • the host device 31 is operatively coupled to the known controller 10 through the host interface 16, and the disk array 33 is operatively coupled to the controller 10 through the IOP 24.
  • Processor 12, host interface 16, IOP 24, and RPA circuit 20 are operatively interconnected through the system bus 28.
  • the RPA memory 22 is operatively coupled to the RPA circuit 20 through the memory bus 66 and is implemented with known DRAM memory circuits.
  • the known controller 10 receives write data from the host device 31 to store in the disk array 33.
  • write data is received by the controller 10 from the host device 31 and transferred across the system bus 28 and memory bus 66 for temporary storage in the RPA memory 22.
  • the processor 12 schedules the write data to be transferred from the RPA memory 22 across the memory bus 66 to the RPA circuit 20 in order to generate parity information for the write data.
  • the resulting parity information is transferred across the memory bus 66 for temporary storage in the RPA memory 22.
  • the processor 12 schedules the write data and the accompanying parity information to be written out of the RPA memory 22, across the memory bus 66 and system bus 28, and to the disk array 33 for storage therein.
  • approximately three units of data and two units of parity information are transferred across the memory bus 66.
  • approximately 14 chunks of data are transferred across the memory bus 66 for each transfer of 4 data chunks from the host device 31 to the disk array 33. That is, approximately 3.5 (14 chunks / 4 data chunks) data transfers occur across the memory bus 66 for each unit of data (4 chunks) transferred from the host device 31 to the disk array 33.
  • the controller 10 requests read data from the disk array 33 in response to a read request from the host device 31.
  • read data and parity information are received by the controller 10 from the functioning disk drives associated with the disk array 33 and transferred across the system bus 28 and memory bus 66 for temporary storage in the RPA memory 22.
  • the processor 12 schedules the read data and parity information from the RPA memory 22 to be transferred across the memory bus 66 to the RPA circuit 20 in order to recover the read data from the failed disk drive.
  • the recovered read data is transferred across the memory bus 66 for temporary storage in the RPA memory 22.
  • the processor 12 schedules the read data, including the read data recovered from the failed disk drive, to be transferred from the RPA memory 22, across the memory bus 66 and system bus 28, and to the host device 31.
  • approximately three units of data and one unit of parity information are transferred across the memory bus 66.
  • approximately 13 chunks of data are transferred across the memory bus 66 for each transfer of 4 data chunks between the host device 31 and the disk array 33.
  • a method for transferring data from a first device to a second device connected by a controller having a parity buffer and a memory having a first storage and a second storage includes the steps of transferring first data from the first device to the first storage; transferring second data from the first device to the second storage; transferring the first data to the second device and storing the first data in the parity buffer; and determining parity data from the second data and the first data stored in the parity buffer.
  • a method for transferring data from a first device to a second device connected by a controller having a parity buffer and a memory having a first storage and a second storage includes the steps of transferring first data from the first device to the first storage and storing the first data in the parity buffer; transferring second data from the first device to the second storage and determining parity data from the second data and the first data stored in the parity buffer; and transferring the first data to the second device.
  • a disk array controller for connecting a host device to a disk array having a first disk, a second disk, and a third disk.
  • the disk array controller includes a memory, a parity buffer, a parity generator, and a parity controller.
  • the parity generator is coupled to the parity buffer.
  • the parity controller is coupled to the memory and the parity generator.
  • the memory includes a first storage and a second storage.
  • the parity controller is configured to transfer first data from the first disk to the first storage, is configured to transfer second data from the second disk to the second storage, is configured to transfer the first data from the first storage to the host device and to the parity generator, and is configured to transfer the second data from the second storage to the parity generator.
  • the parity generator is configured to determine parity data from the second data and the first data stored in the parity buffer.
  • a disk array controller for connecting a host device to a disk array having a first disk, a second disk, and a third disk.
  • the disk array controller includes a memory, a parity buffer, a parity generator, and a parity controller.
  • the parity generator is coupled to the parity buffer.
  • the parity controller is coupled to the memory and the parity generator.
  • the memory includes a first storage and a second storage.
  • the parity controller is configured to transfer first data from the host device to the first storage and to the parity generator, is configured to transfer second data from the host device to the second storage and to the parity generator, is configured to transfer the first data from the first storage to the first disk, is configured to transfer the second data from the second storage to second disk, and is configured to transfer parity data from the parity generator to the third disk.
  • the parity generator is configured to store the first data in the parity buffer; and is configured to determine the parity data from the second data and the first data stored in the parity buffer.
  • FIG. 1 is a simplified block diagram of a known disk array controller
  • FIG. 2 is a block diagram of an improved RAID Parity Assist (RPA) circuit and an RPA memory of FIG. 1 which incorporates features of the present invention therein
  • FIG. 3 is a flowchart illustrating a write request routine which is executed by a disk array controller which incorporates features of the present invention therein;
  • FIG. 4 is a flowchart illustrating a degraded read request routine which is executed by a disk array controller which incorporates features of the present invention therein.
  • the RPA circuit 20 includes a RPA memory controller 60, exclusive-OR (XOR) engine 62, and intermediate parity buffer 64.
  • the RPA memory controller 60 is operatively coupled to the system bus 28 and is operatively coupled to the RPA memory 22 via a memory bus 66.
  • the RPA memory controller 60 is also operatively coupled to the intermediate parity buffer 64 via the XOR engine 62.
  • the RPA memory controller 60 controls (1) the flow of data between the system bus 28, the RPA memory 22, and the intermediate parity buffer 64, and (2) the operation of the XOR engine 62.
  • the XOR engine 62 primarily performs a bit-wise XOR operation on data stored in the intermediate parity buffer 64 with data received from the RPA memory controller 60 as described further below.
  • the XOR engine 62 may also store and/or retrieve data in the intermediate parity buffer 64 without performing an XOR operation thereon.
  • the RPA memory 22 permits the intermediate storage of blocks of read data (i.e. data read from the disk array 33 which is sent to the host device 31) and blocks of write data (i.e. data received from the host device 31 which is written to the disk array 33).
  • the RPA memory 22 provides a caching function wherein the disk array controller 10 may utilize read/write data stored in the RPA memory 22 instead of accessing one or more of the disk drives associated with the disk array 33.
  • the RPA memory 22 permits the disk array controller 10 to perform multiple read operations and write operations concurrently.
  • the RPA memory 22 may be implemented as Dynamic Random Access Memory (DRAM).
  • DRAM Dynamic Random Access Memory
  • the intermediate parity buffer 64 may define one or more buffer locations 64a-64d for storing intermediate parity information that is generated during read/write data transfers across the system bus 28 as described further below.
  • the multiple buffer locations 64a-64d permit the disk array controller 10 to concurrently perform multiple read/write data transfers to and from the disk array 33, as well as to perform read/write data transfers involving multiple blocks of data per each disk drive of the disk array 33.
  • the intermediate parity buffer 64 may be integrated with the RPA memory controller 60 and the XOR engine 62 into a single integrated circuit in order to maximize the data transfer rates therebetween. It should be appreciated that the size of an integrated intermediate parity buffer 64 is relatively small (e.g. 32 kilobytes) due to current technology and manufacturing costs. Larger integrated intermediate parity buffers 64, however, are contemplated with advances in the technology and may result in an increase in performance. Furthermore, intermediate parity buffer 64 may be implemented as a separate circuit which may result in a larger intermediate parity buffer 64. However, current manufacturing costs associated with providing a high speed interface between separate circuits currently make this option cost ineffective.
  • the disk array 33 may include five disk drives, and may be implemented as a RAID Level 3 disk array with four of the five disk drives used exclusively for the storage of data, and the fifth disk drive used exclusively for the storage of parity information.
  • the disk array 33 has a stripe size of 5 chunks with a first chunk of data (chunk 1) stored on a first data disk drive (disk 1), a second chunk of data (chunk 2) stored on a second data disk drive (disk 2), a third chunk of data (chunk 3) stored on a third data disk drive (disk 3), a fourth chunk of data (chunk 4) stored on a fourth data disk drive (disk 4), and a first chunk of parity information (parity chunk A) corresponding to chunks 1-4 of the stripe, stored on a fifth or parity disk drive of the disk array 33.
  • the minimum chunk size of a RAID Level 3 disk array is small compared to the user visible sector size of the disk array 33, so that all disk drives in the disk
  • the routine 100 may be initiated when the host device 31 requests that a stripe of data be written to the disk array 33.
  • the disk array controller 10 causes a first chunk of data (chunk 1) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64 (step 110).
  • chunk 1 is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA memory controller 60, across the memory bus 66 and into the RPA memory 22.
  • This chunk transfer like the others that follow may be directly controlled by the processor 12 or by a DMA engine which may be part of the RPA memory controller 60 or the IOP 24.
  • chunk 1 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller causes chunk 1 to be transferred to the XOR engine 62.
  • the XOR engine then causes chunk 1 to be stored in buffer location 64a of the intermediate parity buffer 64.
  • the disk array controller 10 then causes a second chunk of data (chunk 2) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64 (step 120).
  • chunk 2 is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA controller memory 60, across the memory bus 66 and into the RPA memory 22.
  • chunk 2 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller 60 also causes chunk 2 to be transferred to the XOR engine 65.
  • the XOR engine 62 then causes chunk 2 to be bit-wise XORed with the data stored in buffer location 64a of the intermediate parity buffer 64.
  • the disk array controller 10 causes a third chunk of data (chunk 3) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64.
  • Chunk 3 like chunk 1 and chunk 2 above, is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA controller memory 60, across the memory bus 66 and into the RPA memory 22. While chunk 3 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller 60 also causes chunk 3 to be transferred to the XOR engine 62. The XOR engine 62 then causes chunk 3 to be bit-wise XORed with the data stored in the buffer location 64a.
  • the disk array controller 10 in step 140 causes a fourth chunk of data (chunk 4) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64.
  • chunk 4 is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA controller memory 60, across the memory bus 66 and into the RPA memory 22. While chunk 4 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller 60 also causes chunk 4 to be transferred to the XOR engine 62. The XOR engine 62 then causes chunk 4 to be bit-wise XORed with the data stored in the buffer location 64a.
  • the buffer location 64a contains parity information (parity chunk A) for data chunks 1-4.
  • the RPA memory controller 60 causes parity chunk A to be transferred from the buffer location 64a to the RPA memory 22.
  • the RPA memory controller 60 causes the XOR engine 62 to retrieve parity chunk A from buffer location 64a and to send parity chunk A to the RPA memory controller 60.
  • the RPA memory controller 60 then transfers parity chunk A across the memory bus 66 for temporary storage in the RPA memory 22.
  • the processor 12 schedules a write operation for transferring the write data and parity information to the disk array 33 (step 160).
  • the processor 12 schedules chunk 1 to be transferred to disk 1 , chunk 2 to be transferred to disk 2, chunk 3 to be transferred to disk 3, chunk 4 to be transferred to disk 4, and parity chunk A to be transferred to the parity disk 5 of the disk array 33.
  • the write data (chunks 1-4) and parity chunk A are transferred out of the RPA memory 22, across the memory bus 66, through RPA memory controller 60, across the system bus 28, through the IOP 24 and to the respective disk drives associated with the disk array 33.
  • the write request routine 100 reduces the data traffic across the memory bus 66 when compared to a similar write request operation performed by the prior art.
  • the number of chunks (Y) that are transferred across the memory bus 66 when executing routine 100 may be determined from equation (1):
  • disk array controller 10 transfers ten chunks across the memory bus 66 in order to transfer a stripe of five chunks (four data chunks and one parity chunk). That is:
  • routine 100 provides an approximately 30% reduction (14 chunks - 10 chunks) / 14 chunks in data traffic across the memory bus as compared to the prior art. It should further be appreciated that if the size of the chunk is larger than each of the buffers 64a-64d of the intermediate parity buffer 64, then each chunk may be broken up into sub-chunks that are small enough to be stored within a buffer of the intermediate parity buffer 64. Once the chunks are broken up into sub-chunks, the above method 100 may be implemented in at least two different ways. One implementation is to use multiple buffers 64a-64d and transfer all the sub-chunks of a disk before moving on to the next disk. This method can viewed as creating from the buffers 64a-64d of the intermediate parity buffer 64 a buffer which is large enough to hold a chunk. Therefore, this implementation follows the method 100 except that it makes multiple transfers at each step.
  • Another implementation is to use only one of the buffers of the intermediate parity buffer 64.
  • This implementation transfers a sub-chunk from each disk as described by method 100 but skips step 160. This implementation repeats this process until all the sub-chunks and parity sub- chunks are in the RPA memory 22. Once all the sub-chunks and parity sub-chunks for the stripe are in the RPA memory 22, this implementation executes step 160 which transfers all the sub-chunks and parity sub-chunks for the stripe to the disk array 33.
  • a degraded read request routine 200 which may be executed by the disk array controller 10 when a disk drive associated with the disk array 33 has failed. For example, if disk 2 which contains chunk 2 has failed, the disk array controller 10 must obtain valid read data from the remaining three functioning data disk drives of the disk array 33 (disks 1 , 3 and 4) and from the parity disk drive (disk 5).
  • the routine 200 may be initiated when the host device 31 requests that a stripe be read from the disk array 33.
  • the processor 12 causes valid read data to be transferred from the disk array 33 to the RPA memory 22 (step 210).
  • the processor 12 causes data chunks 1 , 3, 4, and parity chunk A to be transferred from disks 1 , 3, 4, and 5, through the IOP 24, across the system bus 28, through the RPA memory controller 60, across the memory bus 66 and into the RPA memory 22.
  • the disk array controller 10 causes chunk 1 to be transferred to the host device 31 and to the intermediate parity buffer 64 (step 220).
  • chunk 1 is transferred out of the RPA memory 22, across the memory bus 66, through the RPA memory controller 60, across the system bus 28, and through the host interface circuit 16 to the host device 31. While chunk 1 is transferred through the RPA memory controller 60 to the host device 31, RPA memory controller 60 also causes chunk 1 to be transferred to the XOR engine 62. The XOR engine 62 then stores chunk 1 in buffer location 64a of intermediate parity buffer 64.
  • the disk array controller 10 then causes chunk 3 to be transferred to the host device 31 and to the intermediate parity buffer 64.
  • chunk 3 is transferred from the RPA memory 22 across the memory bus 66, through the RPA memory controller 60, across the system bus 28, and through the host interface circuit 16 to the host device 31.
  • chunk 3 is transferred through the RPA memory controller 60 to the host device 31
  • the RPA memory controller 60 also causes chunk 3 to be transferred to the XOR engine 62.
  • the XOR engine 62 then causes chunk 3 to be bit-wise XORed with the data stored in the buffer location 64a of the intermediate parity buffer 64.
  • the disk array controller 10 in step 240 causes chunk 4 to be transferred to the host device 31 and to the intermediate parity buffer 64.
  • chunk 4 is transferred from the RPA memory 22 across the memory bus 66, through the RPA memory controller 60, across the system bus 28, and through the host interface circuit 16 to the host device 31.
  • the RPA memory controller 60 also causes chunk 4 to be transferred to the XOR engine 22.
  • the XOR engine then causes chunk 4 to be bit-wise XORed with the data stored in the buffer location 64a of the intermediate parity buffer 64.
  • step 250 the disk array controller 10 causes chunk 2 to be recovered and sent to the host device 31.
  • the RPA memory controller 60 causes the parity chunk A stored in the RPA memory 22 to be transferred across the memory bus 66, through the RPA memory controller 60, and to the XOR engine 62.
  • the XOR engine 62 then causes the parity chunk A to be bit-wise XORed with the data stored in buffer location 64a.
  • the buffer location 64a contains a restored copy of chunk 2 which was stored on the failed disk 2.
  • the RPA memory controller 60 causes the XOR engine 62 to retrieve data chunk 2 stored in buffer location 64a and to send the retrieved data chunk 2 to the RPA memory controller 60.
  • the RPA memory controller 60 then sends chunk 2 across the system bus 28 and through the host interface circuit 16 to the host device 31.
  • the degraded read request routine 200 reduces the data traffic across the memory bus 66 when compared to a similar degraded read request routine performed by the prior art.
  • the number of chunks (Y) that are transferred across the memory bus 66 when executing the routine 200 may be determined from equation (2):
  • disk array controller 10 in a degraded read operation transfers nine chunks across the memory bus 66 in order to read a stripe from the disk array 33 having four data disk drives and one parity disk drive. That is:
  • routine 200 provided an approximately 30% reduction (13 chunks -9 chunks) / 13 chunks in data traffic across the memory bus as compared to the prior art.
  • the RPA memory controller 60 may also write the restored chunk 2 data into the RPA memory 22 for subsequent use during a disk caching operation. If the RPA memory controller 60 is implemented to write the restored chunk 2 data into the RPA memory 22, then the number of data transfers as a result of a degraded read operation may be determined from above equation (1) instead of above equation (2).
  • the data requested by the host device 31 may be transferred to the host device 31 out of order (i.e. chunk 1 , chunk 3, chunk 4, and then chunk 2) depending upon which disk drive of disk array 33 fails.
  • a data transfer protocol may be required which permits out of order transfers of data chunks.
  • the degraded read operation is performed in a similar manner as the write operation described above (i.e. as chunks are transferred from the disk array 33 to the RPA memory 22, they are XORed with data in a buffer of the intermediate parity buffer 64), then the data may be transferred to the host device 31 in order. It should be appreciated, however, that this in order implementation of degraded read operations increases the complexity and cost of the RPA circuit 20.
  • the read data is transferred from the disk array 33 through the IOP 24, across the system bus 28, through the RPA memory controller 60, across the memory bus 66, and into the RPA memory 22.
  • the processor 12 then schedules a read operation where the chunks of read data are transferred out of the RPA memory 22, across the memory bus 66, through the RPA memory controller 60, across the system bus 28, through the host interface 16 and to the host device 31.

Abstract

A method and apparatus for transferring data from a first device (31) to a second device (33) connected by a controller (60) having a parity buffer (64) and a memory (22) having a first storage and a second storage is disclosed. The method includes the steps of transferring first data from the first device to the first storage; transferring second data from the first device to the second storage; transferring the first data to the second device and storing the first data in the parity buffer; and determining parity data from the second data and the first data stored in the parity buffer.

Description

COMPUTER BUS DATA TRANSFER REDUCTION WITH PARITY C IRCUIT
1. Technical Field
The present invention relates generally to disk array controllers, and more particularly to a method and apparatus for reducing data transfers across a memory bus of a disk array controller.
RAID (Redundant Array of Inexpensive Disks) storage devices have emerged as an alternative to large, expensive disk drives for use with present and future computer system architectures. A .RAID storage device typically includes an array of small, inexpensive hard disk drives, such as the 5 or ZΛ inch disk drives currently used in personal computers and workstations. Although disk array devices have been available for several years, significant improvements in the reliability and performance of small disk drives, and a decline in the cost of such disk drives have resulted in an enhanced interest in RAID systems.
Current RAID disk array design alternatives are described in a publication entitled "The RAIDbook: A Source Book for RAID Technology" third edition, May, 1994 which is available from the RAID Advisory Board, and which is incorporated herein by reference. The publication discusses the different levels of RAID disk arrays and the improvements in performance, reliability, power consumption and scale-ability that RAID disk arrays provide in comparison to single large magnetic disks. The present invention is primarily directed to improvements in the operation of controllers for RAID Level 3, 4 and 5 disk array devices, and to any other disk array devices which determine parity information dependent upon the data to be written to or read from the disk array.
Referring now to FIG. 1 , there is shown a simplified block diagram of a known disk array controller 10 which is operatively coupled between a host device 31 and a RAID disk array 33. The known controller 10 includes processor 12, host interface circuit 16, one or more input/output processors (IOP) 24, RAID Parity Assist (RPA) circuit 20, RPA memory 22, system bus 28 and memory bus 66. The host device 31 is operatively coupled to the known controller 10 through the host interface 16, and the disk array 33 is operatively coupled to the controller 10 through the IOP 24. Processor 12, host interface 16, IOP 24, and RPA circuit 20 are operatively interconnected through the system bus 28. Furthermore, the RPA memory 22 is operatively coupled to the RPA circuit 20 through the memory bus 66 and is implemented with known DRAM memory circuits.
During write operations, the known controller 10 receives write data from the host device 31 to store in the disk array 33. In particular, write data is received by the controller 10 from the host device 31 and transferred across the system bus 28 and memory bus 66 for temporary storage in the RPA memory 22. The processor 12 schedules the write data to be transferred from the RPA memory 22 across the memory bus 66 to the RPA circuit 20 in order to generate parity information for the write data. The resulting parity information is transferred across the memory bus 66 for temporary storage in the RPA memory 22. Finally, the processor 12 schedules the write data and the accompanying parity information to be written out of the RPA memory 22, across the memory bus 66 and system bus 28, and to the disk array 33 for storage therein.
Thus, during write operations, for one unit of data transferred to the disk array 33, approximately three units of data and two units of parity information are transferred across the memory bus 66. For example, in a RAID Level 3 disk array 33 having a stripe size of 5 chunks (4 data chunks and 1 parity chunk where a chunk is defined as the amount of contiguous storage of an array virtual disk that is mapped to contiguous storage on a single member disk of the disk array), approximately 14 chunks of data are transferred across the memory bus 66 for each transfer of 4 data chunks from the host device 31 to the disk array 33. That is, approximately 3.5 (14 chunks / 4 data chunks) data transfers occur across the memory bus 66 for each unit of data (4 chunks) transferred from the host device 31 to the disk array 33. During degraded read operations (i.e. a read operation where one of the disk drives associated with the disk array 33 has failed in some respect), the controller 10 requests read data from the disk array 33 in response to a read request from the host device 31. In particular, read data and parity information are received by the controller 10 from the functioning disk drives associated with the disk array 33 and transferred across the system bus 28 and memory bus 66 for temporary storage in the RPA memory 22. The processor 12 schedules the read data and parity information from the RPA memory 22 to be transferred across the memory bus 66 to the RPA circuit 20 in order to recover the read data from the failed disk drive. The recovered read data is transferred across the memory bus 66 for temporary storage in the RPA memory 22. Finally, the processor 12 schedules the read data, including the read data recovered from the failed disk drive, to be transferred from the RPA memory 22, across the memory bus 66 and system bus 28, and to the host device 31.
Thus, during degraded read operations, for one unit of data transferred from the disk array 33, approximately three units of data and one unit of parity information are transferred across the memory bus 66. For example, in a RAID Level 3 disk array 33 having a stripe size of 5 chunks (4 data chunks and 1 parity chunk, where a chunk is defined as the amount of contiguous storage of an array virtual disk that is mapped to contiguous storage on a single member disk of the disk array), approximately 13 chunks of data are transferred across the memory bus 66 for each transfer of 4 data chunks between the host device 31 and the disk array 33. That is, approximately 3.25 (13 chunks / 4 data chunks) data transfers occur across the memory bus 66 for each unit of data (4 chunks) transferred from disk array 33 to the host device 31. This multiplicity of data transfers back and forth across the memory bus 66 for each desired unit of data transfer may create a bottleneck condition in the known controller 10. This bottleneck condition may greatly reduce the data throughput of the known controller 10. What is needed therefore, is a method and apparatus which improves the throughput of a disk array controller by reducing the number of data transfers across the memory bus of the controller per unit of data transferred between a host device and a disk array.
2. Summary of the Invention In accordance with one embodiment of the present invention, there is provided a method for transferring data from a first device to a second device connected by a controller having a parity buffer and a memory having a first storage and a second storage. The method includes the steps of transferring first data from the first device to the first storage; transferring second data from the first device to the second storage; transferring the first data to the second device and storing the first data in the parity buffer; and determining parity data from the second data and the first data stored in the parity buffer. Pursuant to another embodiment of the present invention, there is provided a method for transferring data from a first device to a second device connected by a controller having a parity buffer and a memory having a first storage and a second storage. The method includes the steps of transferring first data from the first device to the first storage and storing the first data in the parity buffer; transferring second data from the first device to the second storage and determining parity data from the second data and the first data stored in the parity buffer; and transferring the first data to the second device.
Pursuant to yet another embodiment of the present invention, there is provided a disk array controller for connecting a host device to a disk array having a first disk, a second disk, and a third disk. The disk array controller includes a memory, a parity buffer, a parity generator, and a parity controller. The parity generator is coupled to the parity buffer. The parity controller is coupled to the memory and the parity generator. The memory includes a first storage and a second storage. The parity controller is configured to transfer first data from the first disk to the first storage, is configured to transfer second data from the second disk to the second storage, is configured to transfer the first data from the first storage to the host device and to the parity generator, and is configured to transfer the second data from the second storage to the parity generator. Furthermore, the parity generator is configured to determine parity data from the second data and the first data stored in the parity buffer.
In yet another embodiment of the present invention, there is provided a disk array controller for connecting a host device to a disk array having a first disk, a second disk, and a third disk. The disk array controller includes a memory, a parity buffer, a parity generator, and a parity controller. The parity generator is coupled to the parity buffer. The parity controller is coupled to the memory and the parity generator. The memory includes a first storage and a second storage. The parity controller is configured to transfer first data from the host device to the first storage and to the parity generator, is configured to transfer second data from the host device to the second storage and to the parity generator, is configured to transfer the first data from the first storage to the first disk, is configured to transfer the second data from the second storage to second disk, and is configured to transfer parity data from the parity generator to the third disk. Furthermore, the parity generator is configured to store the first data in the parity buffer; and is configured to determine the parity data from the second data and the first data stored in the parity buffer.
It is therefore an object of the present invention to provide a new and useful method for transferring data in a controller.
It is therefore an object of the present invention to provide an improved method for transferring data in a controller. It is another object of the present invention to provide a new and useful controller.
It is another object of the present invention to provide an improved controller.
It is a further object of the present invention to provide a controller with enhanced throughput of disk array data transfers.
It is yet another object of this invention to provide a method for enhancing throughput of disk array data transfers in a controller.
The above and other objects, features, and advantages of the present invention will become apparent from the following description and the attached drawings.
3. Brief Description of the Drawings
FIG. 1 is a simplified block diagram of a known disk array controller; FIG. 2 is a block diagram of an improved RAID Parity Assist (RPA) circuit and an RPA memory of FIG. 1 which incorporates features of the present invention therein; FIG. 3 is a flowchart illustrating a write request routine which is executed by a disk array controller which incorporates features of the present invention therein; and
FIG. 4 is a flowchart illustrating a degraded read request routine which is executed by a disk array controller which incorporates features of the present invention therein.
4. Detailed Description of the Preferred Embodiment
While the invention is susceptible to various modifications and alternative forms, a specific embodiment thereof has been shown by way of example in the drawings and will herein be described in detail. Furthermore, while the present invention is primarily described in reference to a RAID Level 3 disk array device, the present invention is directed to improvements in the operation of controllers for RAID Level 3, 4 and 5 disk array devices, and to any other disk array devices which determine parity information dependent upon the data to be written to or read from the disk array. It should be understood, however, that there is no intent to limit the invention to the particular form disclosed, but on the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention as defined by the appended claims.
Referring now to FIG. 2, there is shown a block diagram of an improved RPA circuit 20 for the disk array controller of FIG. 1 which incorporates the features of the present invention therein. The RPA circuit 20 includes a RPA memory controller 60, exclusive-OR (XOR) engine 62, and intermediate parity buffer 64. The RPA memory controller 60 is operatively coupled to the system bus 28 and is operatively coupled to the RPA memory 22 via a memory bus 66. The RPA memory controller 60 is also operatively coupled to the intermediate parity buffer 64 via the XOR engine 62. The RPA memory controller 60 controls (1) the flow of data between the system bus 28, the RPA memory 22, and the intermediate parity buffer 64, and (2) the operation of the XOR engine 62.
The XOR engine 62 primarily performs a bit-wise XOR operation on data stored in the intermediate parity buffer 64 with data received from the RPA memory controller 60 as described further below. The XOR engine 62 may also store and/or retrieve data in the intermediate parity buffer 64 without performing an XOR operation thereon.
The RPA memory 22 permits the intermediate storage of blocks of read data (i.e. data read from the disk array 33 which is sent to the host device 31) and blocks of write data (i.e. data received from the host device 31 which is written to the disk array 33). As a result, the RPA memory 22 provides a caching function wherein the disk array controller 10 may utilize read/write data stored in the RPA memory 22 instead of accessing one or more of the disk drives associated with the disk array 33. In addition, the RPA memory 22 permits the disk array controller 10 to perform multiple read operations and write operations concurrently. In the embodiment being described, the RPA memory 22 may be implemented as Dynamic Random Access Memory (DRAM). The intermediate parity buffer 64 may define one or more buffer locations 64a-64d for storing intermediate parity information that is generated during read/write data transfers across the system bus 28 as described further below. The multiple buffer locations 64a-64d permit the disk array controller 10 to concurrently perform multiple read/write data transfers to and from the disk array 33, as well as to perform read/write data transfers involving multiple blocks of data per each disk drive of the disk array 33.
In the embodiment being described, the intermediate parity buffer 64 may be integrated with the RPA memory controller 60 and the XOR engine 62 into a single integrated circuit in order to maximize the data transfer rates therebetween. It should be appreciated that the size of an integrated intermediate parity buffer 64 is relatively small (e.g. 32 kilobytes) due to current technology and manufacturing costs. Larger integrated intermediate parity buffers 64, however, are contemplated with advances in the technology and may result in an increase in performance. Furthermore, intermediate parity buffer 64 may be implemented as a separate circuit which may result in a larger intermediate parity buffer 64. However, current manufacturing costs associated with providing a high speed interface between separate circuits currently make this option cost ineffective.
In the embodiment being described, the disk array 33 may include five disk drives, and may be implemented as a RAID Level 3 disk array with four of the five disk drives used exclusively for the storage of data, and the fifth disk drive used exclusively for the storage of parity information. Thus, the disk array 33 has a stripe size of 5 chunks with a first chunk of data (chunk 1) stored on a first data disk drive (disk 1), a second chunk of data (chunk 2) stored on a second data disk drive (disk 2), a third chunk of data (chunk 3) stored on a third data disk drive (disk 3), a fourth chunk of data (chunk 4) stored on a fourth data disk drive (disk 4), and a first chunk of parity information (parity chunk A) corresponding to chunks 1-4 of the stripe, stored on a fifth or parity disk drive of the disk array 33. It should be appreciated that the minimum chunk size of a RAID Level 3 disk array is small compared to the user visible sector size of the disk array 33, so that all disk drives in the disk array 33 participate in the execution of every input/output request from the host device 31.
Referring now to FIG. 3, there is shown a write request routine 100 which is executed by the disk array controller 10. The routine 100 may be initiated when the host device 31 requests that a stripe of data be written to the disk array 33.
In response to the write request from the host device 31 , the disk array controller 10 causes a first chunk of data (chunk 1) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64 (step 110). With reference to FIG. 1 and FIG. 2, chunk 1 is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA memory controller 60, across the memory bus 66 and into the RPA memory 22. This chunk transfer like the others that follow may be directly controlled by the processor 12 or by a DMA engine which may be part of the RPA memory controller 60 or the IOP 24. While chunk 1 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller causes chunk 1 to be transferred to the XOR engine 62. The XOR engine then causes chunk 1 to be stored in buffer location 64a of the intermediate parity buffer 64.
The disk array controller 10 then causes a second chunk of data (chunk 2) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64 (step 120). As with chunk 1 , chunk 2 is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA controller memory 60, across the memory bus 66 and into the RPA memory 22. While chunk 2 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller 60 also causes chunk 2 to be transferred to the XOR engine 65. The XOR engine 62 then causes chunk 2 to be bit-wise XORed with the data stored in buffer location 64a of the intermediate parity buffer 64.
In step 130, the disk array controller 10 causes a third chunk of data (chunk 3) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64. Chunk 3, like chunk 1 and chunk 2 above, is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA controller memory 60, across the memory bus 66 and into the RPA memory 22. While chunk 3 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller 60 also causes chunk 3 to be transferred to the XOR engine 62. The XOR engine 62 then causes chunk 3 to be bit-wise XORed with the data stored in the buffer location 64a.
The disk array controller 10 in step 140 causes a fourth chunk of data (chunk 4) to be transferred to the RPA memory 22 and to the intermediate parity buffer 64. Like chunks 1, 2 and 3 above, chunk 4 is transferred from the host device 31 , through the host interface circuit 16, across the system bus 28, through the RPA controller memory 60, across the memory bus 66 and into the RPA memory 22. While chunk 4 is transferred through the RPA memory controller 60 to the RPA memory 22, the RPA memory controller 60 also causes chunk 4 to be transferred to the XOR engine 62. The XOR engine 62 then causes chunk 4 to be bit-wise XORed with the data stored in the buffer location 64a.
After step 140 is complete, the buffer location 64a contains parity information (parity chunk A) for data chunks 1-4. In step 150, the RPA memory controller 60 causes parity chunk A to be transferred from the buffer location 64a to the RPA memory 22. In particular, the RPA memory controller 60 causes the XOR engine 62 to retrieve parity chunk A from buffer location 64a and to send parity chunk A to the RPA memory controller 60. The RPA memory controller 60 then transfers parity chunk A across the memory bus 66 for temporary storage in the RPA memory 22.
Once the write data (chunks 1-4) and parity chunk A are stored in the RPA memory 22, the processor 12 schedules a write operation for transferring the write data and parity information to the disk array 33 (step 160). In particular, the processor 12 schedules chunk 1 to be transferred to disk 1 , chunk 2 to be transferred to disk 2, chunk 3 to be transferred to disk 3, chunk 4 to be transferred to disk 4, and parity chunk A to be transferred to the parity disk 5 of the disk array 33. When the scheduled write operation is executed, the write data (chunks 1-4) and parity chunk A are transferred out of the RPA memory 22, across the memory bus 66, through RPA memory controller 60, across the system bus 28, through the IOP 24 and to the respective disk drives associated with the disk array 33.
It should be appreciated that the write request routine 100 reduces the data traffic across the memory bus 66 when compared to a similar write request operation performed by the prior art. In particular, the number of chunks (Y) that are transferred across the memory bus 66 when executing routine 100 may be determined from equation (1):
(1) γ = ^2 + j^ X
where X is the number of data chunks to be transferred to the disk array 33, N is the number of data disk drives associated with the disk array 33, and Y is the resulting number of data and parity chunks transferred across the memory bus 66 as a result of transferring X data chunks. Therefore, from the above equation, disk array controller 10 transfers ten chunks across the memory bus 66 in order to transfer a stripe of five chunks (four data chunks and one parity chunk). That is:
Thus, in the embodiment being described, execution of routine 100 provides an approximately 30% reduction (14 chunks - 10 chunks) / 14 chunks in data traffic across the memory bus as compared to the prior art. It should further be appreciated that if the size of the chunk is larger than each of the buffers 64a-64d of the intermediate parity buffer 64, then each chunk may be broken up into sub-chunks that are small enough to be stored within a buffer of the intermediate parity buffer 64. Once the chunks are broken up into sub-chunks, the above method 100 may be implemented in at least two different ways. One implementation is to use multiple buffers 64a-64d and transfer all the sub-chunks of a disk before moving on to the next disk. This method can viewed as creating from the buffers 64a-64d of the intermediate parity buffer 64 a buffer which is large enough to hold a chunk. Therefore, this implementation follows the method 100 except that it makes multiple transfers at each step.
Another implementation is to use only one of the buffers of the intermediate parity buffer 64. This implementation transfers a sub-chunk from each disk as described by method 100 but skips step 160. This implementation repeats this process until all the sub-chunks and parity sub- chunks are in the RPA memory 22. Once all the sub-chunks and parity sub-chunks for the stripe are in the RPA memory 22, this implementation executes step 160 which transfers all the sub-chunks and parity sub-chunks for the stripe to the disk array 33. Referring now to FIG. 4, there is shown a degraded read request routine 200 which may be executed by the disk array controller 10 when a disk drive associated with the disk array 33 has failed. For example, if disk 2 which contains chunk 2 has failed, the disk array controller 10 must obtain valid read data from the remaining three functioning data disk drives of the disk array 33 (disks 1 , 3 and 4) and from the parity disk drive (disk 5).
The routine 200 may be initiated when the host device 31 requests that a stripe be read from the disk array 33. In response to the read request from the host device 31, the processor 12 causes valid read data to be transferred from the disk array 33 to the RPA memory 22 (step 210). With reference to FIG. 1 and FIG. 2, since disk 2 has failed, the processor 12 causes data chunks 1 , 3, 4, and parity chunk A to be transferred from disks 1 , 3, 4, and 5, through the IOP 24, across the system bus 28, through the RPA memory controller 60, across the memory bus 66 and into the RPA memory 22. Once chunks 1 , 3 and 4 and parity chunk A are stored in the RPA memory 22, the disk array controller 10 causes chunk 1 to be transferred to the host device 31 and to the intermediate parity buffer 64 (step 220). In particular, chunk 1 is transferred out of the RPA memory 22, across the memory bus 66, through the RPA memory controller 60, across the system bus 28, and through the host interface circuit 16 to the host device 31. While chunk 1 is transferred through the RPA memory controller 60 to the host device 31, RPA memory controller 60 also causes chunk 1 to be transferred to the XOR engine 62. The XOR engine 62 then stores chunk 1 in buffer location 64a of intermediate parity buffer 64.
In step 230, the disk array controller 10 then causes chunk 3 to be transferred to the host device 31 and to the intermediate parity buffer 64. As with chunk 1 , chunk 3 is transferred from the RPA memory 22 across the memory bus 66, through the RPA memory controller 60, across the system bus 28, and through the host interface circuit 16 to the host device 31. While chunk 3 is transferred through the RPA memory controller 60 to the host device 31 , the RPA memory controller 60 also causes chunk 3 to be transferred to the XOR engine 62. The XOR engine 62 then causes chunk 3 to be bit-wise XORed with the data stored in the buffer location 64a of the intermediate parity buffer 64.
The disk array controller 10 in step 240 causes chunk 4 to be transferred to the host device 31 and to the intermediate parity buffer 64. Like chunks 1 and 3 above, chunk 4 is transferred from the RPA memory 22 across the memory bus 66, through the RPA memory controller 60, across the system bus 28, and through the host interface circuit 16 to the host device 31. While chunk 4 is transferred through the RPA memory controller 60 to the host device 31 , the RPA memory controller 60 also causes chunk 4 to be transferred to the XOR engine 22. The XOR engine then causes chunk 4 to be bit-wise XORed with the data stored in the buffer location 64a of the intermediate parity buffer 64.
In step 250, the disk array controller 10 causes chunk 2 to be recovered and sent to the host device 31. In particular, the RPA memory controller 60 causes the parity chunk A stored in the RPA memory 22 to be transferred across the memory bus 66, through the RPA memory controller 60, and to the XOR engine 62. The XOR engine 62 then causes the parity chunk A to be bit-wise XORed with the data stored in buffer location 64a. When step 250 is complete, the buffer location 64a contains a restored copy of chunk 2 which was stored on the failed disk 2.
The RPA memory controller 60 causes the XOR engine 62 to retrieve data chunk 2 stored in buffer location 64a and to send the retrieved data chunk 2 to the RPA memory controller 60. The RPA memory controller 60 then sends chunk 2 across the system bus 28 and through the host interface circuit 16 to the host device 31.
It should be appreciated that the degraded read request routine 200 reduces the data traffic across the memory bus 66 when compared to a similar degraded read request routine performed by the prior art. In particular, the number of chunks (Y) that are transferred across the memory bus 66 when executing the routine 200 may be determined from equation (2):
(2) Y = 2 +Έ) x where X is the number of data chunks to be transferred to the host device 31 , N is the number of data disk drives associated with the disk array 33, and Y is the resulting number of data and parity chunks transferred across the memory bus 66 as a result of transferring X data chunks. Therefore, disk array controller 10 in a degraded read operation transfers nine chunks across the memory bus 66 in order to read a stripe from the disk array 33 having four data disk drives and one parity disk drive. That is:
7= 2 + ) .4 = 9
Thus, in the embodiment being described, execution of routine 200 provided an approximately 30% reduction (13 chunks -9 chunks) / 13 chunks in data traffic across the memory bus as compared to the prior art.
The RPA memory controller 60 may also write the restored chunk 2 data into the RPA memory 22 for subsequent use during a disk caching operation. If the RPA memory controller 60 is implemented to write the restored chunk 2 data into the RPA memory 22, then the number of data transfers as a result of a degraded read operation may be determined from above equation (1) instead of above equation (2).
Furthermore, the data requested by the host device 31 may be transferred to the host device 31 out of order (i.e. chunk 1 , chunk 3, chunk 4, and then chunk 2) depending upon which disk drive of disk array 33 fails. As a result, a data transfer protocol may be required which permits out of order transfers of data chunks. However, if the degraded read operation is performed in a similar manner as the write operation described above (i.e. as chunks are transferred from the disk array 33 to the RPA memory 22, they are XORed with data in a buffer of the intermediate parity buffer 64), then the data may be transferred to the host device 31 in order. It should be appreciated, however, that this in order implementation of degraded read operations increases the complexity and cost of the RPA circuit 20.
In optimum read operations, there is no disk drive failure associated with the disk array 33. Thus, it should be appreciated that there is no need to read parity information from the parity disk drive into the RPA memory 22, or to send the chucks of data to the intermediate parity buffer 64 after the chunks of data have been written into the RPA memory 22. In particular, in optimum read operations, the read data is transferred from the disk array 33 through the IOP 24, across the system bus 28, through the RPA memory controller 60, across the memory bus 66, and into the RPA memory 22. The processor 12 then schedules a read operation where the chunks of read data are transferred out of the RPA memory 22, across the memory bus 66, through the RPA memory controller 60, across the system bus 28, through the host interface 16 and to the host device 31. While the invention has been illustrated and described in detail in the drawings and foregoing description, such illustration and description is to be considered as exemplary and not restrictive in character, it being understood that only the preferred embodiment has been shown and described and that all changes and modifications that come within the spirit of the invention are desired to be protected.

Claims

What is claimed is:
1. A method for transferring data from a first device to a second device connected by a controller having a parity buffer and a memory having a first storage and a second storage, comprising the steps of: transferring first data from the first device to the first storage; transferring second data from the first device to the second storage; transferring said first data to the second device and storing said first data in the parity buffer; and determining parity data from said second data and said first data stored in the parity buffer.
2. The method of claim 1 , further comprising the step of: transferring said parity data to said second device.
3. The method of claim 1 , wherein the step of determining parity data comprises the step of: transferring said second data to the second device.
4. The method of claim 1 , further comprising the step of: storing said parity data in said parity buffer.
5. The method of claim 1 , further comprising the step of: storing said parity data in a third storage of the memory.
6. The method of claim 1 , wherein the step of determining parity data further comprises the step of: taking the bit-wise exclusive-OR of said second data and said first data stored in the parity buffer.
7. A method for transferring data from a first device to a second device connected by a controller having a parity buffer and a memory having a first storage and a second storage, comprising the steps of: transferring first data from the first device to the first storage and storing said first data in the parity buffer; transferring second data from the first device to the second storage and determining parity data from said second data and said first data stored in the parity buffer; and transferring said first data to the second device.
8. The method of claim 7, further comprising the step of: transferring said parity data to said second device.
9. The method of claim 7, wherein the step of determining parity data comprises the step of: transferring said second data to the second device.
10. The method of claim 7, further comprising the step of: storing said parity data in said parity buffer.
11. The method of claim 7, further comprising the step of: storing said parity data in a third storage of the memory.
12. The method of claim 7, wherein the step of determining parity data includes the step of: taking the bit-wise exclusive-OR of said second data and said first data stored in the parity buffer.
13. A disk array controller for connecting a host device to a disk array having a first disk, a second disk, and a third disk, the disk array controller comprising: a memory comprising a first storage and a second storage; a parity buffer; a parity generator coupled to said parity buffer; and, a parity controller coupled to said memory and said parity generator, said parity controller configured to transfer first data from the first disk to said first storage; configured to transfer second data from the second disk to said second storage; configured to transfer said first data from said first storage to the host device and to said parity generator; and configured to transfer said second data from said second storage to said parity generator; wherein said parity generator is configured to determine parity data from said second data and said first data stored in said parity buffer.
14. The disk array controller of claim 13, wherein: said parity controller is further configured to transfer said parity data to the host device.
15. The disk array controller of claim 14, wherein: said parity controller is further configured to transfer said parity data to a third storage of said memory.
16. The disk array controller of claim 15, wherein: said parity generator determines said parity information by taking the bit-wise exclusive-OR of said second data and said first data stored in said parity buffer.
17. A disk array controller for connecting a host device to a disk array having a first disk, a second disk, and a third disk, the disk array controller comprising: a memory comprising a first storage and a second storage; a parity buffer; a parity generator coupled to said parity buffer; and a parity controller coupled to said memory and said parity generator, said parity controller configured to transfer first data from the host device to said first storage and to said parity generator; configured to transfer second data from the host device to said second storage and to said parity generator; configured to transfer said first data from said first storage to the first disk; configured to transfer said second data from said second storage to second disk; and configured to transfer parity data from said parity generator to the third disk; wherein said parity generator is configured to store said first data in said parity buffer; and is configured to determine said parity data from said second data and said first data stored in said parity buffer.
18. The disk array controller of claim 17, wherein: said parity controller is further configured to transfer said parity data to a third storage of said memory.
19. The disk array controller of claim 18, wherein: said parity generator determines said parity information by taking the bit-wise exclusive-OR of said second data and said first data stored in said parity buffer.
PCT/US1997/022807 1997-12-10 1997-12-10 Computer bus data transfer reduction with parity circuit WO1999030233A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/US1997/022807 WO1999030233A1 (en) 1997-12-10 1997-12-10 Computer bus data transfer reduction with parity circuit
AU55228/98A AU5522898A (en) 1997-12-10 1997-12-10 Computer bus data transfer reduction with parity circuit

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US1997/022807 WO1999030233A1 (en) 1997-12-10 1997-12-10 Computer bus data transfer reduction with parity circuit

Publications (1)

Publication Number Publication Date
WO1999030233A1 true WO1999030233A1 (en) 1999-06-17

Family

ID=22262265

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1997/022807 WO1999030233A1 (en) 1997-12-10 1997-12-10 Computer bus data transfer reduction with parity circuit

Country Status (2)

Country Link
AU (1) AU5522898A (en)
WO (1) WO1999030233A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1241577A2 (en) * 2001-03-14 2002-09-18 Hewlett-Packard Company Methods and arrangements for improved parity-stripe processing

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5146588A (en) * 1990-11-26 1992-09-08 Storage Technology Corporation Redundancy accumulator for disk drive array memory
US5373512A (en) * 1991-09-13 1994-12-13 International Business Machines Corporation Memory controller with parity generator for an I/O control unit
US5396620A (en) * 1993-12-21 1995-03-07 Storage Technology Corporation Method for writing specific values last into data storage groups containing redundancy
US5469566A (en) * 1992-03-12 1995-11-21 Emc Corporation Flexible parity generation circuit for intermittently generating a parity for a plurality of data channels in a redundant array of storage units
EP0767420A1 (en) * 1995-10-02 1997-04-09 International Business Machines Corporation System and method for performing raid operations in a data processing system
EP0768607A2 (en) * 1995-10-13 1997-04-16 Compaq Computer Corporation Disk array controller for performing exclusive or operations

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5146588A (en) * 1990-11-26 1992-09-08 Storage Technology Corporation Redundancy accumulator for disk drive array memory
US5373512A (en) * 1991-09-13 1994-12-13 International Business Machines Corporation Memory controller with parity generator for an I/O control unit
US5469566A (en) * 1992-03-12 1995-11-21 Emc Corporation Flexible parity generation circuit for intermittently generating a parity for a plurality of data channels in a redundant array of storage units
US5396620A (en) * 1993-12-21 1995-03-07 Storage Technology Corporation Method for writing specific values last into data storage groups containing redundancy
EP0767420A1 (en) * 1995-10-02 1997-04-09 International Business Machines Corporation System and method for performing raid operations in a data processing system
EP0768607A2 (en) * 1995-10-13 1997-04-16 Compaq Computer Corporation Disk array controller for performing exclusive or operations

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1241577A2 (en) * 2001-03-14 2002-09-18 Hewlett-Packard Company Methods and arrangements for improved parity-stripe processing
EP1241577A3 (en) * 2001-03-14 2004-09-22 Hewlett-Packard Company Methods and arrangements for improved parity-stripe processing

Also Published As

Publication number Publication date
AU5522898A (en) 1999-06-28

Similar Documents

Publication Publication Date Title
US5883909A (en) Method and apparatus for reducing data transfers across a memory bus of a disk array controller
US5742752A (en) Method for performing a RAID stripe write operation using a drive XOR command set
EP1019835B1 (en) Segmented dma with xor buffer for storage subsystems
US7730257B2 (en) Method and computer program product to increase I/O write performance in a redundant array
EP0594464B1 (en) Method of operating a data storage disk array
US5596708A (en) Method and apparatus for the protection of write data in a disk array
US7975168B2 (en) Storage system executing parallel correction write
JP3742494B2 (en) Mass storage device
US5727144A (en) Failure prediction for disk arrays
US5996046A (en) Parity generation system for generating new parity using old data in temporary storage without accessing main disk storage of disk drive
US20040037120A1 (en) Storage system using fast storage devices for storing redundant data
EP0986000B1 (en) Disk array subsystem
JPH10105344A (en) Data recorder system and data write method therefor
US5787463A (en) Disk array system including a dual-ported staging memory and concurrent redundancy calculation capability
US7130973B1 (en) Method and apparatus to restore data redundancy and utilize spare storage spaces
US6611897B2 (en) Method and apparatus for implementing redundancy on data stored in a disk array subsystem based on use frequency or importance of the data
US6513098B2 (en) Method and apparatus for scalable error correction code generation performance
JP2857288B2 (en) Disk array device
JP2001043031A (en) Disk array controller provided with distributed parity generating function
US6785788B1 (en) System and method for implementing an enhanced raid disk storage system
US5875458A (en) Disk storage device
WO1999030233A1 (en) Computer bus data transfer reduction with parity circuit
CN111475112A (en) Device for improving performance of Oracle database and data reading and writing method
US20010052100A1 (en) Data read/write controlling method, disk array apparatus, and recording medium for recording data read/write controlling program
JP2857289B2 (en) Disk array device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GE GH HU IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SZ UG ZW AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase