GB2298306A - A disk array and tasking means - Google Patents

A disk array and tasking means Download PDF

Info

Publication number
GB2298306A
GB2298306A GB9608322A GB9608322A GB2298306A GB 2298306 A GB2298306 A GB 2298306A GB 9608322 A GB9608322 A GB 9608322A GB 9608322 A GB9608322 A GB 9608322A GB 2298306 A GB2298306 A GB 2298306A
Authority
GB
United Kingdom
Prior art keywords
data
unit
read
write
fifo
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB9608322A
Other versions
GB9608322D0 (en
GB2298306B (en
Inventor
Masahiro Mizuno
Shiro Ogura
Kazuhiko Ito
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Electric Corp
Original Assignee
Mitsubishi Electric Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP26736492A external-priority patent/JP3181398B2/en
Application filed by Mitsubishi Electric Corp filed Critical Mitsubishi Electric Corp
Publication of GB9608322D0 publication Critical patent/GB9608322D0/en
Publication of GB2298306A publication Critical patent/GB2298306A/en
Application granted granted Critical
Publication of GB2298306B publication Critical patent/GB2298306B/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1076Parity data used in redundant arrays of independent storages, e.g. in RAID systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1008Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's in individual solid state devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device

Abstract

A read unit 400 which governs reading of data and a write unit 500 which governs writing of data are located separately in a RAID unit 300 which controls array type magnetic disk units. To read data from magnetic disk units 800, a control circuit 600 uses the read unit 400 to read data through an HDD control unit 700. To write data, the control circuit 600 uses the write unit 500 to write data into the magnetic disk units 800. Since the read unit and write unit are provided separately, processing performance improves. A tasking means 340 is provided for accessing a number of media in the array which selects the next task to be performed by reading an interrupt vector 344, the individual bits being set upon completion of a corresponding disk access.

Description

ARRAY TYPE RECORDING SYSTE"i This invention relates to an array type recording system which is a computer storage system and more particu larly to improvements in performance and reliability of a disk drive system in which a number of disk units are arranged in an array type configuration.
Various documents and patents for disk drive systems each consisting of a number of disk units arranged like an array have been published. One of the documents is a publication on a system which dramatically improves the reliabilits of data stored in mass storage from Berkeley college of California University. This paper "A Case for Redundant Arrays of Inexpensive Disks (RAID)," Proc, ACM SIGMOD Conf., Chicano, IL, June 1988 classifies data reliability improvement systems into five levels ranging from a conventional mirror disk system to a block interleave parity system.
These levels are outlined below: RAID level 1: A normal mirror (shadow) system which stores the same data in two groups of disk units. The RAID level 1 system has been generally used formerly with computer systems for which high reliability is required. However, large redun danc leads to high cost per unit capacity.
RAID level 2: The hamming code format used with DRAM is applied.
Data in a redundant group is stored on disks with bits interleaved. On the other hand, to enable l-bit error correction, an ECC code is written onto a number of check disks (four check disks are required when the number of data disks is ten) per group, one group consisting of about 10 to 25 disk units. Redundancy is slightly large.
RAID level 3: parit disk is fixed for use and data is recorded on data disks in the group with bytes interleaved. Since the error location is found from ECC for each drive. only one parity disk is required. The RAID level 3 system is appro private for synchronizing spindle rotation for high speed transfer.
RAID level 4: parity disk is fixed for use and data is recorded on data disks in the group with blocks interleaved. RAID level 4 differs from level 3 in interleave units. That is, because of recording in block units, the RAID level 4 system is more appropriate for the case in which an access to small data is often made.
RAID level 5: Unlike level 3 or 4, the RAID level 5 system does not have a fixed parity disk and stripes parity data on component disks. Thus, at a write operation, load concentration on parity disks does not occur and lOPS increases (the higher the write percentage, the more advantageous the RAID level 5 sstem is than RAID level 4). Both operating performance and capacity efficiency are good.
A conventional example of a redundant array type disk drive system is "Array Type Disk Drive System and Method" by Array Technology Corporation in USA disclosed in Japanese Patent Laid-Open No.Hei 2-236714, wherein the redundancy level and the number of logical units of component disk units viewed from the host computer can be selected.
The method of striping parity data is shown in Japanese Patent Laid-Open No.Sho 62-293355. "Data Protection Feature" by International Business Machines Corporation in USA.
In the accompanying drawings Figure 28 is a block diagram of the array type disk drive sstem disclosed in Japanese Patent Laid-Open No.Hei 2-236714 mentioned above. for example. In the figure, numeral 2 is a host interface (I/F) serving as a buffer between a host computer (not shown) and an array controller, numeral 3 is a microprocessor which controls an array controller, numeral 4 is a memory, numeral 5 is an EOR engine which generates redundant data and restores data. numeral 6 is a common data bus which connects the host I/F 2, the microprocessor 3, the memory 4, and the EOR engine 5, and numeral 8 is a channel controller, a plurality of channel controllers being connected to the data bus 6.Numeral 9 is a disk unit and numeral 10 is a channel; each of the disk units 9 is connected via the corresponding channel 10 to the corresponding channel controller 8. Numeral 13 is an array controller which controls the disk units 9.
Figure 29 is a drawing for illustrating generation of redundant data on RAID. As shown in Figure 29. stored on one of five disks is redundant data (parity) of data on the other four disks. The parity is calculated by exclusive ORing the data on the four disks. That is, the parity data on the parity disk P results from exclusive-ORing the data on disks 0 to 3. For example. if the data on disk 0 cannot be read due to some fault, providing such parity as redundant data enables the data on disk 0 to be restored. That is, the data resulting from exclusive-ORing the data on disks 1 to 3 and parity disk can be used to restore the data on disk 0.
Although the parity can be calculated by exclusive ORing the data on the four disks as described above, alternatively the old data on the disk onto which new data is to be written and the current parity data stored on the parity disk may be read to exclusive-OR the three types of data.
the new data, old data, and parity data, thereby providing new parity data. This method is described in conjunction with Figure 30. For example, to attempt to record new data DN (2j on disk 2, first the old data is read as DO (2) from disk 2. At the same time. the current parity data DO (P) is read from the parity disk. Next, the three types of data, DN (2), DO (2). and DO (P), are exclusive-ORed to generate new parity data DN (P). Then. the new data DN (2) is recorded on disk 2. Last, the new parity data DN (P) is recorded on the parity disk.
Next. the operation of the disk drive system shown in Figure 28 is described. In Figure 28, the host computer (not shown) always writes and reads data via the host I/F 2 into and from the disk system. When data is stored, instructions and data from the host computer are temporarily stored in the buffer memory 4 via the data bus 6. When data is reproduced, data provided in the buffer memory 4 is transferred via the host I/F 2 to the host computer.
Figures 31A and 31B are an internal block diagram and an operation flowchart of the host I/F 2. In Figure 31A, numeral 21 is an I/F protocol chip and numeral 22 is a microcomputer. The I/F protocol chip 21 is an interface chip to handle SCSI (small computer system interface) and the microcomputer 22 analyzes the contents of data received at the I/F protocol chip 21 and outputs the result to the array controller 13 shown in Figure 28. As shown in the flowchart of Figure 31B, the microcomputer 22 checks a given command for validity, then analyzes the contents of the command and makes address conversion from logical address to physical address according to the analysis result. Thus, the command validity check, command decode, and address conversion are executed sequentially.Since the time required is. for example, between 300 microseconds and 1 millisecond, even if, for example, performance of other hardware devices is improved, the data transfer speed cannot be substantially improved because performance of the microcomputer 22 in the host I/F 2 does not improve.
The operation at RAID level 5 is described in conjunction with Figures 28 and 32. The microprocessor 3 divides data stored in the memory 4 into data blocks and determines data "rite disk units and a redundant data write unit. At RAID level 5, old data in the data blocks into which new data is to be written is required to update redundant data, thus a read operation is executed before a write operation.
Data is transferred between the memory 4 and the channel controllers 8 via the data bus. Redundant data is generated by the EOR engine in synchronization with the data transfer.
Assuming that a data block is set to 512 bytes, for example, when 1024-byte data is written, it is recorded in two blocks 16 and 17 and parity data P is also recorded, as shown in Figure 32. Such a recording state is called striping.
This is described in detail. First, write data disk units 9a and 9b and a redundant data disk unit 9e are determined. Next, the EOR engine 5 is started under the control of the microprocessor 3 for sending an old data read command for redundant data calculation to the channel controllers 8a, 8b, and 8e to which the data disk units 9a and 9b and the redundant data disk unit 9e are connected. After completion of reading the old data in the data disk units 9a and 9b and the redundant data disk unit 9e, new data is mitten into the data disk units 9a and 9b and the new redundant data generated by the EOR engine is written into the redundant data disk unit 9e as instructed by the microprocessor 3. Then, the host computer (not shown) is informed that data write is complete.As described above, when data is written, a preread of old data is required to generate redundant data, prolonging the processing time.
When one data record is divided and recorded over two or more disks as shown in Figure 32, the two or more disks must be accessed to access the data record, degrading performance.
Next. a data read is described. When a data read is instructed by the host computer. the microprocessor 3 calculates the data block and data disk unit where the target data is stored. For example, if the data is stored in the disk unit 9c, a read command is issued to the channel con troller 8c to which the disk unit 9c is connected. Upon completion of reading the data in the disk unit 9c. the data is transferred to the memory 4 and the host computer is informed that the data read is complete.
Next, data recovery and data reconstruction on a standby disk when an error occurs are described. Data recovery is executed when it is becomes impossible to read data in the disk unit 9c, for example. When it is impossible to read data in the disk unit 9c, the microprocessor 3 reads data from all disk units in the redundant group containing the read data block and the EOR engine 5 restores the data in the data block where it is impossible to read data.
For example. assuming that the redundant group consists of disk units 9a. 9b. 9c, 9d. and 9e. data blocks are read from the disk units 9a, 9b, 9d, and 9e, the EOR engine 5 restores the data in the disk unit 9c, and the data is transferred to the memory 4. Then, the host computer is informed that data read is complete.
Thus, even if an error occurring in a disk unit makes it impossible to read data, data can be recovered, improving data reliability.
Data reconstruction is executed when it is becomes im- possible to use the disk unit 9c, for example. In this case, the microprocessor 3 reads data from all disk units in the redundant group containing the data stored in the data unit 9c, the EOR engine 5 restores data in the disk unit 9c, and the restored data is reconstructed on a standby disk.
For example, assuming that the redundant group consists of disk units 9a, 9b, 9c, 9d, and 9e, data is read from the disk units 9a, 9b, 9d, and 9e, the EOR engine 5 restores the data in the disk unit 9c, and the restored data is written onto a standby disk for reconstructing the data in the disk unit 9c on the standby disk. Then, the unavailable disk unit 9c is replaced by the standby disk. Since the replacement operation is performed when the system is operating, the system performance degrades during the replacement processing.
Since the conventional array type disk drive system is configured as described above, a data preread is also required to generate redundant data when data is written in a normal operation, prolonging the processing time.
When replacement occurs on an alternate (standby) disk in disk unit replacement processing, furthermore the system performance is further degraded.
Accordingly, it is an object of the invention to provide array type recording systems in which the read processing time and write processing time are shortened.
GB-A-2271462, from which the present application is divided, is directed to an array type recording system comprising: (a) a plurality of media for recording data; (b) data write means having first-in first-out buffers corresponding to said plurality of recording media on a oneto-one basis, said data write means being a write circuit for writing data onto said plurality of recording media; and (c) data read means having first-in first-out buffers corresponding to said plurality of recording media on a oneto-one basis, said data read means including a read circuit, being different from said write circuit, for reading data from said plurality of recording media.
GB-A- > , which is also divided from GB-A-2271462, is directed to an array type recording system comprising: (a) a plurality of media for recording data; (b) means for reading/writing data from/onto said plurality of recording media as requested; and (c) interface means for issuing a request to said read/write means, said interface means comprising: (cl) a command input section which inputs commands from an external system; (c2) a multiport memory which has a plurality of access ports and stores the commands input through said command input section; and (c3) a plurality of command processing sections in which processing for converting into a request sent to said read/write means from the commands stored in said multiport memory is divided among said plurality of command processing sections for accessing the commands through the access ports of said multiport memory and executing their allotted processing portions in parallel.
GB-A- > , which is also divided from GB-A-2271462, is directed to an array type recording system comprising: (a) a plurality of media for recording data; (b) write means for recording data and redundant data of the data on said plurality of recording media in a scattered manner and for writing a plurality of continuous data entries, said write means writing them onto said plurality of recording media in order in a predetermined direction and locating redundant data of the data entries on said plurality of recording media cyclically and starting to write data entries in the next group at the recording media where redundant data in the preceding data group has been written and writing said redundant data in said preceding data group and data in said next group together on said plurality of recording media; and (c) read means for continuously reading a plurality of continuous data entries written by said write means, when reading data from one redundant group, said read means reading redundant data in the redundant group and ignoring it, then reading data on said plurality of recording media in order in a predetermined direction starting at the recording media where the redundant data has been read.
To this end, according to the invention, there is provided an array type recording system comprising: (a) a plurality of media for recording data; (b) a plurality of task means for issuing an access request to any desired number of recording media of said plurality of recording media; (c) means for reporting access completion to said recording media accessed by said task means; and (d) task start means being responsive to a combination of reports from said report means for selecting the corresponding task means among said plurality of task means and starting its task.
Thus, the task issuing a processing request for recording media is identified in response to a combination of completion reports from the report means, and the identified task is started. When processing is complete on all recording media processed according to the processing request, the task expecting the processing completion can be started automatically.
Preferably, the recording media are magnetic disks.
Thus, a plurality of conventional magnetic disk units can be used.
Alternatively, the recording media are magneto-optic disks.
Thus, the system can be constructed with a large capacity and easy-to-carry recording media as compared with the system which uses a plurality of magnetic disk units.
The invention will be described by way of nonlimitative example with reference to the accompanying drawings, in which: Figure 1 is a block diagram of an array type recording system according to a first embodiment of the invention; Figure 2 is a block diagram of an interface unit of the array type recording system according to the first embodiment of the invention; Figure 3 is a block diagram of another type of interface unit of the array type recording system according to the first embodiment of the invention; Figure 4 is a block diagram of a self-recovery unit of the array type recording system according to the first embodiment of the invention; Figure 5 is a block diagram of a RAID unit of the array type recording system according to the first embodiment of the invention;; Figure 6 is a block diagram of an HDD control unit and magnetic disk units 800 of the array type recording system according to the first embodiment of the invention; Figure 7 is a drawing showing a spiraling data array in the array type recording system according to the first embodiment of the invention; Figure 8 is a drawing showing a modification example of the spiraling data array in the array type recording system according to the first embodiment of the invention; Figure 9 is a block diagram of a read unit of the array type recording system according to the first embodiment of the invention; Figure 10 is a circuit diagram of an exclusive OR circuit of the array type recording system according to the first embodiment of the invention; Figure 11 is a drawing showing the timing for generating recovery data in the first embodiment of the invention;; Figure 12 is a block diagram of a write unit of the array type recording system according to the first embodi met of the invention; Figure 13 is a schematic block diagram of the read unit and the write unit with magnetic disk units between; Figure 14 is a first drawing for illustrating a short read operation according to the first embodiment of the invention; Figure 15 is a second drawing for illustrating a short read operation according to the first embodiment of the invention; Figure 16 is a third drawing for illustrating a short read operation according to the first embodiment of the invention; Figure 17 is a first drawing for illustrating an 1block write operation according to the first embodiment of the invention; Figure 18 is a second drawing for illustrating an 1block write operation according to the first embodiment of the invention;; Figure 19 is a third drawing for illustrating an 1block write operation according to the first embodiment of the invention; Figure 20 is a fourth drawing for illustrating an 1block write operation according to the first embodiment of the invention; Figure 21 is a first drawing for illustrating a group write operation acccrding to the first embodiment of the invention; Figure 22 is a second drawing for illustrating an 1block write operation according to the first embodiment of the invention; Figure 23 is a third drawing for illustrating an 1block write operation according to the first embodiment of the invention; Figure 24 is a fourth drawing for illustrating an 1block write operation according to the first embodiment of the invention; Figure 25 is a fifth drawing for illustrating an 1block write operation according to the first embodiment of the invenTion;; Figure 26 is a block diagram of an interrupt circuit of the array type recording system according to the first embodiment of the invention; Figure 27 is a drawing for illustrating interrupt vectors according to the first embodiment of the invention; Figure 28 is a drawing showing the configuration of a conventional array type magnetic disk drive system; Figure 29 is a drawing for illustrating the operation of the array type magnetic disk drive system; Figure 30 is a drawing for illustrating the operation of the array type magnetic disk drive system; Figure 31A is a drawing for illustrating the operation of a conventional interface unit; Figure 31B is a flowchart for illustrating the operation of the conventional interface unit; and Figure 32 is a drawing showing a conventional striping data array.
Embodiment 1: Figure 1 is a block diagram of an array type recording system according to a first embodiment of the invention.
In the figure. numeral 100 is an interface unit, numeral 200 is a self-recovery unit, numeral 300 is a RAID unit, numeral 400 is a read unit contained in the RAID unit 300, numeral 500 is a write unit contained in the RAID unit 300, and numeral 600 is a control circuit which controls the entire RAID unit including the read unit 400 and the write unit 500. Numeral 700 is a magnetic disk unit control unit (HDD control unit) connected to the RAID unit 300. Numeral 800 is a magnetic disk unit.
Next, the interface unit 100 is described in conjunction with Figures 2 and 3.
In Figure 2, numeral 101 is a SCSI interface, numeral 110 is a SCSI chip which governs the SCSI interface, numeral 120 is a SCSI microprocessing unit (MPU) which analyzes a protocol input from the SCSI chip 110, numeral 130 is a DMA (direct memory access) chip which executes DMA transfer, numeral 140 is a dual port RAM (random access memory) having two access ports" numeral 150 is a buffer having a storage capacity of 0.5 to 1 megabytes, numeral 160 is a control bus connected to the SCSI chip 110, SCSI MPU 120, and dual port RUl 140. Numeral 170 is a data bus connected to the SCSI chip 110. DMA chip 130, dual port RAM 140, and buffer 150.
The dual port RAM 140 has two access ports. in the example. connected to the control bus 160 and the data bus 170. .t processing request input via the SCSI interface 101 to the SCSI chip 110 is divided and output to the control bus 160 and the data bus 170 for processing. The SCSI chip 110 temporarily stores the processing request in the dual port RA'.I 140 and then the processing request stored in the dual port R.EM 140 is accessed by the SCSI MPU 120 and the D'l.i chip 130. The SCSI MPU 120, for example, reads commands (CDB (command description block)) stored in the dual port R.A': 140 via the control bus 160 and checks the commands for validity and decodes them.On the other hand, the DMA chip 130, for example, makes address conversion from logical address to physical address in order to get addresses required while the SCSI MPU 120 executes the command validity check and command decode.
The interface unit 100 is characterized by the fact that it has the separate control bus and data bus and that commands are temporarily stored in the dual port RAM 140, through the ports of which the processors access the commands stored in the dual port RAM 140 at the same time for executing command analysis and necessary conversion in parallel.
The buffer 150 is used when the length of the data to be transferred is long or to perform read modify write operation which reads data once from a magnetic disk unit, modifies it and then rewrites the resultant data into the magnetic disk unit.
The SCSI chip 110 and the DMA chip 130 can debug the connected RAID unit 300, etc., by interfacing with an external system and communicate with a diagnosis system operating in the host computer.
Figure 3 is a block diagram of another example of the interface unit 100.
In Figure 3, numeral 110 is a SCSI chip, numeral 120 is a microcomputer, numeral 121 is a command check circuit, numeral 122 is a logical-to-physical address converter. and numeral 140 is a dual port RAM.
In the example, the SCSI chip 110 stores commands (CDB) transferred from SCSI interface 101 in the dual port RAM 140. The microcomputer 120, the command check circuit 121, and the logical-to-physical address converter 122 can access the dual port RAM 140 at the same time. Therefore. the command check circuit 121 checks the commands (CDB) in the dual port R. 140 for validity. At the same time, the microcomputer 120 decodes the commands. Further, the logical-to-physical address converter 122 generates a physical address from an input logical address. The command check circuit 121 can be made of hardware or a microcomputer different from the microcomputer 120. Likewise, the logical-to-physical address converter can be made of appropriate hardware or a microcomputer different from the microcomputer 120.
Processing performed by the interface unit 100 is separated into three portions for systematic operation, thereby reducing overhead at the interface unit 100 and improving the processing speed of the array type disk drive system Figure 4 is a block diagram of the self-recovery unit 200 shown in Figure 1.
In Figure 4, numeral 210 is a log memory, numeral 220 is a calendar battery to retain a calendar of dates, time of day. etc,, numeral 230 is an MPU which controls the selfrecovery unit. and numeral 240 is an operation panel that can be installed as an option.
The operation panel 240 is provided with indicators such as a Ready LED and a Fault LED and also contains recovery or diagnosis software or a communication function for connecting a diagnosis system for online maintenance. The self-repair unit 200 contains a real-time clock (not shown) and also has the calendar battery 220 to retain a calendar of dates, time of day, etc., and can collect the operation state of the RAID unit 300 together with the date and time in the log memory 210. The MPU 230 always monitors the operation of the RAID unit 300 and collects its log data in the log memory 210. The MPU 230 also diagnoses the RAID unit 300 according to the diagnosis software installed in the operation panel 240. When a magnetic disk unit fails.
recovery work is required to replace the disk unit with an alternative disk. The MPU 230 executes data recovery work on the alternate disk according to the recovery software contained in the operation panel 240 while competing with a processing request accepted by the interface unit 100. For the recovery work on the alternative disk. another disk unit which operates normally is usually fully read from top to end and recovery data is written onto the alternative disk.
The '.IPU 230 executes the full read operation in the meantime when the interface unit 110 accepts a processing request from a higher-ranked device such as the host computer as usual.
Figure 5 is a block diagram showing the detailed configuration of the RAID unit 300 shown in Figure 1.
In Figure 5, numeral 310 is a gate array which serves as the nucleus of the RAID unit 300, numeral 400 is a read unit installed in the gate array 310, numeral 500 is a write unit installed in the gate array 310, and numeral 600 is a control circuit. Numeral 410 is a FIFO installed in the read unit 400, numeral 420 is an exclusive OR circuit provided to generate redundant data, etc. and numeral 450 is a switch for connecting the data bus 170 and FIFO 410. Numeral 510 is a FIFO installed in the write unit 500 and numeral 520 is an exclusive OR circuit provided to generate redundant data, etc. Numeral 320 is a dual port RAM which stores control information of the MPU 230 of the self-recovery unit 200 shown in Figure 4 and numeral 321 is a dual port RAM which stores control information from the control bus 160 of the interface unit 100 described above. Numeral 330 is an 'IPt which accesses data in the dual port R.X0ls 320 and 330 for communicating with the interface unit 100 and the selfrecovery unit 200. Numeral 340 is an interrupt circuit which receives a completion notification of an access to a magnetic disk unit and generates an interrupt. Numeral 350 is a stack to hold requests processed by the MPU 330.
Numeral 360 is dual port RAMs provided for magnetic disk units. Numeral 370 is a polling register for polling messages in the dual port RAMs 360 from the magnetic disk units.
The operation of the RAID unit 300 is described. The RAID unit 300 has the gate array 310 as the nucleus in which the read unit 400 and the write unit 500 are installed separately. The read unit 400 is a read circuit for reading data from a plurality of magnetic disk units and contains FIFOs 410 for recording read data from the magnetic disk units, the FIFOs corresponding to the magnetic disk units on a one-to-one basis. Likewise, the write unit 500 contains FIFOs 510 for writing data onto the magnetic disk units, the FIFOs corresponding to the magnetic disk units on a one-toone basis. The FIFOs of the read unit 400 are selectively connected by the switch 450 to the data bus 170 from the interface unit 100. This also applies to the FIFOs of the write unit 500.
Next, the read unit operation and the write unit operation are described separately.
When a command to read data is transferred via the control bus 160 from the interface unit 100, the read request is sent to the control circuit 600, also called control logic. Likewise, the request is recorded in the dual port R.M 321. The control logic 600 causes the read unit 400 to operate. When receiving the message recorded in the dual port RAM 321, the MPU 330 judges that it is a read request. and writes the message into the dual port RAM 360 corresponding to the magnetic disk unit in which relevant data is recorded. The polling register 370 performs polling b acknouledging interrupt processing from a magnetic disk unit.The magnetic disk unit corresponding to the dual port RA 360 into which the message is written by the MPU 330 transfers data via the bus 380 to the corresponding FIFO 410 of the read unit 400 in response to the message. The control logic 600 previously switches the switch 450 to place the FIFO to which the magnetic disk unit transfers the data in the ready state. After the relevant data is read into the FIFO, the control logic 600 uses the data bus 170 to transfer the data to the interface unit 100.
On the other hand, when a request to write data is issued to the interface unit 100 from the host computer or the like. the interface unit 100 informs the control logic 600 and the dual port RA1 321 through the control bus 160 that a write request is received. The control logic 600 uses the data bus 170 to get the data related to the FIFO corresponding to the magnetic disk unit into which data is to be written.The MPU 330 determines the corresponding magnetic disk unit according to the write request stored in the dual port RAM 321 and writes a message indicating that the write request is received into the related dual port RkN. When the magnetic disk unit is informed that a request to write into the magnetic disk unit is made by using the polling register, the magnetic disk unit uses the bus 380 to read and record data from the FIFO 410 corresponding to the magnetic disk unit.
Next. the control logic 600 is described. The control logic 600 controls the read unit 400 and the write unit 500 in response to requests received on the control bus 160 through the interface unit 100 from the host computer. etc.
Particularly when the FIFOs of the read unit 400 and the write unit 500 are accessed, access patterns and modes are set. For example, 19 access patterns of the first to 19th are provided. The five modes of block read, block write, single read, single write, and recovery read are set. one of which is selected.
The dual port RM: 321 records disk access requests received on the control bus 160 through the interface unit 100 in sequence. On the other hand, the dual port RAM 321 has roles of receiving completion messages of the requests from the MPU 330, recording the request completion messages, and sending them via the control bus 160 to the interface unit 100. For example, if a read request is received through the interface unit 100, when read data is provided in the FIFO 410 and is ready to be transferred from the RAID unit 300 to the interface unit 100, the dual port RAM 321 is used to inform the interface unit 100 that the data is ready to be transferred.If a data write request is received through the interface unit 100, when the write unit 500 and the control logic 500 become ready. the dual port RAM 321 is used to transmit a request to send write data via the data bus 170 to the interface unit 100.
To perform a read write modify operation which modifies once read data and then are writes the resultant data. the dual port RAM 321 is used to request the interface unit 100 to perform the modify write operation.
Next. the configuration of the HDD control unit 700 and magnetic disk units 800 shown in Figure 1 is described in conjunction with Figure 6.
In Figure 6, numeral 700 is the HDD control unit, numeral 710 is an MPU which governs contro3 of the HDD control unit 700, numeral 720 is a SCSI chip used for the HDD control unit 700 to access a disk unit by using SCSI, and numeral 730 is a cache memory which tenporarily stores data. Numeral 800 is a magnetic disk unit such as a 200 megabte, 2.5-inch or 1-gigabyte, 3.5-inch magnetic disk unit.
As shown in the figure, the HDD control unit 700 consists of five control units. each of which contains the MPU 710, SCSI chip 720, and cache memory 730. The magnetic disk units 800 are connected to the corresponding control units.
Although SCSI is used as the interface between the control unit and magnetic disk unit in the example. another interface such as IDE or PCMCIA may be used. In the example, two magnetic disk units are connected to one control unit, but three or four magnetic disk units may be connected to one control unit.
The tPU 710, which is located between the magnetic disk unit 800 and the RAID unit 300 described above. transfers data by string instructions, and can be dedicated to data transfer. The cache memory 730 is made of CAM (content addressable memory) and enables cache retrleval. Further, since the cache memory 730 is provided for each magnetic disk unit, cache retrieval can be made in parallel, enabling a high speed access. When only one cache memory is located and shared by a number of control units (magnetic disk units) as in the conventional disk drive system, what data is used b which control unit (magnetic disk unit) must be managed in the cache memory. The need for exclusive control in the cache memory is eliminated by providing the cache memory 730 for each control unit (magnetic disk unit).
Next, the read operation of the read unit and the write operation of the write unit shown in Figure 5 are described b showing allocation of data in magnetic disk units. In Figure 32 in the description of the prior art, the striping data array is described as a conventional data array.
Figure 7 shows a spiraling data array which is an improvement of the striping data array in order to lay out effective data to continuously handle data whose transfer length is long.
As shown in Figure 32, to write data continuously in the striping data array, the data is written from left to right in the figure. That is, it is written into data blocks 0, 1, 2, 3 in order and its parity is written at the same time. After they are written as one redundant group, redundant data in the next redundant group is written by shifting write disks by one. Thus, the redundant group begins with data block 4, followed by redundant data, followed by data blocks 5, 6, 7. Thus, the redundant data is shifted slantwise from upper left to lower right in the figure for recording and at the same time, previous data is recorded left and subsequent data is recorded from left to right in order.
In contrast, in the spiraling data array, to write or read data blocks 0 to 24 as shown in Figure 7, magnetic disk units 0 (HDDO) to 4 (HDD4) can be read or written in order.
That is, to read data blocks 0 to 4, HDDO to HDD4 are read in order; to read data blocks 5 to 9, again HDDO to HDD4 may be read in order. Likewise, to read data blocks 10 to 14, HDD0 to HDD4 may be read in order. However, as shown in Figure 7, when data block 3 is read, redundant data P must be read from HDD4 at the same time in an idle read. Thus, redundant data P is idly read or ignored at the same time as the last data in the redundant group is read, thereby enabling continuous data to be obtained simply by reading the magnetic disk units spirally in order.
The spiraling data arra can be provided, when the write unit 00 writes data, by placing redundant data P in a number of magnetic disk units cyclically and beginning with the top data of the write of the next redundant group at the magnetic disk unit into which the redundant data in the preceding redundant group has been written. When data is read from the spiraling data array by the read unit 400, when the last data in one redundant group is read. the redundant data in the redundant group is also read at the same time and is ignored and the top data in the next redundant group is read from the magnetic disk unit where the redundant data in that redundant group has been read.
Next, a modification example of the spiraling data array shown in Figure 7 is described in conjunction with Figure 8.
In Figure 7, one redundant group consists of one data block, in which case the spiraling depth is one. In Figure 8, an example where the spiraling depth is four is discussed. Assume that the spiraling depth 4 means that four continuous logical addresses (data blocks) in one magnetic disk unit are read/written consecutively. Therefore, as shown in Figure 8, data blocks 0 to 3 are recorded in HDDO and data blocks 4 to 7 are recorded in HDD1. Likewise, data blocks 8 to 11 are recorded in HDD2. Thus, data blocks 0 to 14 are recorded in the magnetic disk units and redundant data P as much as the spiraling depth 4 is generated and recorded in HDD4.Although not shown in Figure 8, when the next redundant group is recorded, redundant data P is recorded in HDD3 as shown in Figure 7, and when the redundant group after that is recorded. redundant data P is recorded in HDD2; redundant data P is recorded cyclically as shown in Figure 7.
An access to logical addresses 2 to 5 (data blocks 2 to 5) in the state as shown in Figure 8 is described. To read logical addresses 2-5 (data blocks 2-5). data in HDDO and HDD1 is read and transferred to the FIFO of the read unit of the gate array 310. To transfer data from the FIFO 410 of the read unit 400 to the interface unit 100, the data at logical addresses 0 and 1 is discarded and the data at the remaining logical addresses 2 and 3 is transferred. Likewise, the data at logical addresses 4 and 5 among the data at logical addresses 4 to 7 read from HDD1 is transferred to the interface unit 100 and the data at logical addresses 6 and 7 is not transferred to the interface unit 100 and is instead discarded.
Next, a write of data into logical addresses 2-5 is described. First, data at logical addresses 0-3 and 4-7 is read from HDDO and HDD1 and expanded in the FIFO. Write data transferred through the interface unit 100 is shifted and overlapped on the FIFO data. That is, the data at logical addresses 2 to 5 contained in the FIFO is replaced by the write data at that time. Then, the data in the FIFO is written into HDDO and HDD1. That is, the data at logical addresses 0-3 is written into HDDO and the data at logical addresses 4-7 is written into HDDl.
Next. the read unit 400 is described in detail in conjunction with Figure 9. In Figure 9, numeral 410 is data first-in first-out (FIFO), numeral 420 is an exclusive OR circuit which generates redundant data, numeral 430 is a switch for connection and disconnection of the bus 380 extending from magnetic disk units (not shown) to the FIFO 410, numeral 440 is a switch for connecting redundant data, etc., output from the exclusive OR circuit 420 to the FIFO.
and numeral 450 is a switch for connecting the FIFO 410 to the data bus 170 of the interface unit 100. Numeral 460 is a switch for connecting the FIFO 410 to the exclusive OR circuit 420. Numerals 900 and 910 are transfer buses for outputting the operation results of the exclusive OR circuit 420. Numeral 600 is a control circuit.
The control circuit 600 executes a read operation by switching the switches 430, 440, 450, 460 according to an access pattern and mode. It also monitors a clock to write data into the FIFO 410 and supplies a clock to read data from the FIFO 410. The control circuit 600 contains a sequencer (not shown) which stores a spiraling data array as shown in Figure 7 as a spiraling table. It references the spiraling table to determine the magnetic disk units for reading/writing. The spiraling table also stores the spiraling depth as shown in Figure 8, and the control circuit 600 references the spiraling depth for operation.
Read/write operation can also be performed based on an array other than the spiraling data arrays as shown in Figures 7 and 8 by changing the spiraling table. For example, instances corresponding to RAID level 5 are given in Figures 7 and 8, but the general striping data array corresponding to RAID level 5 as shown in Figure 32 may be handled or an array corresponding to RAID level 4 may be stored.
Next, a read operation is described in conjunction with Figure 9.
First, a continuous data read using the array as shown in Figure 7 is discussed.
To read data continuously, data read from HDDO is input to FIFO 410a and data in data block 1 read from HDD1 is read into FIFO 410b. Likewise, data from HDD2 and data from HDD3 are read into FIFOs 410c and 410d. When data is read into FIFO 410d, redundant data P from HDD4 is also read into FIFO 410e. The control circuit 600 transfers the data in data blocks 0-3 among the data read into FIFOs 410a-410d to the data bus 170 by switching the switches 450 in order. Next, data in data block 4 is read from HDD4 into FIFO 410e and data in data blocks 5-7 is read into FIFOs 410a-410c.When data is read into FIFO 410c, redundant data P is read from HDD3 into FIFO 410d. When the second redundant group is read into the FIFOs 410 in such a manner, the control circuit 600 again transfers data in data blocks 4-7 to the data bus 170 in order by switching the switches 450. The redundant data read into FIFO 410d is ignored.
Next, the operation of the exclusive OR circuit is described. For example, if data read from HDDO is erroneous and data cannot be read normally, the data in HDDO can be restored by reading data in HDD1-HDD4. In this case, the data in the redundant group to which the erroneous data in HDDO belongs is read into FIFOs 410b-410e. The data read into FIFOs 410b-410e is also input via switches 460 to the exclusive OR circuit 420.
Figure 10 is a circuit diagram of the exclusive OR circuit 420.
Outputs of FIFOs 410a-410e are input to five AND gates and then five signals are exclusive-ORed by exclusive OR elements. Since data in HDD1-HDD4 is input to FIFOs 410b410e in the example, the switch 460 corresponding to HDDO is turned off and the switches 460 corresponding to the remaining HDDs or HDD1-HDD4 are turned on, inputting the data to the exclusive OR circuit 420. The exclusive OR circuit 420 exclusixe-ORs the data in HDD1-HDD4 and outputs the result to the transfer bus 910. The control circuit 600 turns on the switch 440 corresponding to HDDO, thereby generating correct data in FIFO 410a. The switches 440 corresponding to HDD1-HDD4 remain off.At the same time as the data in HDD1-HDD4 is read into the FIFOs. the exclusive OR circuit generates correct or reccvery data for each bit and writes it into FIFO 410a. thus the recovery data is generated in FIFO 410a in synchronization with the transfer block of the last transferred read data in HDD1-HDD4. The time formerly required for temporarily reading data in HDD1-HDD4 into memory and then generating recovery data by using the data stored in the memory can be saved.
Figure 11 is a drawing showing the time relationship for generating the recovery data.
In the figure. T represents elapse of time. To read data, first. a seek operation is performed at disk units by a read instruction. then data is read by taking time X as shown in Figure 11. Assume that the time required for reading the data in HDD1-HDD4 is X. Since the exclusive OR circuit 420 as shown in Figure 10 is contained, the time Y shown in Figure 11 is required to recover HDDO. The difference between the end points of the time X and the time Y is about the time required for the exclusive OR circuit 420 to operate, that is, element operation delay time in an ideal case.
Next. the detailed configuration of the write unit 500 is described in conjunction with Figure 12.
In Figure 12, numeral 510 is data first-in first-out (FIFO), numeral 520 is an exclusive OR circuit. numeral 530 is a switch for connecting the data bus 170 of the interface unit 100 and the FIFO 510, numeral 540 is a switch for connecting the exclusive OR circuit 520 and the FIFO 510, and numeral 550 is a switch for connecting the FIFO 510 and the bus 380 to the magnetic disk units. Numeral 560 is a switch for connecting the FIFO 510 and the exclusive OR circuit 520. Numeral 900 is a transfer bus for transferring outputs from the exclusive OR circuit 420 of the read unit shorn in Figure 9.
Next, the operation of the write unit shown in Figure 12 is described.
To continuously write data whose transfer length is long. a spiraling data array as shown in Figure 7 is stored in a spiraling table and the control circuit 600 references the spiraling table for writing continuous data according to the array as shown in Figure 7. That is, the control circuit 600 references the spiraling table and transfers data in data blocks 0-3 to FIFOs 510a-510d. The exclusive OR circuit 520 inputs the data transferred to FIFOs 510a-510d and exclusixe-ORs the data. The exclusive OR circuit 520, which has the circuit configuration as shown in Figure 10, outputs the exclusive ORing result to FIFO SlOe. Therefore.
redundant data of data written into HDDO-HDD3 is generated in FIFO 510e in synchronization with the transfer block of the write data transferred last in FIFOs 510a-510d. The ideal timing at which the redundant data is generated in FIFO lOe is the timing delayed as long as the operation delay of the exclusive OR circuit 520 from the timing at which the data in FIFO 510a-510d is transferred last.
Therefore, the time formerly required for temporarily storing write data in memor and then exclusive ORing the data to find redundant data can be saved. When data is provided in FIFO 510a-510e, the control circuit 600 turns on the switches 50 to connect the FIFOs 510 and the bus 380 to the magnetic disk units and records the data contained in FIFOs in the corresponding magnetic disk units.
Next. an operation example is described in which data is written into one data block rather than data being written continuously. For example, to write data into HDDO, it is necessary to read old data OD written so far in HDDO and redundant data OP in the redundant group containing the old data OD. Assume that the redundant data is recorded in HDD4. To write data into HDDO, it is necessary to first read the old data OD from HDDO and the old redundant data OP from HDD4 and generate new redundant data from the old data OD. the old redundant data OP, and new write data. The old data OD and the old redundant data OP are read by the read unit shown in Figure 9. The read data is input to FIFOs 410a and 410e.The read data is also transferred to the exclusive OR circuit 420 which then exclusive-ORs the old data OD and the old redundant data OP and outputs the result to the transfer bus 900. The transferred data on the transfer bus is sent via the switch 560 to the exclusive OR circuit 520 of the write unit 500.
On the other hand, at the write unit 500, the new write data is input via the data bus 170 to FIFO SlOa and then transferred from FIFO 510a to the exclusive OR circuit 520.
The exclusive OR circuit 520 exclusive-ORs the signal resulting from exclusive ORing the OD and OP transferred on the transfer bus 900 and the new write data transferred from FIFO SlOa, and outputs the resultant data via the switch 540 to FIFO SlOe. Thus, new redundant data is generated in FIFO SlOe at the same time as the new write data is provided in FIFO 510a. The control circuit 600 controls the switches 550 to connect FIFOs 510a and 510e to the bus 380 to the magnetic disk units, thereby recording the data in the corresponding magnetic disk units.Formerly, temporary read data and new data were exclusive-ORed to find redundant data; according to the embodiment of the invention in which the exclusive OR circuit 420 of the read unit and the exclu sie OR circuit 520 of the write unit are located, new redundant data can be generated substantially at the same time as new write data is transferred, and the extra time formerly required for the exclusive OR operation can be saved.
Next, the operation of the read unit and the write unit is described for each mode in conjunction with Figures 13 to 25.
Figure 13 shows a schematic configuration of the read unit 400 shown in Figure 9 and the write unit 500 in Figure 11 with magnetic disk units between. The numerals in Figure 13 are identical with those shown in Figures 9 and 12 except that numeral 525 is P GEN FIFO for recording redundant data outputs b the exclusive OR circuit 520.
Next, the operation of reading data from magnetic disk units is described in conjunction with Figures 14 to 16.
As shown in Figure 14, data in data blocks 2-4 is read from HDD2-HDD4. To perform the read operation. as shown in Figure 15, first, data is read from HDD2-HDD4 into RD FIFOs X2-t4. The disks can be accessed in parallel on condition that RD FIFOs #2-#4 are empty and the write address is reset. Next, when the data in data blocks 2-4 is provided in RD FIFO *2-*4, then the data in RD FIFO #2-#4 is transferred to the data bus 170 in order as shown in Figure 16.
Next, the operation of writing data into one block is described in conjunction with Figures 17 to 20.
First, to write data into data block 1 of HDD1 as shown in Figure 17, new write data is transferred to WRT FIFO #1 and P GEN FIFO 525 as shown in Figure 19. At the same time, the corresponding data is read from HDD1 and HDDS (assume that redundant data is recorded in HDD5) into RD FIFOs *1 and =.
As shown in Figure 20, the data read into RD FIFOs #1 and =5 is transferred via the exclusive OR circuit 420 to the exclusive OR circuit 520. At the same time as the data transfer, the new write data held in P GEN FIFO 525 is input to the exclusive OR circuit 520. Then, the exclusive OR circuit 520 exclusive-ORs the old data. old parity, and new write data, and outputs the resultant data to WRT FIFO X5.
At the last step, the new write data held in WRT FIFO #1 is written into HDD1 and the data held in WRT FIFO X5 is written into HDDS as redundant data.
Next, the group write operation for writing one redundant group into magnetic disk units is described in conjunc- tion with Figures 21 to 25.
As shown in Figure 21, data is recorded in data blocks 1 to 4 of magnetic disk units.
First, new write data is transferred to WRT FIFO #1, as shown in Figure 23. Likewise, new write data is transferred to WRT FIFO =2, as shown in Figure 24; new write data to WRT FIFO =3; and new write data to WRT FIFO ;4. Each time the data is input to the FIFO in order. the exclusive OR circuit 520 generates new redundant data in FIFO 525. When data is transferred to WRT FIFO #1, it is also transferred to P GEN FIFO 525. At the same time as the next data is transferred to WRT FIFO t2, this data and the data already existing in P GEN FIFO 52D are exclusive-ORed and the result is recorded in P GEN FIFO 525.After such an operation is repeated for the data written into HDD1 to the data written into HDD4, the data in the FIFOs are written into the corresponding magnetic disk units, as shown in Figure 25.
Next, the FIFO function is described.
The FIFO is used to temporarily store data in a firstin first-out manner. Therefore. normally the data output in a first-in first-out manner does not remain in the FIFO. but in the example, the data once input to the FIFO is held in the FIFO until a command for clearing the data is received.
To hold data in the FIFO. without shifting the data for output for first-in first-out, for example, a counter or the like is counted up and the data indicated by the counter is output in order, thereby accomplishing the FIFO function. or the data in the FIFO is shifted for output as formerly and the data output from the FIFO is again input to the FIFO so that the data in the FIFO cycles. Thus, the data in the FIFO is not cleared and remains in the FIFO and it can be retransferred.
For example, when data is written. if some fault occurs on a magnetic disk unit and the data output from the FIFO is lost, the lost data can be recovered simply by transferring the data held in the FIFO to the magnetic disk unit without the need for again receiving the data from the host computer, etc. This means that the write operation can be retried immediately. On the other hand, a data read is considered.
After the data in the FIFO is output to the interface unit 100. if some fault occurs on the interface unit 100, etc., and a data reread is requested, the data existing in the FIFO may be again simply output to the interface unit 100 without the need for again reading the data from the magnetic disk unit. This means that the read operation can be retried immediately.
Next, the configuration of the interrupt circuit 340 shown in Figure 5 is described in conjunction with Figure 26.
In Figure 26, numeral 340 is an interrupt circuit which detects operation completion of magnetic disk units and starts the processing program causing the magnetic disk units to operate. Numeral 341 is an interrupt vector number dispatch circuit which dispatches interrupt vector numbers, numeral 342 is a converter which generates interrupt vector numbers. and numeral 343 is an interrupt generation circuit which detects operation completion of magnetic disk units and generates an interrupt signal. Numeral 345 is an AND gate, numeral 346 is an OR gate, and numeral 344 is a register in which a value can be programmably set from the MPU 330.
In operation. when receiving a processing request through the host interface 100, the MPU 330 determines which disks are to be used according to the processing request, and controls the operation of the target magnetic disk units. At the same time, the MPU 330 sets a value in the register 344. For example, if the processing request indicates an access to HDDO and HDD4, the register bits corresponding to HDDO and HDD4 are set to 0 and the remaining bits are set to 1. as shown in Figure 26. The magnetic disk units 800 and the HDD control unit 700, upon completion of their processing, inform the interrupt circuit 340 of the completion. In the example, when HDDO completes processing and when HDD4 completes processing, the interrupt circuit 340 is informed of the completion. At this time, the interrupt generation circuit 343 performs operation on the value of the register 344 and completion information by the OR gates 346, and 1 is output from all OR gates to the AND gate 345 which then outputs 1. The converter 342 detects the interrupt generation circuit which goes on, converts the signal into an interrupt vector number, and then sends it to the interrupt vector number dispatch circuit 341. The interrupt vector number dispatch circuit 341 dispatches the interrupt vector number to the MPU 330 which causes control to jump to the address corresponding to the vector number.
thereby starting the processing program, as shown in Figure 27. For example, if vector number 0 is dispatched, the MPU 330 can start the processing program for task 0 located starting at address 100.
Since five sets of interrupt generation circuits 343 are provided in the example. the MPU 330 can hold up to five processing requests in the stack 350 for processing. By setting the registers 344 of the interrupt generation circuits 343 corresponding to a maximum of five processing requests, when the HDD control unit 700 informs the interrupt circuit 340 of completion of processing, the interrupt generation circuit 343 operates and the interrupt vector number dispatch circuit 341 dispatches the vector number corresponding to the processing request to the MPU 330; therefore, the task corresponding to the processing request can be automatically started from the stack 350.The interrupt circuit 340 has circuitry which generates any interrupt vector number according to the combination condition of completion reports of any HDDs, and the corresponding task can be started in response to the vector number. Therefore, to access magnetic disk units and wait for their operation completion, when the desired magnetic disk unit completes all processing, the waiting task can be automatically distinguished from other tasks for starting. The register 344 can be programmably changed by the MPU 330 and the program started in response to the interrupt vector number can also be programmably changed. thus the interrupt circuit 340 can be made very flexible.The number of magnetic disk units whose operation completion is expected or the numbers of tasks to be started corresponding to processing requests can also be programmably changed and so the sstem configuration can be made very flexible.
Embodiment 2: Although the number of magnetic disk units is five in the first embodiment, it is not limited and another number of magnetic disk units may be used. One or two magnetic disk units are connected to one control unit in the first embodiment, but the number of magnetic disk units connected to one control unit is not limited.
Embodiment 3: Although the dual port RAM is used in the first embodiment, any other memory having a plurality of access ports may be used in place of the dual port RAM.
Embodiment 4: Although the array type recording system comprises magnetic disk units in the first embodiment, it may comprise optical disk units. CD ROM units. semiconductor disk memories (SSDs), or the like, or the recording units used in one system may be of not only the same type. but also different types such as optical disk units and magnetic disk units.
As described above, according to the invention. there is provided the array type recording system in which overhead of read/write processing is reduced and whose performance is improved. Particularly, the redundant data generation operation can be simplified and the time required for the operation can be reduced. Also, the recovery data generation operation can be simplified and the time required for the operation can be reduced. Further. data can be handled continuously and memory transfer and data processing are not required. Further, when data is written into a number of blocks, the number of times a parity write operation is performed can be reduced. Further, read/write operation can be retried immediately.

Claims (3)

1. An array type recording system comprising: (a) a plurality of media for recording data; (b) a plurality of task means for issuing an access request to any desired number of recording media of said plurality of recording media; (c) means for reporting access completion to said recording media accessed by said task means; and (d) task start means being responsive to a combination of reports from said report means for selecting the corresponding task means among said plurality of task means and starting its task.
2. The array type recording system as claimed in claim 1, wherein said recording media are magnetic disks.
3. The array type recording system as claimed in claim 1, wherein said recording media are magneto-optic disks.
GB9608322A 1992-10-06 1993-10-05 Array type recording system Expired - Fee Related GB2298306B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP26736492A JP3181398B2 (en) 1992-10-06 1992-10-06 Array type recording device
GB9320521A GB2271462B (en) 1992-10-06 1993-10-05 Array type recording system

Publications (3)

Publication Number Publication Date
GB9608322D0 GB9608322D0 (en) 1996-06-26
GB2298306A true GB2298306A (en) 1996-08-28
GB2298306B GB2298306B (en) 1997-03-05

Family

ID=26303638

Family Applications (3)

Application Number Title Priority Date Filing Date
GB9608324A Expired - Fee Related GB2298307B (en) 1992-10-06 1993-10-05 Array type recording system
GB9608322A Expired - Fee Related GB2298306B (en) 1992-10-06 1993-10-05 Array type recording system
GB9608325A Withdrawn GB2298308A (en) 1992-10-06 1993-10-05 A disk storage array with a spiralling distribution of redundancy data

Family Applications Before (1)

Application Number Title Priority Date Filing Date
GB9608324A Expired - Fee Related GB2298307B (en) 1992-10-06 1993-10-05 Array type recording system

Family Applications After (1)

Application Number Title Priority Date Filing Date
GB9608325A Withdrawn GB2298308A (en) 1992-10-06 1993-10-05 A disk storage array with a spiralling distribution of redundancy data

Country Status (1)

Country Link
GB (3) GB2298307B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999063446A1 (en) * 1998-06-05 1999-12-09 Mylex Corporation Apparatus and method for computer host system and adaptor interrupt reduction including clustered command completion

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19703009A1 (en) * 1997-01-28 1998-04-02 Siemens Nixdorf Inf Syst Redundant data security system for long-term data archiving and back=up

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999063446A1 (en) * 1998-06-05 1999-12-09 Mylex Corporation Apparatus and method for computer host system and adaptor interrupt reduction including clustered command completion
US6185639B1 (en) * 1998-06-05 2001-02-06 International Business Machines Corporation System and method to reduce a computer system's interrupt processing overhead
US6219727B1 (en) 1998-06-05 2001-04-17 International Business Machines Corporation Apparatus and method for computer host system and adaptor interrupt reduction including clustered command completion

Also Published As

Publication number Publication date
GB9608325D0 (en) 1996-06-26
GB9608322D0 (en) 1996-06-26
GB2298308A (en) 1996-08-28
GB2298306B (en) 1997-03-05
GB2298307B (en) 1997-03-05
GB2298307A (en) 1996-08-28
GB9608324D0 (en) 1996-06-26

Similar Documents

Publication Publication Date Title
US5608891A (en) Recording system having a redundant array of storage devices and having read and write circuits with memory buffers
EP0526487B1 (en) Logical track write scheduling system for a parallel disk drive array data storage subsystem
KR100211788B1 (en) Failure prediction for disk arrays
US5566316A (en) Method and apparatus for hierarchical management of data storage elements in an array storage device
US5155835A (en) Multilevel, hierarchical, dynamically mapped data storage subsystem
US5077736A (en) Disk drive memory
US5210866A (en) Incremental disk backup system for a dynamically mapped data storage subsystem
US6243824B1 (en) Array disk subsystem
US4914656A (en) Disk drive memory
US4989205A (en) Disk drive memory
US5758054A (en) Non-volatile memory storage of write operation identifier in data storage device
US4989206A (en) Disk drive memory
US5479611A (en) Disk array apparatus
KR100275900B1 (en) Method for implement divideo parity spare disk in raid sub-system
US5469453A (en) Data corrections applicable to redundant arrays of independent disks
EP0717357B1 (en) Disk array apparatus
US6961818B1 (en) Method, system and computer program product for managing data in a mirrored cache using an access balancing technique
EP0594464A2 (en) Method of operating a data storage disk array
EP0572564A4 (en)
JP2000020248A (en) Disk fault recovering method, disk driver, and recording medium
GB2298306A (en) A disk array and tasking means
KR19990051729A (en) Structure of Raid System with Dual Array Controllers
WO1992004674A1 (en) Computer memory array control
JPH1031563A (en) Storage device
JP2857289B2 (en) Disk array device

Legal Events

Date Code Title Description
746 Register noted 'licences of right' (sect. 46/1977)

Effective date: 20000127

PCNP Patent ceased through non-payment of renewal fee

Effective date: 20031005