WO2002003387A2 - Data storage system having point-to-point configuration - Google Patents

Data storage system having point-to-point configuration Download PDF

Info

Publication number
WO2002003387A2
WO2002003387A2 PCT/US2001/041175 US0141175W WO0203387A2 WO 2002003387 A2 WO2002003387 A2 WO 2002003387A2 US 0141175 W US0141175 W US 0141175W WO 0203387 A2 WO0203387 A2 WO 0203387A2
Authority
WO
WIPO (PCT)
Prior art keywords
directors
transmitting
receiving
packet
data
Prior art date
Application number
PCT/US2001/041175
Other languages
French (fr)
Other versions
WO2002003387A3 (en
Inventor
David Black
Stephen Macarthur
Richard Wheeler
Paul C. Wilson
Mark Zani
Farouk Khan
Christopher S. Maclellan
John K. Walton
Kendell A. Chilton
William Tuccio
Robert A. Thibault
Original Assignee
Emc Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/606,730 external-priority patent/US7007194B1/en
Priority claimed from US09/606,734 external-priority patent/US7073020B1/en
Application filed by Emc Corporation filed Critical Emc Corporation
Priority to GB0207148A priority Critical patent/GB2374697A/en
Priority to JP2002507375A priority patent/JP4885406B2/en
Publication of WO2002003387A2 publication Critical patent/WO2002003387A2/en
Publication of WO2002003387A3 publication Critical patent/WO2002003387A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F13/38Information transfer, e.g. on bus
    • G06F13/382Information transfer, e.g. on bus using universal interface adapter
    • G06F13/387Information transfer, e.g. on bus using universal interface adapter for adaptation of different data processing systems to different peripheral devices, e.g. protocol converters for incompatible systems, open system

Definitions

  • This invention relates generally to data storage systems, and more particularly to data storage systems having redundancy arrangements to protect against total system failure in the event of a failure in a component or subassembly of the storage system.
  • host computer/servers As is known in the art, large host computers and servers (collectively referred to herein as “host computer/servers”) require large capacity data storage systems. These large computer/servers generally includes data processors, which perform many operations on data introduced to the host computer/server through peripherals including the data storage system. The results of these operations are output to peripherals, including the storage system.
  • One type of data storage system is a magnetic disk storage system.
  • a bank of disk drives and the host computer/server are coupled together through an interface.
  • the interface includes "front end” or host computer/server controllers (or directors) and “back-end” or disk controllers (or directors).
  • the interface operates the controllers (or directors) in such a way that they are transparent to the host computer/server. That is, data is stored in, and retrieved from, the bank of disk drives in such a way that the host computer/server merely thinks it is operating with its own local disk drive.
  • One such system is described in U.S. Patent 5,206,939, entitled "System and
  • the interface may also include, in addition to the host computer/server controllers (or directors) and disk controllers (or directors), addressable cache memories.
  • the cache memory is a semiconductor memory and is provided to rapidly store data from the host computer/server before storage in the disk drives, and, on the other hand, store data from the disk drives prior to being sent to the host computer/server.
  • the cache memory being a semiconductor memory, as distinguished from a magnetic memory as in the case of the disk drives, is much faster than the disk drives in reading and writing data.
  • the host computer/server controllers, disk controllers and cache memory are interconnected through a backplane printed circuit board. More particularly, disk controllers are mounted on disk controller printed circuit boards.
  • the host computer/server controllers are mounted on host computer/server controller printed circuit boards.
  • cache memories are mounted on cache memory printed circuit boards.
  • the disk directors, host computer/server directors, and cache memory printed circuit boards plug into the backplane printed circuit board, h order to provide data integrity in case of a failure in a director, the backplane printed circuit board has a pair of buses. One set the disk directors is connected to one bus and another set of the disk directors is connected to the other bus. Likewise, one set the host computer/server directors is connected to one bus and another set of the host computer/server directors is directors connected to the other bus.
  • the cache memories are connected to both buses. Each one of the buses provides data, address and control information.
  • the use of two buses Bl, B2 provides a degree of redundancy to protect against a total system failure in the event that the controllers or disk drives connected to one bus, fail. Further, the use of two buses increases the data transfer bandwidth of the system compared to a system having a single bus.
  • the host computer/server 12 issues a write request to one of the front-end directors 14 (i.e., host computer/server directors) to perform a write command.
  • One of the front-end directors 14 replies to the request and asks the host computer 12 for the data.
  • the director 14 determines the size of the data and reserves space in the cache memory 18 to store the request.
  • the front-end director 14 then produces control signals on one of the address memory busses Bl, B2 connected to such front-end director 14 to enable the transfer to the cache memory 18.
  • the host computer/server 12 then transfers the data to the front-end director 14.
  • the front- end director 14 advises the host computer/server 12 that the transfer is complete.
  • the front- end director 14 looks up in a Table, not shown, stored in the cache memoiy 18 to determine which one of the back-end directors 20 (i.e., disk directors) is to handle this request.
  • the Table maps the host computer/server 12 addresses into an address in the bank 14 of disk drives.
  • the front-end director 14 then puts a notification in a "mail box" (not shown and stored in the cache memory 18) for the back-end director 20, which is to handle the request, the amount of the data and the disk address for the data.
  • Other back-end directors 20 poll the cache memory 18 when they are idle to check their "mail boxes”. If the polled "mail box" indicates a transfer is to be made, the back-end director 20 processes the request, addresses the disk drive in the bank 22, reads the data from the cache memory 18 and writes it into the addresses of a disk drive in the bank 22.
  • a read request is instituted by the host computer/server 12 for data at specified memory locations (i.e., a requested data block).
  • One of the front-end directors 14 receives the read request and examines the cache memory 18 to determine whether the requested data block is stored in the cache memory 18. If the requested data block is in the cache memory 18, the requested data block is read from the cache memory 18 and is sent to the host computer/server 12.
  • the front-end director 14 determines that the requested data block is not in the cache memory 18 (i.e., a so- called "cache miss") and the director 14 writes a note in the cache memory 18 (i.e., the "mail box") that it needs to receive the requested data block.
  • the back-end directors 20 poll the cache memory 18 to determine whether there is an action to be taken (i.e., a read operation of the requested block of data).
  • the one of the back-end directors 20 which poll the cache memory 18 mail box and detects a read operation reads the requested data block and initiates storage of such requested data block stored in the cache memory 18. When the storage is completely written into the cache memory 18, a read complete indication is placed in the "mail box" in the cache memory 18.
  • front-end directors 14 are polling the cache memory 18 for read complete indications. When one of the polling front-end directors 14 detects a read complete indication, such front-end director 14 completes the transfer of the requested data which is now stored in the cache memory 18 to the host computer/server 12.
  • a system interface includes a plurality of first directors, a plurality of second directors, a data transfer section and a message network.
  • the data transfer section includes a cache memory.
  • the cache memory is coupled to the plurality of first and second directors.
  • the messaging network operates independently of the data transfer section and such network is coupled to the plurality of first directors and the plurality of second directors.
  • the first and second directors control data transfer between the first directors and the second directors in response to messages passing between the first directors and the second directors through the messaging network to facilitate data transfer between first directors and the second directors.
  • the data passes through the cache memory in the data transfer section.
  • the system interface each one of the first directors includes a data pipe coupled between an input of such one of the first directors and the cache memory and a controller for transferring the messages between the message network and such one of the first directors.
  • each one of the second directors includes a data pipe coupled between an input of such one of the second directors and the cache memory and a controller for transferring the messages between the message network and such one of the second directors.
  • the directors includes: a data pipe coupled between an input of such one of the first directors and the cache memory; a microprocessor; and a controller coupled to the microprocessor and the data pipe for controlling the transfer of the messages between the message network and such one of the first directors and for controlling the data between the input of such one of the first directors and the cache memory.
  • a data storage system is provided for transferring data between a host computer/server and a bank of disk drives through a system interface.
  • the system interface includes a plurality of first directors coupled to host computer/server, a plurality of second directors coupled to the bank of disk drives, a data transfer section, and a message network.
  • the data transfer section includes a cache memory.
  • the cache memory is coupled to the plurality of first and second directors.
  • the message network is operative independently of the data transfer section and such network is coupled to the plurality of first directors and the plurality of second directors.
  • the first and second directors control data transfer between the host computer and the bank of disk drives in response to messages passing between the first directors and the second directors through the messaging network to facilitate the data transfer between host computer/server and the bank of disk drives with such data passing through the cache memory in the data transfer section.
  • a method for operating a data storage system adapted to transfer data between a host computer/server and a bank of disk drives.
  • the method includes transferring messages through a messaging network with the data being transferred between the host computer/server and the bank of disk drives through a cache memory, such message network being independent of the cache memory.
  • a method is provided for operating a data storage system adapted to transfer data between a host computer/server and a bank of disk drives through a system interface.
  • the interface includes a plurality of first directors coupled to host computer/server, a plurality of second directors coupled to the bank of disk drives; and a data transfer section having a cache memory, such cache memory being coupled to the plurality of first and second directors.
  • the method comprises transferring the data between the host computer/server and the bank of disk drives under control of the first and second directors in response to messages passing between the first directors and the second directors through a messaging network to facilitate the data transfer between host computer/server and the bank of disk drives with such data passing through the cache memory in the data transfer section, such message network being independent of the cache memory.
  • FIG. 1 is a block diagram of a data storage system according to the PRIOR ART
  • FIG. 2 is a block diagram of a data storage system according to the invention.
  • FIG. 2 A shows the fields of a descriptor used in the system interface of the data storage system of FIG. 2;
  • FIG. 2B shows the filed used in a MAC packet used in the system interface of the data storage system of FIG, 2;
  • FIG. 3 is a sketch of an electrical cabinet storing a system interface used in the data storage system of FIG. 2;
  • FIG. 4 is a diagramatical, isometric sketch showing printed circuit boards providing the system interface of the data storage system of FIG. 2;
  • FIG. 5 is a block diagram of the system interface used in the data storage system of FIG.
  • FIG. 6 is a block diagram showing the connections between front-end and back-end directors to one of a pair of message network boards used in the system interface of the data storage system of FIG. 2;
  • FIG. 7 is a block diagram of an exemplary one of the director boards used in the system interface of he data storage system of FIG. 2;
  • FIG. 8 is a block diagram of the system interface used in the data storage system of FIG.
  • FIG. 8A is a diagram of an exemplary global cache memory board used in the system interface of FIG. 8;
  • FIG. 8B is a diagram showing a pair of director boards coupled between a pair of host processors and global cache memory boards used in the system interface of FIG. 8;
  • FIG. 8C is a block diagram of an exemplary crossbar switch used in the front-end and rear-end directors of the system interface of FIG. 8;
  • FIG. 9 is a block diagram of a transmit Direct Memory Access (DMA) used in the system interface of the FIG. 8;
  • DMA Direct Memory Access
  • FIG. 10 is a block diagram of a receive DMA used in the system interface of FIG.8;
  • FIG. 11 shows the relationship between FIGS. 11 A and 1 IB, such FIGS. 11 A and 1 IB together showing a process flow diagram of the send operation of a message network used in the system interface of FIG. 8;
  • FIGS. 1 lC-1 IE are examples of digital words used by the message network in the system interface of FIG. 8;
  • FIG. 1 IF shows bits in a mask used in such message network
  • FIG. 11G shows the result of the mask of FIG. 1 IF applied to the digital word shown in FIG. HE;
  • FIG. 12 shows the relationship between FIGS. 12A and 12B, such FIGS. 12A and 12B Showing a process flow diagram of the receive operation of a message network used in the system interface of FIG. 8;
  • FIG .13 shows the relationship between FIGS. 11A and 1 IB, such FIGS. 11 A and 1 IB together showing a process flow diagram of the acknowledgement operation of a message network used in the system interface of FIG. 8;
  • FIGS. 14A and 14B show process flow diagrams of the transmit DMA operation of the transmit DMA of FIG. 9.
  • FIGS. 15A and 15B show process flow diagrams of the receive DMA operation of the receive DMA of FIG. 10.
  • the system interface 160 includes: a plurality of, here 32 front-end directors 180 ⁇ -180 32 coupled to the host computer/server 120 via ports-123 32 ; a plurality of back-end directors 200!- 200 3 coupled to the bank of disk drives 140 via ports 123 33 -123 6 ; a data transfer section 240, having a global cache memory 220, coupled to the plurality of front-end directors 180 ⁇ -l 80 16 and the back-end directors 200]-200 16 ; and a messaging network 260, operative independently of the data transfer section 240, coupled to the plurality of front-end directors 180 1 -180 32 and the plurality of back-end directors 200 1 -200 32 , as shown .
  • the front-end and back-end directors 180 ⁇ -180 32 , 200 ⁇ -200 32 are functionally similar and include a microprocessor ( ⁇ P) 299 (i.e., a central processing unit (CPU) and RAM), a message engine/ CPU controller 314 and a data pipe 316 to be described in detail in connection with FIGS. 5, 6 and 7. Suffice it to say here, however, that the front-end and back-end directors 180 1 -18 ⁇ 32j 200 1 -20 ⁇ 32 control data transfer between the host computer/server 120 and the bank of disk drives 140 in response to messages passing between the directors I8O1-I8O32, 200 1 -20 ⁇ 32 through the messaging network 260.
  • ⁇ P microprocessor
  • the messages facilitate the data transfer between host computer/server 120 and the bank of disk drives 140 with such data passing through the global cache memory 220 via the data transfer section 240. More particularly, in the case of the front-end directors I8O 1 -I8O 32 , the data passes between the host computer to the global cache memory 220 through the data pipe 316 in the front-end directors 180 1 -18 ⁇ 32 and the messages pass through the message engine/CPU controller 314 in such front-end directors 180 1 -18 ⁇ 3 2 .
  • the data passes between the back-end directors 200 1 -20 ⁇ 32 and the bank of disk drives 140 and the global cache memory 220 through the data pipe 316 in the back-end directors 200 1 -20 ⁇ 32 and again the messages pass through the message engine/CPU controller 314 in such back-end director 200i- 200 32 .
  • the cache memory 220 in the data transfer section 240 is not burdened with the task of transferring the director messaging. Rather the messaging network 260 operates independent of the data transfer section 240 thereby increasing the operating bandwidth of the system interface 160.
  • the request is passed from one of a plurality of, here 32, host computer processors in the host computer 120 to one or more of the pair of the front-end directors 180i-18 ⁇ 3 2 connected to such host computer processor 121 1 -121 32 .
  • each one of the host computer processors 121 1 -121 32 is coupled to here a pair (but not limited to a pair) of the front-end directors I8O 1 -I8O 32, to provide redundancy in the event of a failure in one of the front end-directors I8I 1 -I8I 32 coupled thereto.
  • the bank of disk drives 140 has a plurality of, here 32, disk drives 141 ⁇ -141 32 , each disk drive 141H41 32 being coupled to here apair (but not limited to a pair) of the back-end directors 200 ⁇ 200 32 , to provide redundancy in the event of a failure in one of the back-end directors 200 1 -20 ⁇ 32 coupled thereto).
  • Each front-end director l ⁇ Oi-1803 2 includes a microprocessor ( ⁇ P) 299 (i.e., a central processing unit (CPU) and RAM) and will be described in detail in connection with FIGS. 5 and 7. Suffice it to say here, however, that the microprocessor 299 makes a request for the data from the global cache memory 220.
  • the global cache memory 220 has a resident cache management table, not shown. Every director 180 ⁇ -l 8 ⁇ 3 2 , 200 ⁇ 2003 2 has access to the resident cache management table and every time a front-end director 180 1 -18 ⁇ 32 requests a data transfer, the front-end director I8O 1 -I8O3 2 must query the global cache memory 220 to determine whether the requested data is in the global cache memory 220.
  • the front-end director 180j - 18O 32 mediates a DMA (Direct Memory Access) operation for the global cache memory 220 and the requested data is transferred to the requesting host computer processor 121 1 -12l3 2 .
  • DMA Direct Memory Access
  • the front-end director 180 1 -18 ⁇ 3 2 receiving the data request determines that the requested data is not in the global cache memory 220 (i.e., a "miss") as a result of a query of the cache management table in the global cache memory 220, such front-end director 18 ⁇ !-18 ⁇ 3 2 concludes that the requested data is in the bank of disk drives 140.
  • the front-end director 180 1 -18 ⁇ 32 that received the request for the data must make a request for the data from one of the back-end directors 20 ⁇ !-20 ⁇ 32 in order for such back-end director 200 ⁇ 2003 2 to request the data from the bank of disk drives 140.
  • the mapping of which back-end directors 200 1 -2003 2 control which disk drives 141i-14l 32 in the bank of disk drives 140 is determined during a power-up initialization phase.
  • the map is stored in the global cache memory 220.
  • the front-end director 180 ⁇ -18O 32 makes a request for data from the global cache memory 220 and determines that the requested data is not in the global cache memory 220 (i.e., a "miss")
  • the front-end director I8O 1 -I8O 32 is also advised by the map in the global cache memory 220 of the back-end director 200 20 ⁇ 32 responsible for the requested data in the bank of disk drives 140.
  • the requesting front-end director 18O1-I8O32 then must make a request for the data in the bank of disk drives 140 from the map designated back-end director 200 20 ⁇ 3 2 .
  • This request between the front-end director 180rl8 ⁇ 3 2 and the appropriate one of the back-end directors 2OO1-2OO32 is by a message which passes from the front-end director 18O 1 -18O32 through the message network 260 to the appropriate back- end director 200 ⁇ 200 32 - It is noted then that the message does not pass through the global cache memory 220 (i.e., does not pass through the data transfer section 240) but rather passes through the separate, independent message network 260. Thus, communication between the directors 180 1 -180 32 , 20 ⁇ ! -20 ⁇ 32 is through the message network 260 and not through the global cache memory 220. Consequently, valuable bandwidth for the global cache memory 220 is not used for messaging among the directors I8O1-I8O32, 200i-20 ⁇ 32.
  • the front-end director 180 ⁇ 180 32 sends a message to the appropriate one of the back-end directors 200 ⁇ 20032 through the message network 260 to instruct such back-end director 200i-20 ⁇ 3 2 to transfer the requested data from the bank of disk drives 140 to the global cache memory 220.
  • the back-end director 200 ⁇ 20032 advises the requesting front-end director I8O1-I8O3 2 that the transfer is accomplished by a message, which passes from the back-end director 20 ⁇ !-20 ⁇ 3 2 to the front-end director 180 1 -18 ⁇ 3 2 through the message network 260.
  • the front-end director 180rl 8O32 is thereby advised that such front-end director 180 ⁇ 18 ⁇ 3 2 can transfer the data from the global cache memory 220 to the requesting host computer processor 121 1 -12l 32 as described above when there is a cache "read hit". It should be noted that there might be one or more back-end directors 200 1 -200 32 responsible for the requested data. Thus, if only one back-end director 20 ⁇ !-200 32 is responsible for the requested data, the requesting front-end director 180 ⁇ 18O 32 sends a uni-cast message via the message network 260 to only that specific one of the back-end directors 200 ⁇ 200 32 .
  • a multi-cast message (here implemented as a series of uni-cast messages) is sent by the requesting one of the front-end directors I8O1-I8O3 2 to all of the back-end directors 2OO1-2OO3 2 having responsibility for the requested data.
  • a uni-cast or multi-cast message such message is passed through the message network 260 and not through the data transfer section 240 (i.e., not through the global cache memory 220).
  • the acknowledgement signal may be sent to the requesting host computer processor 121 1 or one or more other host computer processors I2I 1 -I2I 32 via a multi-cast (i.e., sequence of uni-cast) messages through the message network 260 to complete the data read operation.
  • the host computer 120 wishes to write data into storage (i.e., into the bank of disk drives 140).
  • One of the front-end directors I8O1-I8O32 receives the data from the host computer 120 and writes it into the global cache memory 220.
  • the front-end director 18O 1 -18O 32 then requests the transfer of such data after some period of time when the back-end director 200r-20 ⁇ 3 2 determines that the data can be removed from such cache memory 220 and stored in the bank of disk drives 140.
  • the data in the cache memory 220 is tagged with a bit as "fresh data" (i.e., data which has not been transferred to the bank of disk drives 140, that is data which is "write pending").
  • fresh data i.e., data which has not been transferred to the bank of disk drives 140, that is data which is "write pending"
  • the front-end director 1801- 18O3 2 controlling the transfer also informs the host computer 120 that the transfer is complete to thereby free-up the host computer 120 for other data transfers.
  • the back-end director 200 ⁇ 200 32 transfers the data from the global cache memory 220 to the bank of disk drives 140 and resets the tag associated with data in the global cache memory 220 (i.e., un-tags the data) to indicate that the data in the global cache memory 220 has been transferred to the bank of disk drives 140. It is noted that the un-tagged data in the global cache memory 220 remains there until overwritten with new data.
  • the system interface 160 is shown to include an electrical cabinet 300 having stored therein: a plurality of, here eight front-end director boards 190 ⁇ -190 8 , each one having here four of the front-end directors 180 ⁇ 180 32 ; a plurality of, here eight back- end director boards 210 1 -210 8 , each one having here four of the back-end directors and a plurality of, here eight, memory boards 220' which together make up the global cache memory 220.
  • These boards plug into the front side of a backplane 302. (It is noted that the backplane 302 is a mid-plane printed circuit board). Plugged into the backside of the backplane 302 are message network boards 304 l5 304 2 .
  • the backside of the backplane 302 has plugged into it adapter boards, not shown in FIGS. 2-4, which couple the boards plugged into the back-side of the backplane 302 with the computer 120 and the bank of disk drives 140 as shown in FIG. 2. That is, referring again briefly to FIG. 2, an I/O adapter, not shown, is coupled between each one of the front-end directors 180 ⁇ -18 ⁇ 32 and the host computer 120 and an I/O adapter, not shown, is coupled between each one of the back-end directors 200 20 ⁇ 3 2 and the bank of disk drives 140. Referring now to FIG.
  • the system interface 160 is shown to include the director boards 190 ⁇ _ 190 8 , 210 ⁇ 210 8 and the global cache memory 220 , plugged into the backplane 302 and the disk drives 141 ⁇ -141 32 in the bank of disk drives along with the host computer 120 also plugged into the backplane 302 via I/O adapter boards, not shown.
  • the message network 260 (FIG. 2) includes the message network boards 304 1 and 304 2 - Each one of the message network boards 304 t and 304 2 is identical in construction. A pair of message network boards 304] and 304 2 is used for redundancy and for message load balancing.
  • each message network board 304 l5 304 2 includes a controller 306, (i.e., an initialization and diagnostic processor comprising a CPU, system controller interface and memory, as shown in FIG. 6 for one of the message network boards 304 ⁇ , 304 2 , here board 3040 and a crossbar switch section 308 (e.g., a switching fabric made up of here four switches 308 ⁇ 308 ⁇ .
  • a controller 306 i.e., an initialization and diagnostic processor comprising a CPU, system controller interface and memory, as shown in FIG. 6 for one of the message network boards 304 ⁇ , 304 2 , here board 3040 and a crossbar switch section 308 (e.g., a switching fabric made up of here four switches 308 ⁇ 308 ⁇ .
  • each one of the director boards 190 21 ⁇ 8 includes, as noted above four of the directors I8O1-I8O32, 20 ⁇ !-200 3 2 (FIG. 2). It is noted that the director boards 190 ⁇ - 190s having four front-end directors per board, 180 ⁇ 18O 32 are referred to as front-end directors, and the director boards 210 1 -210s having four back-end directors per board, 200 ⁇ 200 32 are referred to as back-end directors.
  • Each one of the directors 180 ⁇ 18032, 200 ⁇ 2003 2 includes a CPU 310, a RAM 312 (which make up the microprocessor 299 referred to above), the message engine/CPU controller 314, and the data pipe 31 .
  • Each one of the director boards 190 ⁇ 2108 includes a crossbar switch 318.
  • the crossbar switch 318 has four input/output ports 319, each one being coupled to the data pipe 316 of a corresponding one of the four directors I8O1-I8O32, 200 20 ⁇ 32 on the director board.l90!-2108.
  • the crossbar switch 318 has eight output input ports collectively identified in FIG. 5 by numerical designation 321 (which plug into the backplane 302.
  • the crossbar switch 318 on the front-end director boards 191 ⁇ -191g is used for coupling the data pipe 316 of a selected one of the four front-end directors 180 ⁇ -1803 2 on the front-end director board 190rl9 ⁇ 8 to the global cache memory 220 via the backplane 302 and I/O adapter, not shown.
  • the crossbar switch 318 on the back-end director boards 210 ⁇ 2108 is used for coupling the data pipe 316 of a selected one of the four back-end directors 200 1 -20 ⁇ 32 on the back-end director board 210 ⁇ 210s to the global cache memory 220 via the backplane 302 and I/O adapter, not shown.
  • the data pipe 316 in the front-end directors 180 l8 ⁇ 32 couples data between the host computer 120 and the global cache memory 220 while the data pipe 316 in the back-end directors 200 ⁇ 20032 couples data between the bank of disk drives 140 and the global cache memory 220.
  • the backplane 302 is a passive backplane because it is made up of only etched conductors on one or more layers of a printed circuit board. That is, the backplane 302 does not have any active components.
  • each one of the director boards 190 ⁇ 2108 includes a crossbar switch 320.
  • Each crossbar switch 320 has four input/output ports 323, each one of the four input/output ports 323 being coupled to the message engine/CPU controller 314 of a corresponding one of the four directors 180 1 -180 3 , 200 ⁇ 200 32 on the director board 190]-210 8 .
  • Each crossbar switch 320 has a pair of output input ports 325 l9 325 2 , which plug into the backplane 302.
  • Each port 325 ⁇ - 3252 is coupled to a corresponding one of the message network boards 304 l5 304 2 , respectively, through the backplane 302.
  • the crossbar switch 320 on the front-end director boards 190 l9 ⁇ 8 is used to couple the messages between the message engine/CPUcontroller 314 of a selected one of the four front-end directors 180 ⁇ -l 80 32 on the front-end director boards 190 ⁇ -190g and the message network 260, FIG. 2.
  • the back-end director boards 210r210 8 are used to couple the messages produced by a selected one of the four back-end directors 200 ⁇ 2003 2 on the back-end director board 21 Oi-210s between the message engine/CPU controller 314 of a selected one of such four back-end directors and the message network 260 (FIG. 2).
  • the message network 260 (FIG. 2) includes the crossbar switch 320 and the message network boards 304 ls 304 2 . Each message is a 64-byte descriptor, shown in FIG. 2 A) which is created by the CPU 310 (FIG.
  • Each one of the MAC packets has: an 8-byte header which includes source (i.e., transmitting director) and destination (i.e., receiving director) address; a payload; and terminates with a 4-byte Cyclic Redundancy Check (CRC), as shown in FIG. 2B.
  • the acknowledgement packet i.e., signal
  • the message packet has a 32-byte payload section.
  • the Fabric Management Packet (FMP) has a 256-byte payload section.
  • the MAC packet is sent to the crossbar switch 320.
  • the destination portion of the packet is used to indicate the destination for the message and is decoded by the switch 320 to determine which port the message is to be routed.
  • the decoding process uses a decoder table 327 in the switch 318, such table being initialized by controller during power-up by the initialization and diagnostic processor (controller) 306 (FIG. 5).
  • the table 327 (FIG. 7) provides the relationship between the destination address portion of the MAC packet, which identifies the routing for the message and the one of the four directors 180 ⁇ 18O 32 , 200 1 -2003 2 on the director board 190 ⁇ -19 ⁇ 8 , 210r21 ⁇ 8 or to one of the message network boards 304 ⁇ , 304 2 to which the message is to be directed.
  • each one of the message network boards 304 l5 304 2 has sixteen input/output ports 322 ⁇ 322 ⁇ , each one being coupled to a corresponding one of the output/input ports 325 1 , 325 2 , respectively, of a corresponding one of the director boards 190rl9 ⁇ 8 , 210 ⁇ -21 ⁇ 8 through the backplane 302, as shown.
  • FIG. 5 considering exemplary message network board 304 ls FIG.
  • each switch 308 1 -308 4 also includes three coupling ports 324 324 3 .
  • the coupling ports 324 ⁇ 324 3 are used to interconnect the switches 322i-322 4 , as shown in FIG. 6.
  • input/output ports 322 ⁇ 322 8 are coupled to output/input ports 325 ⁇ of front-end director boards 190 ⁇ 190 8 and input/output ports 322 9 -322 16 are coupled to output/input ports 325 t of back-end director boards 210 ⁇ -210 8 , as shown.
  • each one of the message network boards 304 ⁇ , 304 2 includes a processor 306 (FIG. 5) and a crossbar switch section 308 having four switches 308 ⁇ -308 4 , as shown in FIGS. 5 and 6.
  • the switches 308 1 -308 4 are interconnected as shown so that messages can pass between any pair of the input/output ports 322 1 -322 16 .
  • a message from any one of the front-end directors 180 l8 ⁇ 32 can be coupled to another one of the front-end directors 180 ⁇ I8O 32 and/or to any one of the back-end directors 200 ⁇ -200 3 2.
  • a message from any one of the back-end directors 180 ⁇ 180 32 can be coupled to another one of the back-end directors 180 ⁇ _l 8O32 and/or to any one of the front-end directors 200 1 -20 ⁇ 3 .
  • each MAC packet (FIG. 2B) includes in an address destination portion and a data payload portion.
  • the MAC header is used to indicate the destination for the MAC packet and such MAC header is decoded by the switch to determine which port the MAC packet is to be routed.
  • the decoding process uses a table in the switch 308 ⁇ -308 4 , such table being initialized by processor 306 during power-up. The table provides the relationship between the MAC header, which identifies the destination for the MAC packet and the route to be taken through the message network.
  • the switches 320 and the switches 308 1 -308 in switch section 308 provides packet routing which enables each one of the directors 20 ⁇ !-200 32 to transmit a message between itself and any other one of the directors, regardless of whether such other director is on the same director board 190 1 - 190s, 210 ⁇ -210s or on a different director board.
  • the MAC packet has an additional bit B in the header thereof, as shown in FIG.
  • this additional bit B is toggled between a logic 1 and a logic 0 so that one message passes through one of the redundant message network boards 304 ls 3042 and the next message to pass through the other one of the message network boards 304 1; 304 2 to balance the load requirement on the system.
  • the non-failed one of the boards 304 ⁇ , 304 2 is used exclusively until the failed message network board is replaced.
  • an exemplary one of the director boardsl ⁇ Ot-l ⁇ Os, 210 1 -210s, here director board 190 ⁇ is shown to include directors 180 1? I8O3, 18O 5 and 180 7 .
  • An exemplary one of the directors 180 1 -180 , here director I8O 1 is shown in detail to include the data pipe 316, the message engine/CPU controller 314, the RAM 312, and the CPU 310 all coupled to the CPU interface bus 317, as shown.
  • the exemplary director 180 ⁇ also includes: a local cache memory 319 (which is coupled to the CPU 310); the crossbar switch 318; and, the crossbar switch 320, described briefly above in connection with FIGS. 5 and 6.
  • the data pipe 316 includes a protocol translator 400, a quad port RAM 402 and a quad port RAM controller 404 arranged as shown.
  • the protocol translator 400 converts between the protocol of the host computer 120, in the case of a front-end director 180 ⁇ 18032, (and between the protocol used by the disk drives in bank 140 in the case of a back-end director 200 1 -20 ⁇ 32 ) and the protocol between the directors 18O 1 -180 3 , 2OO 1 -2OO 32 and the global memory 220 (FIG. 2).
  • the protocol used the host computer 120 may, for example, be fibre channel, SCSI, ESCON or FICON, for example, as determined by the manufacture of the host computer 120 while the protocol used internal to the system interface 160 (FIG.
  • the quad port RAM 402 is a FIFO controlled by controller 404 because the rate data coming into the RAM 402 may be different from the rate data leaving the RAM 402.
  • the RAM 402 has four ports, each adapted to handle an 18 bit digital word.
  • the protocol translator 400 produces 36 bit digital words for the system interface 160 (FIG. 2) protocol, one 18 bit portion of the word is coupled to one of a pair of the ports of the quad port RAM 402 and the other 18 bit portion of the word is coupled to the other one of the pair of the ports of the quad port RAM 402.
  • the quad port RAM has a pair of ports 402A, 402B, each one of to ports 402A, 402B being adapted to handle an 18 bit digital word.
  • Each one of the ports 402 A, 402B is independently controllable and has independent, but arbitrated, access to the memory array within the RAM 402. Data is transferred between the ports 402A, 402B and the cache memory 220 (FIG. 2) through the crossbar switch 318, as shown.
  • the crossbar switch 318 includes a pair of switches 406A, 406B.
  • Each one of the switches 406A, 406B includes four input/output director-side ports D ⁇ -D (collectively referred to above in connection with FIG. 5 as port 319) and four input/output memory-side ports M 1 -M4, Ms-Ms, respectively, as indicated.
  • the input/output memory-side ports MrM , M 5 -M 8 were collectively referred to above in connection with FIG. 5 as port 317).
  • the director-side ports DrD 4 of switch 406A are connected to the 402 A ports of the quad port RAMs 402 in each one the directors 180 l5 I8O 3 , 180 5 and I8O 7 , as indicated.
  • director-side ports of switch 406B are connected to the 402B ports of the quad port RAMs 402 in each one the directors 180 l5 18O 3 , 180 5 , and 180 7 , as indicated.
  • the ports D 1 -D 4 are selectively coupled to the ports M 1 -M 4 in accordance with control words provided to the switch 406A by the controllers in directors I8O 1 , 18O 3 , 18O 5 , 180 on busses R A rRA4, respectively, and the ports D D 4 are coupled to ports M 5 -Ms in accordance with the control words provided to switch 406B by the controllers in directors 180 ls 18O 3 , 18O 5 , I8O7 on busses RBI-R B 4 5 as indicated.
  • the signals on buses RAI-RA4 are request signals.
  • port 402A of any one of the directors 180 l5 I8O 3 , 18O 5 , 180 ⁇ may be coupled to any one of the ports M1-M 4 of switch 406A, selectively in accordance with the request signals on buses RA I - RA .
  • port 402B of any one of the directors 18 ⁇ ! -180 4 may be coupled to any one of the ports M 5 -Mg of switch 406B, selectively in accordance with the request signals on buses RBI- R B4 .
  • the coupling between the director boards 190 ⁇ 190 8 , 210 ⁇ 210s and the global cache memory 220 is shown in FIG. 8.
  • each one of the host computer processors 121 1 -I2I 32 in the host computer 120 is coupled to a pair of the front-end directors I8O 1 -I8O 32, to provide redundancy in the event of a failure in one of the front end-directors 1811- I8I 32 coupled thereto.
  • the bank of disk drives 140 has a plurality of, here 32, disk drives 141 1 -14l 32 , each disk drive 141 ⁇ -14l 32 being coupled to a pair of the back-end directors 2OO1-2OO 32, to provide redundancy in the event of a failure in one of the back-end directors 200!- 2OO 32 coupled thereto).
  • exemplary host computer processorl21 ls such processor 121 ⁇ is coupled to a pair of front-end directors 180 ⁇ , I8O 2 .
  • director I8O1 fails, the host computer processor 121 ⁇ can still access the system interface 160, albeit by the other front-end director 180 2 .
  • directors I8O 1 and I8O 2 are considered redundancy pairs of directors.
  • front-end directors I8O 3 , 180 4 ; I8O5, 180 6; I8O7, 180 8; 180 9 , I8O10; I8O11, 18O12; I8O13, 18O14; I8O1 5 , 180 16; 180 17 , 180 l8; I8O19, 18O20; I8O21, 18O22; 180 23 , 180 24 ; I8O25, 180 26 ; 180 27 , 180 28 ; 180 29 , 180 30 ; and I8O31, 180 32 (only directors I8O 31 and I8O 32 being shown in FIG. 2).
  • disk drive 141 1 is coupled to a pair of back-end directors 200 l5 200 2 .
  • director 200 ! fails, the disk drive 141 1 can still access the system interface 160, albeit by the other back- end director 180 2 .
  • directors 200! and 2OO2 are considered redundancy pairs of directors.
  • other redundancy pairs of back-end directors are: back-end directors 200 3 , 200 ; 200 , 200 6; 200 7 ,200 8; 200 9; 20 ⁇ !
  • the global cache memory 220 includes a plurality of, here eight, cache memory boards 22 ⁇ ! -220s, as shown. Still further, referring to FIG. 8 A, an exemplary one of the cache memory boards, here board 220 1 is shown in detail and is described in detail in U. S. Patent No.
  • the board 220i includes a plurality of, here four RAM memory arrays, each one of the arrays has a pair of redundant ports, i.e., an A port and a B port.
  • the board itself has sixteen ports; a set of eight A ports MA I -MA S and a set of eight B ports M B !- MB 8 -
  • a ports MAI-MA are coupled to the M ⁇ port of each of the front-end director boards 190 l5 190 3 , 190s, and 190 7 , respectively, as indicated in FIG. 8.
  • B ports M B1 -MB 4 are coupled to the Mi port of each of the front-end director boards 190 2 , 190 4 , 190 6 , and 190s, respectively, as indicated in FIG. 8.
  • the other four of the eight A port here A ports MA S - A S are coupled to the Mi port of each of the back-end director boards 210 ⁇ , 210 3 , 210 5 , and 210 7 , respectively, as indicated in FIG. 8.
  • the other four of the eight B port here B ports M B5 -M 48 are coupled to the Mi port of each of the back-end director boards 2IO 2 , 210 4 , 210 6 , and 210 8 , respectively, as indicated in FIG.
  • each one of the four A ports MAI-MA4 can be coupled to the A port of any one of the memory arrays through the logic network 2211 A-
  • port MAI such port can be coupled to the A port of the four memory arrays.
  • each one of the four A ports MAS-MAS can be coupled to the A port of any one of the memory arrays through the logic network 221 m.
  • each one of the four B ports MB 1 -MB 4 can be coupled to the B port of any one of the memory arrays through logic network 221 I B .
  • each one of the four B ports M BS -MB S can be coupled to the B port of any one of the memory arrays through the logic network 221 2 ⁇ .
  • port M BI such port can be coupled to the B port of the four memory arrays.
  • there are two paths data and control from either a front-end director 18 ⁇ ! -18 ⁇ 32 or a back-end director 200 ⁇ 2OO32 can reach each one of the four memory arrays on the memory board.
  • each one of the directors has a pair of redundant ports, i.e. a 402A port and a 402 B port (FIG. 7).
  • the A port (i.e., port 402A) of one of the directors in the pair is connected to one of the pair of redundant memory ports and the B port (i.e., 402B) of the other one of the directors in such pair is connected to the other one of the pair of redundant memory ports.
  • FIG. 8B an exemplary pair of redundant directors is shown, here, for example, front-end director 180i and front end-director I8O 2 . It is first noted that the directors I8O 1 , 18O2 in each redundant pair of directors must be on different director boards, here boards 190 l5 190 2 , respectively. Thus, here front-end director boards 19 ⁇ !
  • back-end director boards 210 ⁇ -210 8 have thereon: back-end directors 200 1; 2OO 3 , 2OO 5 and 200 7 ; back-end directors 200 2 , 200 , 200 6 and 200 8 ; back-end directors 200 9) 200 ⁇ , 200 13 and 200 15 ; back-end directors 200 10) 200 ⁇ 2 , 200 14 and 200 16 ; back-end directors 200 17; 200 19 , 200 21 and 200 23 ; back-end directors 200 18 , 200 20 , 200 2 2 and 200 2 ; back-end directors 200 25 , 200 27 , 200 29 and 200 31 ; back-end directors 200 18; 200 20 , 200 22 and 200 2 ;
  • FIG. 8A is on front-end director board 190 ⁇ and its redundant front-end director 180 2 , shown in FIG. 8B, is on anther front-end director board, here for example, front-end director board 1 0 2 .
  • the port 402 A of the quad port RAM 402 i.e., the A port referred to above
  • the port 402B of the quad port RAM 402 i.e., the B port referred to above
  • switch 406B of crossbar switch 318 is connected to switch 406B of crossbar switch 318.
  • the ports M 1 -M 4 of switch 406A of director 180 ⁇ are connected to the MAI ports of global cache memory boards 22 ⁇ !
  • each one of the director boards 190 ⁇ 2108 has such a switch 318 and such switch 318 includes a pair of switches 406 A, 406B.
  • Each one of the switches 406 A, 406B is identical in construction, an exemplary one thereof, here switch 406A being shown in detail in FIG. 8C.
  • switch 406A includes four input/output director-side ports D ! -D 4 as described in connection with exemplary director board 190 ⁇ .
  • the exemplary switch 406A includes a plurality of, here four, switch sections 430 ⁇ -430 4 .
  • Each one of the switch sections 430i-430 4 is identical in construction and is coupled between a corresponding one of the input/output director-side ports U ! -D 4 and a corresponding one of the output/input memory-side ports M 1 -M 4 , respectively, as shown.
  • the output/input memory-side ports of switch 406B FIG.
  • switch 406A is responsive to request signals on busses R I- A4 from quad port controller 404 in directors 180 ls I8O 3 , 18O 5 , 18O 7 (FIG. 7)
  • switch 406B is responsive in like manner to request signals on busses RBI-RB4 from controller 404 in directors 180 ls I8O 3 , 18O5 and 180 7 ). More particularly, controller 404 of director I8O 1 produces request signals on busses RAI or R B I.
  • controller 404 of director 18O 3 produces request signals on busses RA2 or R B 2
  • controller 404 of director I8O 5 produces request signals on busses RA 3 or R ⁇ 3
  • controller 404 of director I8O 7 produces request signals on busses R 4 or R ⁇ 4 .
  • switch section 403! is shown in FIG. 8C to include a FIFO 432 fed by the request signal on bus R ⁇ A- (It should be understood that the FIFOs, not shown, in switch sections 430 2 -430 4 are fed by request signals RA 2 - A4, respectively).
  • the switch section 406 ⁇ also includes a request generation 434, and arbiter 436, and selectors 442 and 446, all arranged as shown.
  • the data at the memory-side ports M1-M4 are on busses DM1 -DM4 are fed as inputs to selector 446.
  • selector 446 Also fed to selector 446 is a control signal produced by the request generator on bus 449 in response to the request signal RA !
  • the control signal on bus 449 indicates to the selector 446 the one of the memory- side ports M 1 -M 4 which is to be coupled to director-side port Di.
  • the other switch sections 430 2 - 430 4 operate in like manner with regard to director-side ports Di-D 4 , respectively and the memory-side ports M ⁇ -M 4 .
  • the data portion of the word at port Di (i.e., the word on bus DD1) is also coupled to the other switch sections 430 2 -430 4 .
  • the data portion of the words at ports D 2 -D (i.e., the words on busses DD2-DD4, respectively), are fed to the switch sections 430 ⁇ -430 4 , as indicated. That is, each one of the switch sections 430r430 has the data portion of the words on ports D ⁇ -D 4 (i.e., busses DD1-DD4), as indicated.
  • the data portion of the word at port Mi (i.e., the word on bus DM1) is also coupled to the other switch sections 430 2 -430 4 .
  • each one of the switch sections 430 ⁇ -430 4 has the data portion of the words on ports M 1 -M 4 (i.e., busses DM1 -DM4), as indicated.
  • a request on bus R AI to switch section 430i is a request from the director 180i which identifies the one of the four ports M 1 -M in switch 430 ⁇ is to be coupled to port 402A of director 180i (director side port D ).
  • port 402A of director I8O 1 may be coupled to one of the memory side ports M ⁇ -M selectively in accordance with the data on bus R AI -
  • a request on buses R A2 , R A3 , R A4 to switch section 43 ⁇ 2 -430 , respectively are requests from the directors I8O 3 , 18O 5 , and 180 7 , respectively, which identifies the one of the four ports MrM 4 in switch 430 ⁇ -430 4 is to be coupled to port 402 A of directors I8O3, 18O5 and I8O7, respectively.
  • the requests R AI are stored as they are produced by the quad port RAM controller 440 (FIG. 7) in receive FIFO 432.
  • the request generator 434 receives from FIFO 432 the requests and determines which one of the four memory-side ports M 1 -M 4 is to be coupled to port 402A of director I8O 1 . These requests for memory-side ports M 1 -M 4 are produced on lines RA1,1 - RAI, 4, respectively.
  • line RA1,1 i.e., the request for memory side port Mi
  • switch sections 430 2 -430 4 which are coupled to port 402 A of directors I8O3, 180 5 , and I8O 7
  • line RA2,1, RA3,1 and RA4,1, respectively are also fed to the arbiter 436, as indicated. .
  • the arbiter 436 resolves multiple requests for memory-side port Mi on a first come-first serve basis.
  • the arbiter 436 then produces a control signal on bus 435 indicating the one of the directors 180 ⁇ , I8O 3 , 18O5 or I8O 7 which is to be coupled to memory-
  • the control signal on bus 435 is fed to selector 442. Also fed to selector 442 are the data portion of the data at port Di, i.e., the data on data bus DD1) along with the data portion of the data at ports D 2 -D4, i.e., the data on data busses DD2-DD4, respectively, as indicated.
  • the control signal on bus 435 causes the selector 442 to couple to the output thereof the data busses DD1- DD4 from the one of the directors I8O 1 , 18O 3 , 180 5 , I8O7 being granted access to memory-side port Mi by the arbiter 436.
  • the selected outputs of selector 442 is coupled to memory-side port Mi.
  • the system interface 160 (FIG. 2) is presented to the system interface 160 (FIG. 2) in batches from many host computer processors 121 1 -12l3 2 .
  • the data from the host computer processors 121 1 -12l3 re interleaved with each other as they are presented to a director I8O 1 -18 ⁇ 3 .
  • the batch from each host computer processor I8O 1 -18O 32 (i.e., source) is tagged by the protocol translator 400. More particularly by a Tacheon ASIC in the case of a fibre channel connection.
  • the controller 404 has a look-up table formed during initialization. As the data comes into the protocol translator 400 and is put into the quad port RAM 420 under the control of controller 404, the protocol translator 400 informs the controller that the data is in the quad port RAM 420.
  • the controller 404 looks at the configuration of its look-up table to determine the global cache memory 220 location (e.g., cache memory board 22 ⁇ !-220 8 ) the data is to be stored into. The controller 404 thus produces the request signals on the appropriate bus R AI , R BI , and then tells the quad port RAM 402 that there is a block of data at a particular location in the quad port RAM 402, move it to the particular location in the global cache, memory 220.
  • the global cache memory 220 location e.g., cache memory board 22 ⁇ !-220 8
  • the crossbar switch 318 also takes a look at what other controllers 404 in the directors I8O3, 18O5, and I8O 7 on that particular director board 190 ⁇ are asking by making request signal on busses R A2 , RB 2 , A3 , R B3 , R A4 , R B4 , respectively.
  • the arbitration of multiple requests is handled by the arbiter 436 as described above in connection with FIG. 8C.
  • the exemplary director I8 1 is shown to include in the message engine/CPU controller 314.
  • the message engine/CPU controller 314 is contained in a field programmable gate array (FPGA).
  • the message engine (ME) 315 is coupled to the CPU bus 317 and the DMA section 408 as shown.
  • the message engine (ME) 315 includes a Direct Memory Access (DMA) section 408, a message engine (ME) state machine 410, a transmit buffer 424 and receive buffer 424, a MAC packetizer/depacketizer 428, send and receive pointer registers 420, and a parity generator 321.
  • the DMA section 408 includes a DMA transmitter 418, shown and to be described below in detail in connection with FIG. 9, and a DMA receiver 424, shown and to be described below in detail in connection with FIG.10, each of which is coupled to the CPU bus interface 317, as shown in FIG. 7.
  • the message engine (ME) 315 includes a transmit data buffer 422 coupled to the DMA transmitter 418, a receive data buffer 424 coupled to the DMA receiver 421 , registers 420 coupled to the CPU bus 317 through an address decoder 401 , the packetizer/de-packetizer 428, described above, coupled to the transmit data buffer 422, the receive data buffer 424 and the crossbar switch 320, as shown, and a parity generator 321 coupled between the transmit data buffer 422 and the crossbar switch 320. More particularly, the packetizer portion 428P is used to packetize the message payload into a MAC packet (FIG.
  • the message engine (ME) 315 also includes: a receive write pointer 450, a receive read pointer 452; a send write pointer 454, and a send read pointer 456.
  • FIGS. 11 and 12 the transmission of a message from a director I8O 1 -I8O3 2 , 2OO 1 -2OO32 and the reception of a message by a director 2IO 1 -2IO3 2 , here exemplary director ISOi shown in FIG. 7) will be described.
  • a director I8O 1 -I8O3 2 , 2OO 1 -2OO32 and the reception of a message by a director 2IO 1 -2IO3 2 here exemplary director ISOi shown in FIG. 7
  • the controller 306 FIG. 5
  • the controller 306 the controller 306 (FIG. 5) of both message network boards 304 l5 304 2 initialize the message routing mapping described above for the switches 308!-308 4 in switch section 308 and for the crossbar switches 320.
  • a request is made by the host computer 120.
  • the request is sent to the protocol translator 400.
  • the protocol translator 400 sends the request to the microprocessor 299 via CPU bus 317 and buffer 301.
  • the CPU 310 (FIG. 7) in the microprocessor 299 of exemplary director I8O1 determines that a message is to be sent to another one of the directors I8O2-I8O32, 200 20032, (e.g., the CPU 310 determines that there has been a "miss" in the global cache memory 220 (FIG. 2) and wants to send a message to the appropriate one of the back-end directors 20 ⁇ !-20 ⁇ 3 2 , as described above in connection with FIG. 2), the CPU 310 builds a 64 byte descriptor (FIG.
  • Step 500 which includes a 32 byte message payload indicating the addresses of the batch of data to be read from the bank of disk drives 140 (FIG. 2) (Step 500) and a 32 byte command field (Step 510) which indicates the message destination via an 8-byte bit vector, i.e., the director, or directors, which are to receive the message.
  • An 8-byte portion of the command field indicates the director or directors, which are to receive the message. That is, each one of the 64 bits in the 8-byte portion corresponds to one of the 64 directors.
  • a logic 1 in a bit indicates that the corresponding director is to receive a message and a logic 0 indicates that such corresponding director is not to receive the message.
  • the resulting 64-byte descriptor is generated by the CPU 310 (FIG. 7) (Step 512) is written into the RAM 312 (Step 514), as shown in FIG. 11. More particularly, the RAM 512 includes a pair of queues; a send queue and a receive queue, as shown in FIG. 7.
  • the RAM 312 is coupled to the CPU bus 317 through an Error Detection and Correction (EDAC)/Memory control section 303, as shown.
  • EDAC Error Detection and Correction
  • the CPU 310 then indicates to the message engine (ME) 315 state machine 410 (FIG. 7) that a descriptor has been written into the RAM 312.
  • the message engine (ME) 315 also includes: a receive write pointer or counter 450, the receive read pointer or counter 452, the send write pointer or counter 454, and the send read pointer or counter 454, shown in FIG. 7. All four pointers 450, 452, 454 and 456 are reset to zero on power-up.
  • the message engine/CPU controller 314 also includes: the de-packetizer portion 428D of packetizer/de-packetizer 428, coupled to the receive data buffer 424 (FIG.
  • the CPU 310 when the CPU 310 indicates that a descriptor has been written into the RAM 312 and is now ready to be sent, the CPU 310 increments the send write pointer and sends it to the send write pointer register 454 via the register decoder 401.
  • the contents of the send write pointer register 454 indicates the number of messages in the send queue 312S of RAM 312, which have not been sent.
  • the state machine 410 checks the send write pointer register 454 and the send read pointer register 456, Step 518.
  • both the send write pointer register 454 and the send read pointer register 456 are initially reset to zero during power-up. Thus, if the send read pointer register 456 and the send write pointer register 454 are different, the state machine knows that there is a message is in RAM 312 and that such message is ready for transmission. If a message is to be sent, the state machine 410 initiates a transfer of the stored 64-byte descriptor to the message engine (ME) 315 via the DMA transmitter 418, FIG. 7 (Steps 520, 522). The descriptor is sent from the send queues 312S in RAM 312 until the send read pointer 456 is equal to the send write pointer 454.
  • ME message engine
  • the CPU 310 generates a destination vector indicating the director, or directors, which are to receive the message.
  • the command field is 32-bytes, eight bytes thereof having a bit representing a corresponding one of the 64 directors to receive the message.
  • each of the bit positions 1-64 represents directors 180 ⁇ 1803 2 , 200 1 -20 ⁇ 3 1 , respectively.
  • the eight-byte vector indicates that the destination director is only front-end director 108 ⁇
  • the eight-byte vector indicates that the destination director is only front-end director IO8 2 .
  • the destination for the message is to more than one director, i.e., a multi-cast message.
  • a logic 1 is only in bit positions 2, 3, 63 and 64.
  • the eight-byte vector indicates that the destination directors are only front-end director IO8 2 and I8O3 and back- end directors 2003! and 2OO32.
  • Step 524, 525, for a uni-cast transmission If the message engine (ME) 315 state machine 410 indicates that the director is available by examining the transmit vector mask (FIG. 1 IF) stored in register 420, the message engine (ME) 315 encapsulates the message payload with a MAC header and CRC inside the packetizer portion 428P, discussed above (Step 526).
  • An example of the mask is shown in FIG. 1 IF.
  • the mask has 64 bit positions, one for each one of the directors. Thus, as with the destination vectors described above in connection with FIGS. 1 lC-1 IE, bit positions 1- 64 represents directors 180 1 -18032, 200i-20 ⁇ 32, respectively.
  • a logic 1 in a bit position in the mask indicates that the representative director is available and a logic 0 in such bit position indicates that the representative director is not available.
  • FIG. 1 IF only director 2OO32 is unavailable.
  • the destination vector after passing through the mask of FIG. 1 IF modifies the destination vector to that shown in FIG. 11G.
  • director 2OO 32 will not receive the message.
  • Such mask modification to the destination vector is important because, as will be described, the messages on a multi-cast are sent sequentially and not in parallel. Thus, elimination of message transmission to an unavailable director or directors increases the message transmission efficiency of the system.
  • the message engine (ME) 315 transfers the MAC packet to the crossbar switch 320 (Step 528) and the MAC packet is routed to the destination by the message network 260 (Step 530) via message network boards 304 l5 304 2 or on the same director board via the crossbar switch 320 on such director board.
  • the message read operation is described.
  • the director waits for a message.
  • the message engine (ME) 315 state machine 410 receives the packet (Step 602).
  • the state machine 410 checks the receive bit vector mask (FIG.
  • Step 604 If the state machine 410 determines that the message is from an improper source (i.e., a faulty director as indicated in the mask, FIG. 1 IF, for example), the packet is discarded (Step 606). On the other hand, if the state machine 410 determines that the packet is from a proper or valid director (i.e., source), the message engine (ME) 315 de-encapsulates the message from the packet (Step 608) in de-packetizer 428D. The state machine 410 in the message engine (ME) 315 initiates a 32- byte payload transfer via the DMA receive operation (Step 610).
  • an improper source i.e., a faulty director as indicated in the mask, FIG. 1 IF, for example
  • the packet is discarded (Step 606).
  • the message engine (ME) 315 de-encapsulates the message from the packet (Step 608) in de-packetizer 428D.
  • the DMA writes the 32 byte message to the memory receive queue 312R in the RAM 312 (Step 612).
  • the message engine (ME) 315 state machine 410 increments the receive write pointer register 450 (Step 614).
  • the CPU 310 checks whether the receive write pointer 450 is equal to the receive read pointer 452 (Step 616). If they are equal, such condition indicates to the CPU 310 that a message has not been received (Step 618).
  • Step 700 the receive DMA engine 420 successfully completes a message transfer to the receive queue in RAM 312 (FIG. 7).
  • the state machine 410 in the message engine (ME) 315 generates an acknowledgement MAC packet and transmits the MAC packet to the sending director via the message network 260 (FIG. 2) (Steps 702, 704).
  • the message engine (ME) 315 at the sending director de-encapsulates a 16 byte status payload in the acknowledgement MAC packet and transfers such status payload via a receive DMA operation (Step 706).
  • the DMA of the sending (i.e., source) director writes to a status field of the descriptor within the RAM memory send queue 314S (Step 708).
  • the state machine 410 of the message engine (ME) 315 of the sending director increments its send read pointer 454 (Step 712).
  • the CPU 310 of the sending director (which received the acknowledgement message) processes the descriptor status and removes the descriptor from the send queue 312S of RAM 312 (Step 714). It should be noted that the send and receive queues 312S and 312R are each circular queues.
  • the MAC packets are, during normal operation, transmitted alternatively to one of the pair of message network boards 304 l5 304 2 by hardware a selector S in the crossbar switch 320.
  • the selector S is responsive to the bit B in the header of the MAC packet (FIG. 2B) and, when such bit B is one logic state the data is coupled to one of the message networks boards 402A and in response to the opposite logic state the data is coupled to the other one of the message networks boards 402B. That is, when one message is transmitted to board 304i the next message is transmitted to board 304 2 .
  • FIG. 9 the details of an exemplary transmit DMA 418 is shown.
  • a descriptor has been created by the CPU 310 (FIG. 7) and is then stored in the RAM 312. If the send write pointer 450 (FIG. 7) and send read pointer 452, described above, have different counts an indication is provided by the state machine 410 in the message engine (ME) 315 (FIG. 7) that the created descriptor is available for DMA transmission to the message engine (ME) 315, the payload off the descriptor is packetized into a MAC packet and sent through the message network 360 (FIG. 2) to one or more directors I8O1-I8O32, 20 ⁇ !-20 ⁇ 32.
  • the descriptor created by the CPU 310 is first stored in the local cache memory 319 and is later transferred to the send queue 312S in RAM 312.
  • the message engine (ME) 315 state machine 410 initiates a DMA transmission as discussed above in connection with Step 520 (FIG. 11).
  • the descriptor resides in send queues 312R within the RAM 312.
  • each descriptor which contains the message is a fixed size, here 64- bytes. As each new, non-transmitted descriptor is created by the CPU 310, it is sequentially stored in a sequential location, or address in the send queue 312S.
  • the address is a 32-bit address.
  • the state machine 410 in the message engine (ME) 315 (FIG. 7), sends the queue address on bus 411 to an address register 413 in the DMA transmitter 418 (FIG. 9) along with a transmit write enable signal Tx_WE signal.
  • the DMA transmitter 418 requests the CPU bus 317 by asserting a signal on Xmit_Br.
  • the CPU bus arbiter 414 (FIG. 7) performs a bus arbitration and when appropriate the arbiter 414 grants the DMA transmitter 418 access to the CPU bus 317.
  • the Xmit Cpu state machine 419 then places the address currently available in the address register 413 on the Address bus portion 317 A of CPU bus 317 by loading the output address register 403. Odd parity is generated by a Parity generator 405 before loading the output address register 403.
  • the address in register 403 is placed on the CPU bus 317 (FIG. 7) for RAM 312 send queue 312S, along with appropriate read control signals via CPU bus 317 portion 317C.
  • the data at the address from the RAM 312 passes, via the data bus portion 317D of CPU bus 317, through a parity checker 415 to a data input register 417.
  • the control signals from the CPU 10 are fed to a Xmit CPU state machine 419 via CPU bus 317 bus portion 317C.
  • One of the control signals indicates whether the most recent copy of the requested descriptor is in the send queue 312S of the RAM 312 or still resident in the local cache memory 319. That is, the most recent descriptor at any given address is first formed by the CPU 310 in the local cache memory 319 and is later transferred by the CPU 310 to the queue in the RAM 312. Thus, there may be two descriptors with the same address; one in the RAM 312 and one in the local cache memory 319 (FIG. 7), the most recent one being in the local cache memory 319.
  • the transmit DMA 418 must obtain the descriptor for DMA transmission from the RAM 312 and this descriptor is stored in the transmit buffer register 421 using signal 402 produced by the state machine 419 to load these registers 421.
  • the control signal from the CPU 310 to the Xmit CPU state machine 419 indicates whether the most recent descriptor is in the local cache memory 319. If the most recent descriptor is in the local cache memory 319, the Xmit CPU state machine 419 inhibits the data that was just read from send queue 312S in the RAM 312 and which has been stored in register 421 from passing to selector 423. In such case, state machine 419 must perform another data transfer at the same address location.
  • the most recent message is then transferred by the CPU 310 from the local cache memory 319 to the send queue 312S in the RAM 312.
  • the transmit message state machine 419 then re-arbitrates for the CPU bus 317 and after it is granted such CPU bus 317, the Xmit CPU state machine 419 then reads the descriptor from the RAM 312. This time, however, there the most recent descriptor is available in the send queue 312s in the RAM 312.
  • the descriptor in the RAM 312 is now loaded into the transmit buffer register 421 in response to the assertion of the signal 402 by the Xmit CPU state machine 419.
  • the descriptor in the register 421 is then transferred through selector 423 to message bus interface 409 under the control of a Xmit message (msg) state machine 427. That is, the descriptor in the transmit buffer register 421 is transferred to the transmit data buffer 422 (FIG. 7) over the 32 bit transmit message bus interface 409 by the Xmit message (msg) state machine 427.
  • the data in the transmit data buffer 422 (FIG. 7) is packetized by the packetizer section of the packetizer/de-packetizer 428 as described in Step 530 in FIG. 11. More particularly, and referring also to FIG. 14 A, the method of operating the transmit DMA 418 (FIG. 9) is shown. As noted above, each descriptor is 64-byte.
  • the transfer of the descriptor takes place over two interfaces namely, the CPU bus 317 and the transmit message interface bus 409 (FIG. 7).
  • the CPU bus 317 is 64 bits wide and eight, 64-bit double-words constitute a 64-byte descriptor.
  • the Xmit CPU state machine 419 generates the control signals which result in the transfer of the descriptor from the RAM 312 into the transmit buffer register 421 (FIG. 7).
  • the 64-byte descriptor is transferred in two 32-byte burst accesses on the CPU bus 317. Each one of the eight double words is stored sequentially in the transmit buffer register 421 (FIG. 9).
  • Step 800 the message engine 315 state machine 410 loads the transmit DMA address register 413 with the address of the descriptor to be transmitted in the send queue 312S in RAM 312. This is done by the asserting the Tx_WE signal and this puts Xmit CPU state machine 419 in step 800, loads the address register 413 and proceeds to step 802.
  • step 802 The Xmit Cpu state machine 419 loads the CPU transfer counter 431 (FIG. 9) with a 32-byte count, which is 2. This is the number of 32 byte transfers that would be required to transfer the 64-byte descriptor, Step 802.
  • the Xmit Cpu state machine 419 now proceeds to Step 804.
  • the transmit DMA state machine 419 checks the validity of the address that is loaded into its address register 413.
  • the address loaded into the address register 413 is checked against the values loaded into the memory address registers 435.
  • the memory address registers 435 contain the base address and the offset of the send queue 312s in the RAM 312. The sum of the base address and the offset is the range of addresses for the send queue 312S in RAM 312.
  • the address check circuitry 437 constantly checks whether the address in the address register 413 is with in the range of the send queue 312S in the RAM 312. If the address is found to be outside the range of the send queue 312S the transfer is aborted, this status is stored in the status register 404 and then passed back to the message engine 315 state machine 410 in Step 416.
  • Step 805. If the address is within the range, i.e., valid, the transmit DMA state machine 419 proceeds with the transfer and proceeds to Step 806.
  • the transmit DMA state machine 419 requests the CPU bus 317 by asserting the Xmit_BR signal to the arbiter 414 and then proceeds to Step 807.
  • Step 807 the Xmit Cpu state machine 419 constantly checks if it has been granted the bus by the arbiter. When the CPU bus 317 is granted, the Xmit CPU state machine proceeds to Step 808.
  • Step 808 the Xmit Cpu state machine 419 generates an address and a data cycle which essentially reads 32-bytes of the descriptor from the send queue 312S in the RAM 312 into its transmit buffer register 421.
  • the Xmit Cpu state machine 419 now proceeds to step 810.
  • Step 810 the Xmit Cpu state machine 419 loads the descriptor that was read into its buffer registers 421 and proceeds to Step 811.
  • Step 811 a check is made for any local cache memory 319 coherency errors (i.e., checks whether the most recent data is in the cache memory 319 and not in the RAM 312) on these 32-bytes of data.
  • the Xmit Cpu state machine 419 discards this data and proceeds to Step 806.
  • the Xmit Cpu state machine 419 now requests for the CPU bus 317 again and when granted, transfers another 32-bytes of data into the transmit buffer register 421, by which time the CPU has already transferred the latest copy of the descriptor into the RAM 312. In cases when the 32-bytes of the descriptor initially fetched from the RAM 312 was not resident in the local CPU cache memory 319 (i.e., if no cache coherency errors were detected), the Xmit Cpu state machine 419 proceeds to Step 812.
  • Step 812 the Xmit CPU state machine 419 decrements counters 431 and increments the address register 413 so that such address register 413 points to the next address.
  • the Xmit Cpu state machine then proceeds to step 814.
  • the Transmit CPU state machine 419 checks to see if the transfer counter 431 has expired, i.e., counted to zero, if the count was found to be non-zero, it then, proceeds to Step 804 to start the transfer of the next 32-bytes of the descriptor. In case the counter 431 is zero, the process goes to Step 816 to complete the transfer.
  • the message interface 409 is 32 bits wide and sixteen, 32 bit words constitute a 64-byte descriptor.
  • the 64-byte descriptor is transferred in batches of 32 bytes each.
  • the Xmit msg state machine 427 controls and manages the interface 409.
  • the Xmit Cpu state machine asserts the signal 433 to indicate that the first 32 bytes have been successfully transferred over the CPU bus 317 (Step 818, FIG. 14B), this puts the Xmit msg state machine into Step 818 and starts the transfer on the message interface.
  • step 820 the Xmit msg machine 427 resets burst/transfer counters 439 and initiates the transfer over the message interface 409.
  • the transfer is initiated over the message interface 409 by asserting the "transfer valid" (TX_DATA_Vaild) signal indicating to the message engine 315 state machine 410 that valid data is available on the bus 409.
  • the transmit msg machine 427 transfers 32 bits of data on every subsequent clock until its burst counter in burst/transfer counter 439 reaches a value equal to eight, Step 822.
  • the burst counter in burst/transfer counter 439 is incremented with each 32-bit word put on the message bus 409 by a signal on line 433.
  • Step 824 a check is made by the state machine 427 as to whether the transmit counter 431 has expired, i.e., is zero, Step 824.
  • the expiry of the transfer counter in burst/transfer counter 439 indicates the 64 byte descriptor has been transferred to the transmit buffer 422 in message engine 315. If it has expired, the transmit message state machine 427 proceeds to Step 826.
  • the Xmit msg state machine asserts the output End of Transfer (Tx_EOT) indicating the end of transfer over the message bus 409 .In this state, after the assertion of the Tx_EOT signal the status of the transfer captured in the status register 404 is sent to the message engine 315 state machine 410.
  • the DMA operation is complete with the descriptor being stored in the transmit buffer 422 (FIG. 7).
  • the transfer counter in burst/transfer counter 439 has not expired, the process goes to Step 800 and repeats the above described procedure to transfer the 2 nd 32 bytes of descriptor data, at which time the transfer will be complete.
  • the receive DMA 420 is shown.
  • a message received from another director is to be written into the RAM 312 (FIG. 7).
  • the receive DMA 420 is adapted to handle three types of information: error information which is 8 bytes in size; acknowledgement information which is 16 bytes in size; and receive message payload and/or fabric management information which is 32 byes in size.
  • the message engine 315 state machine 410 asserts the Rx_WE signal, indicating to the Receive DMA 420 that it is ready transfer the Data in its Rec buffer 426 FIG. 7.
  • the data in the Receive buffer could be the 8-byte error information, the 16-byte Acknowledgment information or the 32-byte Fabric management/ Receive message payload information.
  • the Receive message machine 450 (FIG. 10) loads the address into the address register 452 and the transfer count indicating the type of information, into the receive transfer counter 454.
  • the address loaded into the address register 452 is checked by the address check circuitry 456 to see if it is with in the range of the Receive queue addresses, in the RAM 312. This is done by checking the address against the values loaded into the memory registers 457 (i.e., a base address register and an offset register therein).
  • the base address register contains the start address of the receive queue 312R residing in the RAM 312 and the offset register contains the size of this receive queue 312R in RAM 312. Therefore the additive sum of, the values stored in the base address register and the offset register specifies the range of addresses of the receive queue in the RAM 312R.
  • the memory registers 457 are loaded during initialization.
  • the message engine 315 state machine 410 the proceeds to place the data on a 32-bit message engine 315 data bus 407, FIG. 10.
  • a Rx_data_valid signal accompanies each 32 bits of data, indicating that the data on the message engine data bus 407 is valid.
  • the receive message state machine 450 loads the data on the data bus into the receive buffer register 460.
  • the end of the transfer over the message engine data bus 407d is indicated by the assertion of the Rx_EOT signal at which time the Receive message state machine 450 loads the last 32 bits of data on the message engine data bus 407D of bus 407, into the receive buffer registers 460.
  • the Receive CPU machine 462 now, requests for the CPU bus 317 by asserting the signal REC_Br. After an arbitration by CPU bus arbiter 414 (FIG.
  • the receive DMA 420 (FIG. 10) is given access to the CPU bus 317.
  • the Receive CPU state machine 462 proceeds to transfer the data in its buffer registers 424 over the CPU bus 317 into the Receive queue 312R in the RAM 312. Simultaneously, this data is also transferred into a duplicate buffer register 466.
  • the data at the output of the receive buffer register 460 passes to one input of a selector 470 and also passes to a duplicate data receive buffer register 460.
  • the output of the duplicate receive buffer register 466 is fed to a second input of the selector 470. As the data is being transferred by the Receive CPU state machine 462, it is also checked for cache coherency errors.
  • the receive DMA machine 420 waits for the CPU 310 to copy the old data in its local cache memory 319 back to the receive queue 312R in the RAM 312 and then overwrites this old data with a copy of the new data from the duplicate buffer register 466.
  • the receive CPU state machine 462 produces a select signal on line 463 which couples the data in the duplicate buffer register 466 to the output of selector 470 and then to the bus 317 for store in the random access memory 312.
  • the successful write into the RAM 312 completes the DMA transfer.
  • the receive DMA 420 then signals the message engine 315 state machine 410 on the status of the transfer. The status of the transfer is captured in the status register 459.
  • Step 830 the Receive message machine 450 checks if the write enable signal Rx_WE is asserted. If found asserted, the receive DMA 420 proceeds to load the address register 452 and the transfer counter 454. The value loaded into the transfer counter 454 determines the type of DMA transfer requested by the Message engine state machme 310 in FIG 7. The assertion of the Rx_WE signal starts the DMA receive transfer operation. This puts the Rx msg state machine 450 in Step 832. In Step 832 the Rec msg state machine 450 loads the address register 452, the transfer counter 454 and then proceeds to Step 834.
  • Step 834 it checks to see if the Rx_DATA_VALID signal is asserted. If asserted it proceeds to step 836.
  • the Rx msg state machine loads the buffer register 460 (FIG. 10) in Step 836 with the data on the message engine data bus 407D of bus 407 fig 10.
  • the Rx_DATA_VALID signal accompanies each piece of data put on the bus 407.
  • the data is sequentially loaded into the buffer registers 460 (FIG.10).
  • the End of the transfer on the message engine data bus 407D of bus 407 is indicated by the assertion of the Rx_EOT signal.
  • the Receive message state machine 450 When the Receive message state machine 450 is in the End of transfer state Step 840 it signals the Receive CPU state machine 462 and this starts the transfer on the CPU bus 317 side.
  • Step 844 the End of the transfer on the Message engine data bus 407D portion of bus 407 starts the Receive CPU state machine 462 and puts it in Step 842.
  • the Receive CPU state machine 462 checks for validity of the address in this state (Step 844). This is done by the address check circuitry 456. If the address loaded in the address register 452 is outside the range of the receive queue 312R in the RAM 312, the transfer is aborted and the status is captured in the Receive status register 459 and the Rec Cpu state machine 462 proceeds to Step 845. On a valid address the Receive CPU state machine 462 goes to Step 846. In Step 846 the Receive Cpu state machine 462 requests for access of the CPU bus 317.
  • Step 848 it checks for a grant on the bus 317. On a qualified grant it proceeds to step 850.
  • Step 850 The Rec Cpu state machine 462 performs an address and a data cycle, which essentially writes the data in the buffer registers 460 into the receive queue 312R in RAM 312. Simultaneously with the write to the RAM 312, the data put on the CPU bus 317 is also loaded into the duplicate buffer register 466. At same time, the CPU 310 also indicates on one of the control lines, if the data corresponding to the address written to in the RAM 312 is available in its local cache memory 319. At the end of the address and data cycle the Rec Cpu state machine 462 proceeds to Step 850.
  • Step 846 the Receive CPU state machine 462 now generates another address and data cycle to the previous address and this time the data from the duplicate buffer 466 is put on to the CPU data bus 317. If there were no cache coherency errors the Receive CPU state machine 462 proceeds to Step 852 where it decrements the transfer counter 454 and increment the address in the address register 452. The Receive Cpu state machine 462 then proceeds to Step 854. In Step 854, the state machine 462 checks if the transfer counter has expired, i.e., is zero.
  • Step 844 On a non zero transfer count the receive Cpu state machine 462 proceeds to Step 844 and repeats the above described procedure until the transfer becomes zero.
  • a zero transfer count when in step 854 completes the write into the receive queue 312R in RAM 312 and the Rec Cpu state machine proceeds to 845.
  • step 845 it conveys status stored in the status register back to status is conveyed to the message engine 315 state machine 410.
  • the interrupt control status register 412 will be described in more detail.
  • a packet is sent by the pocketsize portion of the packetizer/de- packetizer 428 to the crossbar switch 320 for transmission to one or more of the directors.
  • the packet sent by the packetizer portion of the packetizer/de-packetizer 428 passes through a parity generator PG in the message engine 315 prior to passing to the crossbar switch 320.
  • a parity bit is added to the packet by parity bit generator PG prior to passing to the crossbar switch 320.
  • the parity of the packet is checked in the parity checker portion of a parity checker/generator (PG/C) in the crossbar switch 320.
  • the result of the check is sent by the PG/C in the crossbar switch 320 to the interrupt control status register 412 in the director 180 ⁇ .
  • a packet when a packet is transmitted from the crossbar switch 320 to the message engine 315 of exemplary director 180 ⁇ , the packet passes through a parity generator portion of the parity checker/generator (PG/C) in the crossbar switch 320 prior to being transmitted to the message engine 315 in director 180 ⁇ .
  • the parity of the packet is then checked in the parity checker portion of the parity checker (PC) in director 180 ⁇ and is the result (i.e., status) is transmitted to the status register 412.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

A system interface (160) includes a plurality of first directors (180), a plurality of second directors (200), a data transfer section (240) and a message network (260) which operates independently of the data transfer section. The data transfer section includes a cache memory (220) coupled to the plurality of first and second directors as is the message network. The first and second directors control data transfer between the first directors and the second directors in response to messages passing between the first directors and the second directors through the messaging network to facilitate data transfer between first directors and the second directors. The data passes through the cache memory in the data transfer section. The system interface operates to transfer data between a host computer/server (120) and a bank of disk drives (140).

Description

DATA STORAGE SYSTEM HAVING POINT-TO-POINT CONFIGURATION
BACKGROUND OF THE INVENTION
This invention relates generally to data storage systems, and more particularly to data storage systems having redundancy arrangements to protect against total system failure in the event of a failure in a component or subassembly of the storage system.
As is known in the art, large host computers and servers (collectively referred to herein as "host computer/servers") require large capacity data storage systems. These large computer/servers generally includes data processors, which perform many operations on data introduced to the host computer/server through peripherals including the data storage system. The results of these operations are output to peripherals, including the storage system.
One type of data storage system is a magnetic disk storage system. Here a bank of disk drives and the host computer/server are coupled together through an interface. The interface includes "front end" or host computer/server controllers (or directors) and "back-end" or disk controllers (or directors). The interface operates the controllers (or directors) in such a way that they are transparent to the host computer/server. That is, data is stored in, and retrieved from, the bank of disk drives in such a way that the host computer/server merely thinks it is operating with its own local disk drive. One such system is described in U.S. Patent 5,206,939, entitled "System and
Method for Disk Mapping and Data Retrieval", inventors Moshe Yanai, Natan Vishlitzky, Bruno
Alterescu and Daniel Castel, issued April 27, 1993, and assigned to the same assignee as the present invention.
As described in such U.S. Patent, the interface may also include, in addition to the host computer/server controllers (or directors) and disk controllers (or directors), addressable cache memories. The cache memory is a semiconductor memory and is provided to rapidly store data from the host computer/server before storage in the disk drives, and, on the other hand, store data from the disk drives prior to being sent to the host computer/server. The cache memory being a semiconductor memory, as distinguished from a magnetic memory as in the case of the disk drives, is much faster than the disk drives in reading and writing data.
The host computer/server controllers, disk controllers and cache memory are interconnected through a backplane printed circuit board. More particularly, disk controllers are mounted on disk controller printed circuit boards. The host computer/server controllers are mounted on host computer/server controller printed circuit boards. And, cache memories are mounted on cache memory printed circuit boards. The disk directors, host computer/server directors, and cache memory printed circuit boards plug into the backplane printed circuit board, h order to provide data integrity in case of a failure in a director, the backplane printed circuit board has a pair of buses. One set the disk directors is connected to one bus and another set of the disk directors is connected to the other bus. Likewise, one set the host computer/server directors is connected to one bus and another set of the host computer/server directors is directors connected to the other bus. The cache memories are connected to both buses. Each one of the buses provides data, address and control information.
The arrangement is shown schematically in FIG. 1. Thus, the use of two buses Bl, B2 provides a degree of redundancy to protect against a total system failure in the event that the controllers or disk drives connected to one bus, fail. Further, the use of two buses increases the data transfer bandwidth of the system compared to a system having a single bus. Thus, in operation, when the host computer/server 12 wishes to store data, the host computer 12 issues a write request to one of the front-end directors 14 (i.e., host computer/server directors) to perform a write command. One of the front-end directors 14 replies to the request and asks the host computer 12 for the data. After the request has passed to the requesting one of the front-end directors 14, the director 14 determines the size of the data and reserves space in the cache memory 18 to store the request. The front-end director 14 then produces control signals on one of the address memory busses Bl, B2 connected to such front-end director 14 to enable the transfer to the cache memory 18. The host computer/server 12 then transfers the data to the front-end director 14. The front- end director 14 then advises the host computer/server 12 that the transfer is complete. The front- end director 14 looks up in a Table, not shown, stored in the cache memoiy 18 to determine which one of the back-end directors 20 (i.e., disk directors) is to handle this request. The Table maps the host computer/server 12 addresses into an address in the bank 14 of disk drives. The front-end director 14 then puts a notification in a "mail box" (not shown and stored in the cache memory 18) for the back-end director 20, which is to handle the request, the amount of the data and the disk address for the data. Other back-end directors 20 poll the cache memory 18 when they are idle to check their "mail boxes". If the polled "mail box" indicates a transfer is to be made, the back-end director 20 processes the request, addresses the disk drive in the bank 22, reads the data from the cache memory 18 and writes it into the addresses of a disk drive in the bank 22.
When data is to be read from a disk drive in bank 22 to the host computer/server 12 the system operates in a reciprocal manner. More particularly, during a read operation, a read request is instituted by the host computer/server 12 for data at specified memory locations (i.e., a requested data block). One of the front-end directors 14 receives the read request and examines the cache memory 18 to determine whether the requested data block is stored in the cache memory 18. If the requested data block is in the cache memory 18, the requested data block is read from the cache memory 18 and is sent to the host computer/server 12. If the front-end director 14 determines that the requested data block is not in the cache memory 18 (i.e., a so- called "cache miss") and the director 14 writes a note in the cache memory 18 (i.e., the "mail box") that it needs to receive the requested data block. The back-end directors 20 poll the cache memory 18 to determine whether there is an action to be taken (i.e., a read operation of the requested block of data). The one of the back-end directors 20 which poll the cache memory 18 mail box and detects a read operation reads the requested data block and initiates storage of such requested data block stored in the cache memory 18. When the storage is completely written into the cache memory 18, a read complete indication is placed in the "mail box" in the cache memory 18. It is to be noted that the front-end directors 14 are polling the cache memory 18 for read complete indications. When one of the polling front-end directors 14 detects a read complete indication, such front-end director 14 completes the transfer of the requested data which is now stored in the cache memory 18 to the host computer/server 12.
The use of mailboxes and polling requires time to transfer data between the host computer/server 12 and the bank 22 of disk drives thus reducing the operating bandwidth of the interface.
SUMMARY OF THE INVENTION
In accordance with the present invention, a system interface is provided. Such interface includes a plurality of first directors, a plurality of second directors, a data transfer section and a message network. The data transfer section includes a cache memory. The cache memory is coupled to the plurality of first and second directors. The messaging network operates independently of the data transfer section and such network is coupled to the plurality of first directors and the plurality of second directors. The first and second directors control data transfer between the first directors and the second directors in response to messages passing between the first directors and the second directors through the messaging network to facilitate data transfer between first directors and the second directors. The data passes through the cache memory in the data transfer section. With such an arrangement, the cache memory in the data transfer section is not burdened with the task of transferring the director messaging but rather a messaging network is provided, operative independent of the data transfer section, for such messaging thereby increasing the operating bandwidth of the system interface.
In one embodiment of the invention, the system interface each one of the first directors includes a data pipe coupled between an input of such one of the first directors and the cache memory and a controller for transferring the messages between the message network and such one of the first directors.
In one embodiment each one of the second directors includes a data pipe coupled between an input of such one of the second directors and the cache memory and a controller for transferring the messages between the message network and such one of the second directors. In one embodiment the directors includes: a data pipe coupled between an input of such one of the first directors and the cache memory; a microprocessor; and a controller coupled to the microprocessor and the data pipe for controlling the transfer of the messages between the message network and such one of the first directors and for controlling the data between the input of such one of the first directors and the cache memory. In accordance with another feature of the invention, a data storage system is provided for transferring data between a host computer/server and a bank of disk drives through a system interface. The system interface includes a plurality of first directors coupled to host computer/server, a plurality of second directors coupled to the bank of disk drives, a data transfer section, and a message network. The data transfer section includes a cache memory. The cache memory is coupled to the plurality of first and second directors. The message network is operative independently of the data transfer section and such network is coupled to the plurality of first directors and the plurality of second directors. The first and second directors control data transfer between the host computer and the bank of disk drives in response to messages passing between the first directors and the second directors through the messaging network to facilitate the data transfer between host computer/server and the bank of disk drives with such data passing through the cache memory in the data transfer section.
In accordance with yet another embodiment, a method is provided for operating a data storage system adapted to transfer data between a host computer/server and a bank of disk drives. The method includes transferring messages through a messaging network with the data being transferred between the host computer/server and the bank of disk drives through a cache memory, such message network being independent of the cache memory. In accordance with another embodiment, a method is provided for operating a data storage system adapted to transfer data between a host computer/server and a bank of disk drives through a system interface. The interface includes a plurality of first directors coupled to host computer/server, a plurality of second directors coupled to the bank of disk drives; and a data transfer section having a cache memory, such cache memory being coupled to the plurality of first and second directors. The method comprises transferring the data between the host computer/server and the bank of disk drives under control of the first and second directors in response to messages passing between the first directors and the second directors through a messaging network to facilitate the data transfer between host computer/server and the bank of disk drives with such data passing through the cache memory in the data transfer section, such message network being independent of the cache memory. BRIEF DESCRIPTION OF THE DRAWINGS
These and other features of the invention will become more readily apparent from the following detailed description when read together with the accompanying drawings, in which:
FIG. 1 is a block diagram of a data storage system according to the PRIOR ART;
FIG. 2 is a block diagram of a data storage system according to the invention; FIG. 2 A shows the fields of a descriptor used in the system interface of the data storage system of FIG. 2;
FIG. 2B shows the filed used in a MAC packet used in the system interface of the data storage system of FIG, 2;
FIG. 3 is a sketch of an electrical cabinet storing a system interface used in the data storage system of FIG. 2;
FIG. 4 is a diagramatical, isometric sketch showing printed circuit boards providing the system interface of the data storage system of FIG. 2;
FIG. 5 is a block diagram of the system interface used in the data storage system of FIG.
2; FIG. 6 is a block diagram showing the connections between front-end and back-end directors to one of a pair of message network boards used in the system interface of the data storage system of FIG. 2;
FIG. 7 is a block diagram of an exemplary one of the director boards used in the system interface of he data storage system of FIG. 2; FIG. 8 is a block diagram of the system interface used in the data storage system of FIG.
2;
FIG. 8A is a diagram of an exemplary global cache memory board used in the system interface of FIG. 8; FIG. 8B is a diagram showing a pair of director boards coupled between a pair of host processors and global cache memory boards used in the system interface of FIG. 8;
FIG. 8C is a block diagram of an exemplary crossbar switch used in the front-end and rear-end directors of the system interface of FIG. 8; FIG. 9 is a block diagram of a transmit Direct Memory Access (DMA) used in the system interface of the FIG. 8;
FIG. 10 is a block diagram of a receive DMA used in the system interface of FIG.8; FIG. 11 shows the relationship between FIGS. 11 A and 1 IB, such FIGS. 11 A and 1 IB together showing a process flow diagram of the send operation of a message network used in the system interface of FIG. 8;
FIGS. 1 lC-1 IE are examples of digital words used by the message network in the system interface of FIG. 8;
FIG. 1 IF shows bits in a mask used in such message network,
FIG. 11G shows the result of the mask of FIG. 1 IF applied to the digital word shown in FIG. HE;
FIG. 12 shows the relationship between FIGS. 12A and 12B, such FIGS. 12A and 12B Showing a process flow diagram of the receive operation of a message network used in the system interface of FIG. 8;
FIG .13 shows the relationship between FIGS. 11A and 1 IB, such FIGS. 11 A and 1 IB together showing a process flow diagram of the acknowledgement operation of a message network used in the system interface of FIG. 8;
FIGS. 14A and 14B show process flow diagrams of the transmit DMA operation of the transmit DMA of FIG. 9; and
FIGS. 15A and 15B show process flow diagrams of the receive DMA operation of the receive DMA of FIG. 10.
DETAILED DESCRIPTION Referring now to FIG. 2, a data storage system 100 is shown for transferring data between a host computer/server 120 and a bank of disk drives 140 through a system interface 160. The system interface 160 includes: a plurality of, here 32 front-end directors 180ι-18032 coupled to the host computer/server 120 via ports-12332; a plurality of back-end directors 200!- 2003 coupled to the bank of disk drives 140 via ports 12333-1236 ; a data transfer section 240, having a global cache memory 220, coupled to the plurality of front-end directors 180ι-l 8016 and the back-end directors 200]-20016; and a messaging network 260, operative independently of the data transfer section 240, coupled to the plurality of front-end directors 1801-18032 and the plurality of back-end directors 2001-20032, as shown . The front-end and back-end directors 180τ-18032, 200ι-20032 are functionally similar and include a microprocessor (μP) 299 (i.e., a central processing unit (CPU) and RAM), a message engine/ CPU controller 314 and a data pipe 316 to be described in detail in connection with FIGS. 5, 6 and 7. Suffice it to say here, however, that the front-end and back-end directors 1801-18θ32j 2001-20θ32 control data transfer between the host computer/server 120 and the bank of disk drives 140 in response to messages passing between the directors I8O1-I8O32, 2001-20θ32 through the messaging network 260. The messages facilitate the data transfer between host computer/server 120 and the bank of disk drives 140 with such data passing through the global cache memory 220 via the data transfer section 240. More particularly, in the case of the front-end directors I8O1-I8O32, the data passes between the host computer to the global cache memory 220 through the data pipe 316 in the front-end directors 1801-18θ32 and the messages pass through the message engine/CPU controller 314 in such front-end directors 1801-18θ32. In the case of the back-end directors 2001-20θ32 the data passes between the back-end directors 2001-20θ32 and the bank of disk drives 140 and the global cache memory 220 through the data pipe 316 in the back-end directors 2001-20θ32 and again the messages pass through the message engine/CPU controller 314 in such back-end director 200i- 20032. With such an arrangement, the cache memory 220 in the data transfer section 240 is not burdened with the task of transferring the director messaging. Rather the messaging network 260 operates independent of the data transfer section 240 thereby increasing the operating bandwidth of the system interface 160.
In operation, and considering first a read request by the host computer/server 120 (i.e., the host computer/server 120 requests data from the bank of disk drives 140), the request is passed from one of a plurality of, here 32, host computer processors
Figure imgf000008_0001
in the host computer 120 to one or more of the pair of the front-end directors 180i-18θ32 connected to such host computer processor 1211-12132. (It is noted that in the host computer 120, each one of the host computer processors 1211 -12132 is coupled to here a pair (but not limited to a pair) of the front-end directors I8O1-I8O32, to provide redundancy in the event of a failure in one of the front end-directors I8I1-I8I32 coupled thereto. Likewise, the bank of disk drives 140 has a plurality of, here 32, disk drives 141ι-14132, each disk drive 141H4132 being coupled to here apair (but not limited to a pair) of the back-end directors 200^20032, to provide redundancy in the event of a failure in one of the back-end directors 2001-20θ32 coupled thereto). Each front-end director lδOi-18032 includes a microprocessor (μP) 299 (i.e., a central processing unit (CPU) and RAM) and will be described in detail in connection with FIGS. 5 and 7. Suffice it to say here, however, that the microprocessor 299 makes a request for the data from the global cache memory 220. The global cache memory 220 has a resident cache management table, not shown. Every director 180ι-l 8θ32, 200^20032 has access to the resident cache management table and every time a front-end director 1801-18θ32 requests a data transfer, the front-end director I8O1-I8O32 must query the global cache memory 220 to determine whether the requested data is in the global cache memory 220. If the requested data is in the global cache memory 220 (i.e., a read "hit"), the front-end director 180j - 18O32, more particularly the microprocessor 299 therein, mediates a DMA (Direct Memory Access) operation for the global cache memory 220 and the requested data is transferred to the requesting host computer processor 1211-12l32. If, on the other hand, the front-end director 1801-18θ32 receiving the data request determines that the requested data is not in the global cache memory 220 (i.e., a "miss") as a result of a query of the cache management table in the global cache memory 220, such front-end director 18θ!-18θ32 concludes that the requested data is in the bank of disk drives 140. Thus the front-end director 1801-18θ32 that received the request for the data must make a request for the data from one of the back-end directors 20θ!-20θ32 in order for such back-end director 200^20032 to request the data from the bank of disk drives 140. The mapping of which back-end directors 2001-20032 control which disk drives 141i-14l32 in the bank of disk drives 140 is determined during a power-up initialization phase. The map is stored in the global cache memory 220. Thus, when the front- end director 180} -18O32 makes a request for data from the global cache memory 220 and determines that the requested data is not in the global cache memory 220 (i.e., a "miss"), the front-end director I8O1-I8O32 is also advised by the map in the global cache memory 220 of the back-end director 200 20θ32 responsible for the requested data in the bank of disk drives 140. The requesting front-end director 18O1-I8O32 then must make a request for the data in the bank of disk drives 140 from the map designated back-end director 200 20θ32. This request between the front-end director 180rl8θ32 and the appropriate one of the back-end directors 2OO1-2OO32 (as determined by the map stored in the global cache memory 200) is by a message which passes from the front-end director 18O1-18O32 through the message network 260 to the appropriate back- end director 200^20032- It is noted then that the message does not pass through the global cache memory 220 (i.e., does not pass through the data transfer section 240) but rather passes through the separate, independent message network 260. Thus, communication between the directors 1801-18032, 20θ!-20θ32 is through the message network 260 and not through the global cache memory 220. Consequently, valuable bandwidth for the global cache memory 220 is not used for messaging among the directors I8O1-I8O32, 200i-20θ32.
Thus, on a global cache memory 220 "read miss", the front-end director 180^18032 sends a message to the appropriate one of the back-end directors 200^20032 through the message network 260 to instruct such back-end director 200i-20θ32 to transfer the requested data from the bank of disk drives 140 to the global cache memory 220. When accomplished, the back-end director 200^20032 advises the requesting front-end director I8O1-I8O32 that the transfer is accomplished by a message, which passes from the back-end director 20θ!-20θ32 to the front-end director 1801-18θ32 through the message network 260. In response to the acknowledgement signal, the front-end director 180rl 8O32 is thereby advised that such front-end director 180^ 18θ32 can transfer the data from the global cache memory 220 to the requesting host computer processor 1211-12l32 as described above when there is a cache "read hit". It should be noted that there might be one or more back-end directors 2001-20032 responsible for the requested data. Thus, if only one back-end director 20θ!-20032 is responsible for the requested data, the requesting front-end director 180^18O32 sends a uni-cast message via the message network 260 to only that specific one of the back-end directors 200^20032. On the other hand, if more than one of the back-end directors 2OO1-2OO32 is responsible for the requested data, a multi-cast message (here implemented as a series of uni-cast messages) is sent by the requesting one of the front-end directors I8O1-I8O32 to all of the back-end directors 2OO1-2OO32 having responsibility for the requested data. In any event, with both a uni-cast or multi-cast message, such message is passed through the message network 260 and not through the data transfer section 240 (i.e., not through the global cache memory 220). Likewise, it should be noted that while one of the host computer processors
Figure imgf000010_0001
might request data, the acknowledgement signal may be sent to the requesting host computer processor 1211 or one or more other host computer processors I2I1-I2I32 via a multi-cast (i.e., sequence of uni-cast) messages through the message network 260 to complete the data read operation. Considering a write operation, the host computer 120 wishes to write data into storage (i.e., into the bank of disk drives 140). One of the front-end directors I8O1-I8O32 receives the data from the host computer 120 and writes it into the global cache memory 220. The front-end director 18O1-18O32 then requests the transfer of such data after some period of time when the back-end director 200r-20θ32 determines that the data can be removed from such cache memory 220 and stored in the bank of disk drives 140. Before the transfer to the bank of disk drives 140, the data in the cache memory 220 is tagged with a bit as "fresh data" (i.e., data which has not been transferred to the bank of disk drives 140, that is data which is "write pending"). Thus, if there are multiple write requests for the same memory location in the global cache memory 220 (e.g., a particular bank account) before being transferred to the bank of disk drives 140, the data is overwritten in the cache memory 220 with the most recent data. Each time data is transferred to the global cache memory 220, the front-end director 1801- 18O32 controlling the transfer also informs the host computer 120 that the transfer is complete to thereby free-up the host computer 120 for other data transfers.
When it is time to transfer the data in the global cache memory 220 to the bank of disk drives 140, as determined by the back-end director 200^20032, the back-end director 200^20032 transfers the data from the global cache memory 220 to the bank of disk drives 140 and resets the tag associated with data in the global cache memory 220 (i.e., un-tags the data) to indicate that the data in the global cache memory 220 has been transferred to the bank of disk drives 140. It is noted that the un-tagged data in the global cache memory 220 remains there until overwritten with new data. Referring now to FIGS. 3 and 4, the system interface 160 is shown to include an electrical cabinet 300 having stored therein: a plurality of, here eight front-end director boards 190ι-1908, each one having here four of the front-end directors 180^18032; a plurality of, here eight back- end director boards 2101-2108, each one having here four of the back-end directors
Figure imgf000011_0001
and a plurality of, here eight, memory boards 220' which together make up the global cache memory 220. These boards plug into the front side of a backplane 302. (It is noted that the backplane 302 is a mid-plane printed circuit board). Plugged into the backside of the backplane 302 are message network boards 304l5 3042. The backside of the backplane 302 has plugged into it adapter boards, not shown in FIGS. 2-4, which couple the boards plugged into the back-side of the backplane 302 with the computer 120 and the bank of disk drives 140 as shown in FIG. 2. That is, referring again briefly to FIG. 2, an I/O adapter, not shown, is coupled between each one of the front-end directors 180ι-18θ32and the host computer 120 and an I/O adapter, not shown, is coupled between each one of the back-end directors 200 20θ32 and the bank of disk drives 140. Referring now to FIG. 5, the system interface 160 is shown to include the director boards 190ι_ 1908, 210^2108 and the global cache memory 220, plugged into the backplane 302 and the disk drives 141ι-14132 in the bank of disk drives along with the host computer 120 also plugged into the backplane 302 via I/O adapter boards, not shown. The message network 260 (FIG. 2) includes the message network boards 3041 and 3042- Each one of the message network boards 304t and 3042is identical in construction. A pair of message network boards 304] and 3042 is used for redundancy and for message load balancing. Thus, each message network board 304l5 3042, includes a controller 306, (i.e., an initialization and diagnostic processor comprising a CPU, system controller interface and memory, as shown in FIG. 6 for one of the message network boards 304ι, 3042, here board 3040 and a crossbar switch section 308 (e.g., a switching fabric made up of here four switches 308^308^.
Referring again to FIG. 5, each one of the director boards 190 21θ8 includes, as noted above four of the directors I8O1-I8O32, 20θ!-20032 (FIG. 2). It is noted that the director boards 190ι- 190s having four front-end directors per board, 180^18O32 are referred to as front-end directors, and the director boards 2101-210s having four back-end directors per board, 200^20032 are referred to as back-end directors. Each one of the directors 180^18032, 200^20032 includes a CPU 310, a RAM 312 (which make up the microprocessor 299 referred to above), the message engine/CPU controller 314, and the data pipe 31 .
Each one of the director boards 190^2108 includes a crossbar switch 318. The crossbar switch 318 has four input/output ports 319, each one being coupled to the data pipe 316 of a corresponding one of the four directors I8O1-I8O32, 200 20θ32 on the director board.l90!-2108. The crossbar switch 318 has eight output input ports collectively identified in FIG. 5 by numerical designation 321 (which plug into the backplane 302. The crossbar switch 318 on the front-end director boards 191ι-191g is used for coupling the data pipe 316 of a selected one of the four front-end directors 180Ϊ-18032 on the front-end director board 190rl9θ8 to the global cache memory 220 via the backplane 302 and I/O adapter, not shown. The crossbar switch 318 on the back-end director boards 210^2108 is used for coupling the data pipe 316 of a selected one of the four back-end directors 2001-20θ32 on the back-end director board 210^210s to the global cache memory 220 via the backplane 302 and I/O adapter, not shown. Thus, referring to FIG. 2, the data pipe 316 in the front-end directors 180 l8θ32 couples data between the host computer 120 and the global cache memory 220 while the data pipe 316 in the back-end directors 200^20032 couples data between the bank of disk drives 140 and the global cache memory 220. It is noted that there are separate point-to-point data paths Pι-P64 (FIG. 2) between each one of the directors I8O1-I8O32,
Figure imgf000012_0001
and the global cache memory 220. It is also noted that the backplane 302 is a passive backplane because it is made up of only etched conductors on one or more layers of a printed circuit board. That is, the backplane 302 does not have any active components.
Referring again to FIG. 5, each one of the director boards 190^2108 includes a crossbar switch 320. Each crossbar switch 320 has four input/output ports 323, each one of the four input/output ports 323 being coupled to the message engine/CPU controller 314 of a corresponding one of the four directors 1801-1803 , 200^20032 on the director board 190]-2108. Each crossbar switch 320 has a pair of output input ports 325l9 3252, which plug into the backplane 302. Each port 325ι- 3252 is coupled to a corresponding one of the message network boards 304l5 3042, respectively, through the backplane 302. The crossbar switch 320 on the front-end director boards 190 l9θ8 is used to couple the messages between the message engine/CPUcontroller 314 of a selected one of the four front-end directors 180ι-l 8032 on the front-end director boards 190ι-190g and the message network 260, FIG. 2. Likewise, the back-end director boards 210r2108 are used to couple the messages produced by a selected one of the four back-end directors 200^20032 on the back-end director board 21 Oi-210s between the message engine/CPU controller 314 of a selected one of such four back-end directors and the message network 260 (FIG. 2). Thus, referring also to FIG. 2, instead of having a separate dedicated message path between each one of the directors 1801-18032, 2001-20θ32 and the message network 260 (which would require M individual connections to the backplane 302 for each of the directors, where M is an integer), here only M/4 individual connections are required). Thus, the total number of connections between the directors 180^18032, 200ι-20032 and the backplane 302 is reduced to l/4th. Thus, it should be noted from FIGS. 2 and 5 that the message network 260 (FIG. 2) includes the crossbar switch 320 and the message network boards 304ls 3042. Each message is a 64-byte descriptor, shown in FIG. 2 A) which is created by the CPU 310 (FIG. 5) under software control and is stored in a send queue in RAM 312. When the message is to be read from the send queue in RAM 312 and transmitted through the message network 260 (FIG. 2) to one or more other directors via a DMA operation to be described, it is packetized in the packetizer portion of packetizer/de-packetizer 428 (FIG. 7) into a MAC type packet, shown in FIG. 2B, here using the NGIO protocol specification. There are three types of packets: a message packet section; an acknowledgement packet; and a message network fabric management packet, the latter being used to establish the message network routing during initialization (i.e., during power-up). Each one of the MAC packets has: an 8-byte header which includes source (i.e., transmitting director) and destination (i.e., receiving director) address; a payload; and terminates with a 4-byte Cyclic Redundancy Check (CRC), as shown in FIG. 2B. The acknowledgement packet (i.e., signal) has a 4-byte acknowledgment payload section. The message packet has a 32-byte payload section. The Fabric Management Packet (FMP) has a 256-byte payload section. The MAC packet is sent to the crossbar switch 320. The destination portion of the packet is used to indicate the destination for the message and is decoded by the switch 320 to determine which port the message is to be routed. The decoding process uses a decoder table 327 in the switch 318, such table being initialized by controller during power-up by the initialization and diagnostic processor (controller) 306 (FIG. 5). The table 327 (FIG. 7) provides the relationship between the destination address portion of the MAC packet, which identifies the routing for the message and the one of the four directors 180^18O32, 2001-20032 on the director board 190ι-19θ8, 210r21θ8 or to one of the message network boards 304ι, 3042 to which the message is to be directed.
More particularly, and referring to FIG. 5, a pair of output/input ports 325 \, 3252 is provided for each one of the crossbar switches 320, each one being coupled to a corresponding one of the pair of message network boards 304l5 3042. Thus, each one of the message network boards 304l5 3042 has sixteen input/output ports 322^322^, each one being coupled to a corresponding one of the output/input ports 3251, 3252, respectively, of a corresponding one of the director boards 190rl9θ8, 210ι-21θ8 through the backplane 302, as shown. Thus, considering exemplary message network board 304ls FIG. 6, each switch 3081-3084 also includes three coupling ports 324 3243. The coupling ports 324^ 3243 are used to interconnect the switches 322i-3224, as shown in FIG. 6. Thus, considering message network board 304ls input/output ports 322^3228 are coupled to output/input ports 325ι of front-end director boards 190^1908 and input/output ports 3229-32216 are coupled to output/input ports 325t of back-end director boards 210ι-2108, as shown. Likewise, considering message network board 3042, input/output ports 322!-3228 thereof are coupled, via the backplane 302, to output/input ports 3252 of front-end director boards 190!- 190s and input/output ports 322 -32216 are coupled, via the backplane 302, to output/input ports 3252 of back-end director boards 2101-2108.
As noted above, each one of the message network boards 304ι, 3042 includes a processor 306 (FIG. 5) and a crossbar switch section 308 having four switches 308ι-3084, as shown in FIGS. 5 and 6. The switches 3081-3084 are interconnected as shown so that messages can pass between any pair of the input/output ports 3221 -32216. Thus, it follow that a message from any one of the front-end directors 180 l8θ32 can be coupled to another one of the front-end directors 180^ I8O32 and/or to any one of the back-end directors 200ι-20032. Likewise, a message from any one of the back-end directors 180^18032 can be coupled to another one of the back-end directors 180ι_l 8O32 and/or to any one of the front-end directors 2001-20θ3 .
As noted above, each MAC packet (FIG. 2B) includes in an address destination portion and a data payload portion. The MAC header is used to indicate the destination for the MAC packet and such MAC header is decoded by the switch to determine which port the MAC packet is to be routed. The decoding process uses a table in the switch 308ι-3084, such table being initialized by processor 306 during power-up. The table provides the relationship between the MAC header, which identifies the destination for the MAC packet and the route to be taken through the message network. Thus, after initialization, the switches 320 and the switches 3081-308 in switch section 308 provides packet routing which enables each one of the directors
Figure imgf000015_0001
20θ!-20032 to transmit a message between itself and any other one of the directors, regardless of whether such other director is on the same director board 1901- 190s, 210ι-210s or on a different director board. Further, the MAC packet has an additional bit B in the header thereof, as shown in FIG. 2B, which enables the message to pass through message network board 304i or through message network board 3042- During normal operation, this additional bit B is toggled between a logic 1 and a logic 0 so that one message passes through one of the redundant message network boards 304ls 3042 and the next message to pass through the other one of the message network boards 3041; 3042 to balance the load requirement on the system. However, in the event of a failure in one of the message network boards 304l5 3042, the non-failed one of the boards 304ι, 3042 is used exclusively until the failed message network board is replaced.
Referring now to FIG. 7, an exemplary one of the director boardslθOt-lθOs, 2101-210s, here director board 190ι is shown to include directors 1801? I8O3, 18O5 and 1807. An exemplary one of the directors 1801-180 , here director I8O1 is shown in detail to include the data pipe 316, the message engine/CPU controller 314, the RAM 312, and the CPU 310 all coupled to the CPU interface bus 317, as shown. The exemplary director 180ι also includes: a local cache memory 319 (which is coupled to the CPU 310); the crossbar switch 318; and, the crossbar switch 320, described briefly above in connection with FIGS. 5 and 6. The data pipe 316 includes a protocol translator 400, a quad port RAM 402 and a quad port RAM controller 404 arranged as shown. Briefly, the protocol translator 400 converts between the protocol of the host computer 120, in the case of a front-end director 180^18032, (and between the protocol used by the disk drives in bank 140 in the case of a back-end director 2001-20θ32) and the protocol between the directors 18O1-1803, 2OO1-2OO32 and the global memory 220 (FIG. 2). More particularly, the protocol used the host computer 120 may, for example, be fibre channel, SCSI, ESCON or FICON, for example, as determined by the manufacture of the host computer 120 while the protocol used internal to the system interface 160 (FIG. 2) may be selected by the manufacturer of the interface 160. The quad port RAM 402 is a FIFO controlled by controller 404 because the rate data coming into the RAM 402 may be different from the rate data leaving the RAM 402. The RAM 402 has four ports, each adapted to handle an 18 bit digital word. Here, the protocol translator 400 produces 36 bit digital words for the system interface 160 (FIG. 2) protocol, one 18 bit portion of the word is coupled to one of a pair of the ports of the quad port RAM 402 and the other 18 bit portion of the word is coupled to the other one of the pair of the ports of the quad port RAM 402. The quad port RAM has a pair of ports 402A, 402B, each one of to ports 402A, 402B being adapted to handle an 18 bit digital word. Each one of the ports 402 A, 402B is independently controllable and has independent, but arbitrated, access to the memory array within the RAM 402. Data is transferred between the ports 402A, 402B and the cache memory 220 (FIG. 2) through the crossbar switch 318, as shown. The crossbar switch 318 includes a pair of switches 406A, 406B. Each one of the switches 406A, 406B includes four input/output director-side ports DΪ-D (collectively referred to above in connection with FIG. 5 as port 319) and four input/output memory-side ports M1-M4, Ms-Ms, respectively, as indicated. The input/output memory-side ports MrM , M5-M8 were collectively referred to above in connection with FIG. 5 as port 317). The director-side ports DrD4 of switch 406A are connected to the 402 A ports of the quad port RAMs 402 in each one the directors 180l5 I8O3, 1805 and I8O7, as indicated. Likewise, director-side ports of switch 406B are connected to the 402B ports of the quad port RAMs 402 in each one the directors 180l5 18O3, 1805, and 1807, as indicated. The ports D1-D4 are selectively coupled to the ports M1-M4 in accordance with control words provided to the switch 406A by the controllers in directors I8O1, 18O3, 18O5, 180 on busses RArRA4, respectively, and the ports D D4 are coupled to ports M5-Ms in accordance with the control words provided to switch 406B by the controllers in directors 180ls 18O3, 18O5, I8O7 on busses RBI-RB45 as indicated. The signals on buses RAI-RA4 are request signals. Thus, port 402A of any one of the directors 180l5 I8O3, 18O5, 180η may be coupled to any one of the ports M1-M4 of switch 406A, selectively in accordance with the request signals on buses RAI- RA . Likewise, port 402B of any one of the directors 18θ!-1804 may be coupled to any one of the ports M5-Mg of switch 406B, selectively in accordance with the request signals on buses RBI- RB4. The coupling between the director boards 190^1908, 210^210s and the global cache memory 220 is shown in FIG. 8.
More particularly, and referring also to FIG. 2, as noted above, each one of the host computer processors 1211 -I2I32 in the host computer 120 is coupled to a pair of the front-end directors I8O1-I8O32, to provide redundancy in the event of a failure in one of the front end-directors 1811- I8I32 coupled thereto. Likewise, the bank of disk drives 140 has a plurality of, here 32, disk drives 1411-14l32, each disk drive 141ι-14l32 being coupled to a pair of the back-end directors 2OO1-2OO32, to provide redundancy in the event of a failure in one of the back-end directors 200!- 2OO32 coupled thereto). Thus, considering exemplary host computer processorl21 ls such processor 121ι is coupled to a pair of front-end directors 180ι, I8O2. Thus, if director I8O1 fails, the host computer processor 121ι can still access the system interface 160, albeit by the other front-end director 1802. Thus, directors I8O1 and I8O2 are considered redundancy pairs of directors. Likewise, other redundancy pairs of front-end directors are: front-end directors I8O3, 1804; I8O5, 1806; I8O7, 1808; 1809, I8O10; I8O11, 18O12; I8O13, 18O14; I8O15, 18016; 18017, 180l8; I8O19, 18O20; I8O21, 18O22; 18023, 18024; I8O25, 18026; 18027, 18028; 18029, 18030; and I8O31, 18032 (only directors I8O31 and I8O32 being shown in FIG. 2). Likewise, disk drive 1411 is coupled to a pair of back-end directors 200l5 2002. Thus, if director 200 ! fails, the disk drive 1411 can still access the system interface 160, albeit by the other back- end director 1802. Thus, directors 200! and 2OO2 are considered redundancy pairs of directors. Likewise, other redundancy pairs of back-end directors are: back-end directors 2003, 200 ; 200 , 2006;2007,2008; 2009;20θ!0; 200ll5 20012; 20013, 20014;20θ!5, 20016; 20017, 20018;20019, 20020; 2OO21, 20022;20θ23, 20024; 20025, 20026; 20027, 20028; 200 9, 20030; and2003ι, 20032(only directors 2OO3! and 20032 being shown in FIG. 2). Further, referring also to FIG. 8, the global cache memory 220 includes a plurality of, here eight, cache memory boards 22θ!-220s, as shown. Still further, referring to FIG. 8 A, an exemplary one of the cache memory boards, here board 2201 is shown in detail and is described in detail in U. S. Patent No. 5,943,287 entitled "Fault Tolerant Memory System", John K. Walton, inventor, issued August 24, 1999 and assigned to the same assignee as the present invention, the entire subject matter therein being incorporated herein by reference. Thus, as shown in FIG. 8A, the board 220i includes a plurality of, here four RAM memory arrays, each one of the arrays has a pair of redundant ports, i.e., an A port and a B port. The board itself has sixteen ports; a set of eight A ports MAI -MAS and a set of eight B ports MB!- MB8- Four of the eight A port, here A ports MAI-MA are coupled to the M\ port of each of the front-end director boards 190l5 1903, 190s, and 1907, respectively, as indicated in FIG. 8. Four of the eight B port, here B ports MB1-MB4 are coupled to the Mi port of each of the front-end director boards 1902, 1904, 1906, and 190s, respectively, as indicated in FIG. 8. The other four of the eight A port, here A ports MAS- AS are coupled to the Mi port of each of the back-end director boards 210ι, 2103, 2105, and 2107, respectively, as indicated in FIG. 8. The other four of the eight B port, here B ports MB5-M48 are coupled to the Mi port of each of the back-end director boards 2IO2, 2104, 2106, and 2108, respectively, as indicated in FIG. 8 Considering the exemplary four A ports MAI-MA4, each one of the four A ports MAI-MA4 can be coupled to the A port of any one of the memory arrays through the logic network 2211 A- Thus, considering port MAI, such port can be coupled to the A port of the four memory arrays. Likewise, considering the four A ports MAS-MAS, each one of the four A ports MAS-MAS can be coupled to the A port of any one of the memory arrays through the logic network 221 m. Likewise, considering the four B ports MB!-MB4, each one of the four B ports MB1-MB4 can be coupled to the B port of any one of the memory arrays through logic network 221 IB. Likewise, considering the four B ports MBS-MBS, each one of the four B ports MBS-MBS can be coupled to the B port of any one of the memory arrays through the logic network 2212β. Thus, considering port MBI, such port can be coupled to the B port of the four memory arrays. Thus, there are two paths data and control from either a front-end director 18θ!-18θ32 or a back-end director 200^ 2OO32 can reach each one of the four memory arrays on the memory board. Thus, there are eight sets of redundant ports on a memory board, i.e., ports MAI, MBΪ; MA2, MB2; MA3, MB3; MA4, MB ; M S, MB5; MA6, MB6; MA7, MB7; and MAS, MBS- Further, as noted above each one of the directors has a pair of redundant ports, i.e. a 402A port and a 402 B port (FIG. 7). Thus, for each pair of redundant directors, the A port (i.e., port 402A) of one of the directors in the pair is connected to one of the pair of redundant memory ports and the B port (i.e., 402B) of the other one of the directors in such pair is connected to the other one of the pair of redundant memory ports. More particularly, referring to FIG. 8B, an exemplary pair of redundant directors is shown, here, for example, front-end director 180i and front end-director I8O2. It is first noted that the directors I8O1, 18O2 in each redundant pair of directors must be on different director boards, here boards 190l5 1902, respectively. Thus, here front-end director boards 19θ!-190s have thereon: front-end directors 1801; I8O3, 1805 and 1807; front-end directors 180 , 1804, 1806 and 1808; front end directors 180 > 180π, 180B and 18015; front end directors 18010, 180^, 1801 and 18016; front-end directors 18017, 18019, 18O21 and I8O23; front-end directors 18018, 1802o, 18θ22 and I8O24; front-end directors 18025, 18027, 18029 and 18031; front-end directors 18018, 18020, 18022 and 18O24. Thus, here back-end director boards 210ι-2108 have thereon: back-end directors 2001; 2OO3, 2OO5 and 2007; back-end directors 2002, 200 , 2006 and 2008; back-end directors 2009) 200ιι, 20013 and 20015; back-end directors 20010) 200ι2, 20014 and 20016; back-end directors 20017; 20019, 20021 and 20023; back-end directors 20018, 20020, 20022 and 2002 ; back-end directors 20025, 20027, 20029 and 20031; back-end directors 20018; 20020, 20022 and 2002 ; Thus, here front-end director 180l5 shown in FIG. 8A, is on front-end director board 190ι and its redundant front-end director 1802, shown in FIG. 8B, is on anther front-end director board, here for example, front-end director board 1 02. As described above, the port 402 A of the quad port RAM 402 (i.e., the A port referred to above) is connected to switch 406A of crossbar switch 318 and the port 402B of the quad port RAM 402 (i.e., the B port referred to above) is connected to switch 406B of crossbar switch 318. Likewise, for redundant director 1802, However, the ports M1-M4 of switch 406A of director 180ι are connected to the MAI ports of global cache memory boards 22θ!-2004, as shown, while for its redundancy director I8O2, the ports M!-M4 of switch 406A are connected to the redundant MB! ports of global cache memory boards 22θ!-2004, as shown. .
Referring in more detail to the crossbar switch 318 (FIG. 7), as noted above, each one of the director boards 190^2108 has such a switch 318 and such switch 318 includes a pair of switches 406 A, 406B. Each one of the switches 406 A, 406B is identical in construction, an exemplary one thereof, here switch 406A being shown in detail in FIG. 8C. Thus switch 406A includes four input/output director-side ports D!-D4 as described in connection with exemplary director board 190ι. Thus, for the director board 190ι shown in FIG. 7, the four input/output director- side ports D!-D of switch 406 A are each coupled to the port 402A of a corresponding one of the directors I8O1, I8O3, 1805, and 1807 on the director board 190ι. Referring again to FIG. 8C, the exemplary switch 406A includes a plurality of, here four, switch sections 430ι-4304. Each one of the switch sections 430i-4304 is identical in construction and is coupled between a corresponding one of the input/output director-side ports U!-D4 and a corresponding one of the output/input memory-side ports M1-M4, respectively, as shown. (It should be understood that the output/input memory-side ports of switch 406B (FIG. 7) are designated as ports Ms-Ms, as shown. It should also be understood that while switch 406A is responsive to request signals on busses R I- A4 from quad port controller 404 in directors 180ls I8O3, 18O5, 18O7 (FIG. 7), switch 406B is responsive in like manner to request signals on busses RBI-RB4 from controller 404 in directors 180ls I8O3, 18O5 and 1807). More particularly, controller 404 of director I8O1 produces request signals on busses RAI or RBI. In like manner, controller 404 of director 18O3 produces request signals on busses RA2 or RB2, controller 404 of director I8O5 produces request signals on busses RA3 or Rβ3, and controller 404 of director I8O7 produces request signals on busses R 4 or Rβ4.
Considering exemplary switch section 430ι, such switch section 403! is shown in FIG. 8C to include a FIFO 432 fed by the request signal on bus RΪA- (It should be understood that the FIFOs, not shown, in switch sections 4302-4304 are fed by request signals RA2- A4, respectively). The switch section 406\ also includes a request generation 434, and arbiter 436, and selectors 442 and 446, all arranged as shown. The data at the memory-side ports M1-M4 are on busses DM1 -DM4 are fed as inputs to selector 446. Also fed to selector 446 is a control signal produced by the request generator on bus 449 in response to the request signal RA! stored in FIFO 432. The control signal on bus 449 indicates to the selector 446 the one of the memory- side ports M1-M4 which is to be coupled to director-side port Di. The other switch sections 4302- 4304 operate in like manner with regard to director-side ports Di-D4, respectively and the memory-side ports Mι-M4.
It is to be noted that the data portion of the word at port Di (i.e., the word on bus DD1) is also coupled to the other switch sections 4302-4304. It is further noted that the data portion of the words at ports D2-D (i.e., the words on busses DD2-DD4, respectively), are fed to the switch sections 430ι-4304, as indicated. That is, each one of the switch sections 430r430 has the data portion of the words on ports Dι-D4 (i.e., busses DD1-DD4), as indicated. It is also noted that the data portion of the word at port Mi (i.e., the word on bus DM1) is also coupled to the other switch sections 4302-4304. It if further noted that the data portion of the words at ports M2-M4 (i.e., the words on busses DM2-DM4, respectively), are fed to the switch sections 4302-430 , as indicated. That is, each one of the switch sections 430ι-4304 has the data portion of the words on ports M1-M4 (i.e., busses DM1 -DM4), as indicated.
As will be described in more detail below, a request on bus RAI to switch section 430i is a request from the director 180i which identifies the one of the four ports M1-M in switch 430ι is to be coupled to port 402A of director 180i (director side port D ). Thus, port 402A of director I8O1 may be coupled to one of the memory side ports Mι-M selectively in accordance with the data on bus RAI- Likewise, a request on buses RA2, RA3, RA4 to switch section 43θ2-430 , respectively, are requests from the directors I8O3, 18O5, and 1807, respectively, which identifies the one of the four ports MrM4 in switch 430ι-4304 is to be coupled to port 402 A of directors I8O3, 18O5 and I8O7, respectively. More particularly, the requests RAI are stored as they are produced by the quad port RAM controller 440 (FIG. 7) in receive FIFO 432. The request generator 434 receives from FIFO 432 the requests and determines which one of the four memory-side ports M1-M4 is to be coupled to port 402A of director I8O1. These requests for memory-side ports M1-M4 are produced on lines RA1,1 - RAI, 4, respectively. Thus, line RA1,1 (i.e., the request for memory side port Mi) is fed to arbiter 436 and the requests from switch sections 4302-4304 (which are coupled to port 402 A of directors I8O3, 1805, and I8O7) on line RA2,1, RA3,1 and RA4,1, respectively are also fed to the arbiter 436, as indicated. . The arbiter 436 resolves multiple requests for memory-side port Mi on a first come-first serve basis. The arbiter 436 then produces a control signal on bus 435 indicating the one of the directors 180ι, I8O3, 18O5 or I8O7 which is to be coupled to memory-
Figure imgf000021_0001
The control signal on bus 435 is fed to selector 442. Also fed to selector 442 are the data portion of the data at port Di, i.e., the data on data bus DD1) along with the data portion of the data at ports D2-D4, i.e., the data on data busses DD2-DD4, respectively, as indicated. Thus, the control signal on bus 435 causes the selector 442 to couple to the output thereof the data busses DD1- DD4 from the one of the directors I8O1, 18O3, 1805, I8O7 being granted access to memory-side port Mi by the arbiter 436. The selected outputs of selector 442 is coupled to memory-side port Mi. It should be noted that when the arbiter 436 receives a request via the signals on lines RAI , 1 , RA2, 1 , RA3 , 1 and RA4, 1 , acknowledgements are returned by the arbiter 436 via acknowledgement signals on line AK1,1, AK1,2, AK1,3, AK1,4, respectively such signals being fed to the request generators 434 in switch section 430ι, 4302, 4303, 430 , respectively. Thus, the data on any port D D4 can be coupled to and one of the ports M1-M4 to effectuate the point-to-point data paths Pι-P64 described above in connection with FIG. 2. Referring again to FIG. 7, data from host computer 120 (FIG. 2) is presented to the system interface 160 (FIG. 2) in batches from many host computer processors 1211-12l32. Thus, the data from the host computer processors 1211-12l3 re interleaved with each other as they are presented to a director I8O1-18θ3 . The batch from each host computer processor I8O1-18O32 (i.e., source) is tagged by the protocol translator 400. More particularly by a Tacheon ASIC in the case of a fibre channel connection. The controller 404 has a look-up table formed during initialization. As the data comes into the protocol translator 400 and is put into the quad port RAM 420 under the control of controller 404, the protocol translator 400 informs the controller that the data is in the quad port RAM 420. The controller 404 looks at the configuration of its look-up table to determine the global cache memory 220 location (e.g., cache memory board 22θ!-2208) the data is to be stored into. The controller 404 thus produces the request signals on the appropriate bus RAI, RBI, and then tells the quad port RAM 402 that there is a block of data at a particular location in the quad port RAM 402, move it to the particular location in the global cache, memory 220. The crossbar switch 318 also takes a look at what other controllers 404 in the directors I8O3, 18O5, and I8O7 on that particular director board 190ι are asking by making request signal on busses RA2, RB2, A3, RB3, RA4, RB4, respectively. The arbitration of multiple requests is handled by the arbiter 436 as described above in connection with FIG. 8C. Referring again to FIG. 7, the exemplary director I8 1 is shown to include in the message engine/CPU controller 314. The message engine/CPU controller 314 is contained in a field programmable gate array (FPGA). The message engine (ME) 315 is coupled to the CPU bus 317 and the DMA section 408 as shown. The message engine (ME) 315 includes a Direct Memory Access (DMA) section 408, a message engine (ME) state machine 410, a transmit buffer 424 and receive buffer 424, a MAC packetizer/depacketizer 428, send and receive pointer registers 420, and a parity generator 321. The DMA section 408 includes a DMA transmitter 418, shown and to be described below in detail in connection with FIG. 9, and a DMA receiver 424, shown and to be described below in detail in connection with FIG.10, each of which is coupled to the CPU bus interface 317, as shown in FIG. 7. The message engine (ME) 315 includes a transmit data buffer 422 coupled to the DMA transmitter 418, a receive data buffer 424 coupled to the DMA receiver 421 , registers 420 coupled to the CPU bus 317 through an address decoder 401 , the packetizer/de-packetizer 428, described above, coupled to the transmit data buffer 422, the receive data buffer 424 and the crossbar switch 320, as shown, and a parity generator 321 coupled between the transmit data buffer 422 and the crossbar switch 320. More particularly, the packetizer portion 428P is used to packetize the message payload into a MAC packet (FIG. 2B) passing from the transmit data buffer 422 to the crossbar switch 320 and the de-packetizer portion 428D is used to de-packetize the MAC packet into message payload data passing from the crossbar switch 320 to the receive data buffer 424. The packetization is here performed by a MAC core which builds a MAC packet and appends to each message such things as a source and destination address designation indicating the director sending and receiving the message and a cyclic redundancy check (CRC), as described above. The message engine (ME) 315 also includes: a receive write pointer 450, a receive read pointer 452; a send write pointer 454, and a send read pointer 456.
Referring now to FIGS. 11 and 12, the transmission of a message from a director I8O1-I8O32, 2OO1-2OO32 and the reception of a message by a director 2IO1-2IO32, here exemplary director ISOi shown in FIG. 7) will be described. Considering first transmission of a message, reference is made to FIGS. 7 and 11. First, as noted above, at power-up the controller 306 (FIG. 5) of both message network boards 304l5 3042 initialize the message routing mapping described above for the switches 308!-3084 in switch section 308 and for the crossbar switches 320. As noted above, a request is made by the host computer 120. The request is sent to the protocol translator 400. The protocol translator 400 sends the request to the microprocessor 299 via CPU bus 317 and buffer 301. When the CPU 310 (FIG. 7) in the microprocessor 299 of exemplary director I8O1 determines that a message is to be sent to another one of the directors I8O2-I8O32, 200 20032, (e.g., the CPU 310 determines that there has been a "miss" in the global cache memory 220 (FIG. 2) and wants to send a message to the appropriate one of the back-end directors 20θ!-20θ32, as described above in connection with FIG. 2), the CPU 310 builds a 64 byte descriptor (FIG. 2 A) which includes a 32 byte message payload indicating the addresses of the batch of data to be read from the bank of disk drives 140 (FIG. 2) (Step 500) and a 32 byte command field (Step 510) which indicates the message destination via an 8-byte bit vector, i.e., the director, or directors, which are to receive the message. An 8-byte portion of the command field indicates the director or directors, which are to receive the message. That is, each one of the 64 bits in the 8-byte portion corresponds to one of the 64 directors. Here, a logic 1 in a bit indicates that the corresponding director is to receive a message and a logic 0 indicates that such corresponding director is not to receive the message. Thus, if the 8-byte word has more than one logic 1 bit more than one director will receive the same message. As will be described, the same message will not be sent in parallel to all such directors but rather the same message will be sent sequentially to all such directors. In any event, the resulting 64-byte descriptor is generated by the CPU 310 (FIG. 7) (Step 512) is written into the RAM 312 (Step 514), as shown in FIG. 11. More particularly, the RAM 512 includes a pair of queues; a send queue and a receive queue, as shown in FIG. 7. The RAM 312 is coupled to the CPU bus 317 through an Error Detection and Correction (EDAC)/Memory control section 303, as shown. The CPU 310 then indicates to the message engine (ME) 315 state machine 410 (FIG. 7) that a descriptor has been written into the RAM 312. It should be noted that the message engine (ME) 315 also includes: a receive write pointer or counter 450, the receive read pointer or counter 452, the send write pointer or counter 454, and the send read pointer or counter 454, shown in FIG. 7. All four pointers 450, 452, 454 and 456 are reset to zero on power-up. As is also noted above, the message engine/CPU controller 314 also includes: the de-packetizer portion 428D of packetizer/de-packetizer 428, coupled to the receive data buffer 424 (FIG. 7) and a packetizer portion 428P of the packetizer/de-packetizer 428, coupled to the transmit data buffer 422 (FIG. 7). Thus, referring again to FIG. 11, when the CPU 310 indicates that a descriptor has been written into the RAM 312 and is now ready to be sent, the CPU 310 increments the send write pointer and sends it to the send write pointer register 454 via the register decoder 401. Thus, the contents of the send write pointer register 454 indicates the number of messages in the send queue 312S of RAM 312, which have not been sent. The state machine 410 checks the send write pointer register 454 and the send read pointer register 456, Step 518. As noted above, both the send write pointer register 454 and the send read pointer register 456 are initially reset to zero during power-up. Thus, if the send read pointer register 456 and the send write pointer register 454 are different, the state machine knows that there is a message is in RAM 312 and that such message is ready for transmission. If a message is to be sent, the state machine 410 initiates a transfer of the stored 64-byte descriptor to the message engine (ME) 315 via the DMA transmitter 418, FIG. 7 (Steps 520, 522). The descriptor is sent from the send queues 312S in RAM 312 until the send read pointer 456 is equal to the send write pointer 454.
As described above in connection with Step 510, the CPU 310 generates a destination vector indicating the director, or directors, which are to receive the message. As also indicated above the command field is 32-bytes, eight bytes thereof having a bit representing a corresponding one of the 64 directors to receive the message. For example, referring to FIG. 11C, each of the bit positions 1-64 represents directors 180^18032, 2001-20θ31, respectively. Here, in this example, because a logic 1 is only in bit position 1, the eight-byte vector indicates that the destination director is only front-end director 108^ In the example in FIG. 1 ID, because a logic 1 is only in bit position 2, the eight-byte vector indicates that the destination director is only front-end director IO82. In the example in FIG. 1 IE, because a logic 1 is more than one bit position, the destination for the message is to more than one director, i.e., a multi-cast message. In the example in FIG. 1 IE, a logic 1 is only in bit positions 2, 3, 63 and 64. Thus, the eight-byte vector indicates that the destination directors are only front-end director IO82 and I8O3 and back- end directors 2003! and 2OO32. There is a mask vector stored in a register of register section 420 (FIG. 7) in the message engine (ME) 315 which identifies director or directors which may be not available to use (e.g. a defective director or a director not in the system at that time), Step 524, 525, for a uni-cast transmission). If the message engine (ME) 315 state machine 410 indicates that the director is available by examining the transmit vector mask (FIG. 1 IF) stored in register 420, the message engine (ME) 315 encapsulates the message payload with a MAC header and CRC inside the packetizer portion 428P, discussed above (Step 526). An example of the mask is shown in FIG. 1 IF. The mask has 64 bit positions, one for each one of the directors. Thus, as with the destination vectors described above in connection with FIGS. 1 lC-1 IE, bit positions 1- 64 represents directors 1801-18032, 200i-20θ32, respectively. Here in this example, a logic 1 in a bit position in the mask indicates that the representative director is available and a logic 0 in such bit position indicates that the representative director is not available. Here, in the example shown in FIG. 1 IF, only director 2OO32 is unavailable. Thus, if the message has a destination vector as indicated in FIG. 1 IE, the destination vector, after passing through the mask of FIG. 1 IF modifies the destination vector to that shown in FIG. 11G. Thus, director 2OO32 will not receive the message. Such mask modification to the destination vector is important because, as will be described, the messages on a multi-cast are sent sequentially and not in parallel. Thus, elimination of message transmission to an unavailable director or directors increases the message transmission efficiency of the system.
Having packetized the message into a MAC packet via the packetizer portion of the packetizer/de-packetizer 428 (FIG. 7), the message engine (ME) 315 transfers the MAC packet to the crossbar switch 320 (Step 528) and the MAC packet is routed to the destination by the message network 260 (Step 530) via message network boards 304l5 3042 or on the same director board via the crossbar switch 320 on such director board. Referring to FIG. 12, the message read operation is described. Thus, in Step 600 the director waits for a message. When a message is received, the message engine (ME) 315 state machine 410 receives the packet (Step 602). The state machine 410 checks the receive bit vector mask (FIG. 11 stored in register 426) against the source address of the packet (Step 604). If the state machine 410 determines that the message is from an improper source (i.e., a faulty director as indicated in the mask, FIG. 1 IF, for example), the packet is discarded (Step 606). On the other hand, if the state machine 410 determines that the packet is from a proper or valid director (i.e., source), the message engine (ME) 315 de-encapsulates the message from the packet (Step 608) in de-packetizer 428D. The state machine 410 in the message engine (ME) 315 initiates a 32- byte payload transfer via the DMA receive operation (Step 610). The DMA writes the 32 byte message to the memory receive queue 312R in the RAM 312 (Step 612). The message engine (ME) 315 state machine 410 then increments the receive write pointer register 450 (Step 614). The CPU 310 then checks whether the receive write pointer 450 is equal to the receive read pointer 452 (Step 616). If they are equal, such condition indicates to the CPU 310 that a message has not been received (Step 618). On the other hand, if the receive write pointer 450 and the receive read pointer 452 are not equal, such condition indicates to the CPU 310 that a message has been received and the CPU 310 processes the message in the receive queue 314R of RAM 312 and then the CPU 310 increments the receive read pointer and writes it into the receive read pointer register 452. Thus, messages are stored in the receive queue 312R of RAM 312 until the contents of the receive read pointer 452 and the contents of the receive write pointer 450, which are initialized to zero during power-up, are equal. Referring now to FIG. 13, the acknowledgement of a message operation is described. In Step 700 the receive DMA engine 420 successfully completes a message transfer to the receive queue in RAM 312 (FIG. 7). The state machine 410 in the message engine (ME) 315 generates an acknowledgement MAC packet and transmits the MAC packet to the sending director via the message network 260 (FIG. 2) (Steps 702, 704). The message engine (ME) 315 at the sending director de-encapsulates a 16 byte status payload in the acknowledgement MAC packet and transfers such status payload via a receive DMA operation (Step 706). The DMA of the sending (i.e., source) director writes to a status field of the descriptor within the RAM memory send queue 314S (Step 708). The state machine 410 of the message engine (ME) 315 of the sending director (which received the acknowledgement message) increments its send read pointer 454 (Step 712). The CPU 310 of the sending director (which received the acknowledgement message) processes the descriptor status and removes the descriptor from the send queue 312S of RAM 312 (Step 714). It should be noted that the send and receive queues 312S and 312R are each circular queues.
As noted above, the MAC packets are, during normal operation, transmitted alternatively to one of the pair of message network boards 304l5 3042 by hardware a selector S in the crossbar switch 320. The selector S is responsive to the bit B in the header of the MAC packet (FIG. 2B) and, when such bit B is one logic state the data is coupled to one of the message networks boards 402A and in response to the opposite logic state the data is coupled to the other one of the message networks boards 402B. That is, when one message is transmitted to board 304i the next message is transmitted to board 3042.
Referring again to FIG. 9, the details of an exemplary transmit DMA 418 is shown. As noted above, after a descriptor has been created by the CPU 310 (FIG. 7) and is then stored in the RAM 312. If the send write pointer 450 (FIG. 7) and send read pointer 452, described above, have different counts an indication is provided by the state machine 410 in the message engine (ME) 315 (FIG. 7) that the created descriptor is available for DMA transmission to the message engine (ME) 315, the payload off the descriptor is packetized into a MAC packet and sent through the message network 360 (FIG. 2) to one or more directors I8O1-I8O32, 20θ!-20θ32. More particularly, the descriptor created by the CPU 310 is first stored in the local cache memory 319 and is later transferred to the send queue 312S in RAM 312. When the send write pointer 450 and send read pointer 452 have different counts, the message engine (ME) 315 state machine 410 initiates a DMA transmission as discussed above in connection with Step 520 (FIG. 11). Further, as noted above, the descriptor resides in send queues 312R within the RAM 312. Further, as noted above, each descriptor which contains the message is a fixed size, here 64- bytes. As each new, non-transmitted descriptor is created by the CPU 310, it is sequentially stored in a sequential location, or address in the send queue 312S. Here, the address is a 32-bit address. When the transmit DMA is initiated, the state machine 410 in the message engine (ME) 315 (FIG. 7), sends the queue address on bus 411 to an address register 413 in the DMA transmitter 418 (FIG. 9) along with a transmit write enable signal Tx_WE signal. The DMA transmitter 418 requests the CPU bus 317 by asserting a signal on Xmit_Br. The CPU bus arbiter 414 (FIG. 7) performs a bus arbitration and when appropriate the arbiter 414 grants the DMA transmitter 418 access to the CPU bus 317. The Xmit Cpu state machine 419 then places the address currently available in the address register 413 on the Address bus portion 317 A of CPU bus 317 by loading the output address register 403. Odd parity is generated by a Parity generator 405 before loading the output address register 403. The address in register 403 is placed on the CPU bus 317 (FIG. 7) for RAM 312 send queue 312S, along with appropriate read control signals via CPU bus 317 portion 317C. The data at the address from the RAM 312 passes, via the data bus portion 317D of CPU bus 317, through a parity checker 415 to a data input register 417. The control signals from the CPU 10 are fed to a Xmit CPU state machine 419 via CPU bus 317 bus portion 317C. One of the control signals indicates whether the most recent copy of the requested descriptor is in the send queue 312S of the RAM 312 or still resident in the local cache memory 319. That is, the most recent descriptor at any given address is first formed by the CPU 310 in the local cache memory 319 and is later transferred by the CPU 310 to the queue in the RAM 312. Thus, there may be two descriptors with the same address; one in the RAM 312 and one in the local cache memory 319 (FIG. 7), the most recent one being in the local cache memory 319. In either event, the transmit DMA 418 must obtain the descriptor for DMA transmission from the RAM 312 and this descriptor is stored in the transmit buffer register 421 using signal 402 produced by the state machine 419 to load these registers 421. The control signal from the CPU 310 to the Xmit CPU state machine 419 indicates whether the most recent descriptor is in the local cache memory 319. If the most recent descriptor is in the local cache memory 319, the Xmit CPU state machine 419 inhibits the data that was just read from send queue 312S in the RAM 312 and which has been stored in register 421 from passing to selector 423. In such case, state machine 419 must perform another data transfer at the same address location. The most recent message is then transferred by the CPU 310 from the local cache memory 319 to the send queue 312S in the RAM 312. The transmit message state machine 419 then re-arbitrates for the CPU bus 317 and after it is granted such CPU bus 317, the Xmit CPU state machine 419 then reads the descriptor from the RAM 312. This time, however, there the most recent descriptor is available in the send queue 312s in the RAM 312. The descriptor in the RAM 312 is now loaded into the transmit buffer register 421 in response to the assertion of the signal 402 by the Xmit CPU state machine 419. The descriptor in the register 421 is then transferred through selector 423 to message bus interface 409 under the control of a Xmit message (msg) state machine 427. That is, the descriptor in the transmit buffer register 421 is transferred to the transmit data buffer 422 (FIG. 7) over the 32 bit transmit message bus interface 409 by the Xmit message (msg) state machine 427. The data in the transmit data buffer 422 (FIG. 7) is packetized by the packetizer section of the packetizer/de-packetizer 428 as described in Step 530 in FIG. 11. More particularly, and referring also to FIG. 14 A, the method of operating the transmit DMA 418 (FIG. 9) is shown. As noted above, each descriptor is 64-byte. Here, the transfer of the descriptor takes place over two interfaces namely, the CPU bus 317 and the transmit message interface bus 409 (FIG. 7). The CPU bus 317 is 64 bits wide and eight, 64-bit double-words constitute a 64-byte descriptor. The Xmit CPU state machine 419 generates the control signals which result in the transfer of the descriptor from the RAM 312 into the transmit buffer register 421 (FIG. 7). The 64-byte descriptor is transferred in two 32-byte burst accesses on the CPU bus 317. Each one of the eight double words is stored sequentially in the transmit buffer register 421 (FIG. 9). Thus, in Step 800, the message engine 315 state machine 410 loads the transmit DMA address register 413 with the address of the descriptor to be transmitted in the send queue 312S in RAM 312. This is done by the asserting the Tx_WE signal and this puts Xmit CPU state machine 419 in step 800, loads the address register 413 and proceeds to step 802. In step 802, The Xmit Cpu state machine 419 loads the CPU transfer counter 431 (FIG. 9) with a 32-byte count, which is 2. This is the number of 32 byte transfers that would be required to transfer the 64-byte descriptor, Step 802. The Xmit Cpu state machine 419 now proceeds to Step 804. In step 804, the transmit DMA state machine 419 checks the validity of the address that is loaded into its address register 413. The address loaded into the address register 413 is checked against the values loaded into the memory address registers 435. The memory address registers 435 contain the base address and the offset of the send queue 312s in the RAM 312. The sum of the base address and the offset is the range of addresses for the send queue 312S in RAM 312. The address check circuitry 437 constantly checks whether the address in the address register 413 is with in the range of the send queue 312S in the RAM 312. If the address is found to be outside the range of the send queue 312S the transfer is aborted, this status is stored in the status register 404 and then passed back to the message engine 315 state machine 410 in Step 416. The check for valid addresses is done in Step 805. If the address is within the range, i.e., valid, the transmit DMA state machine 419 proceeds with the transfer and proceeds to Step 806. In the step 806, the transmit DMA state machine 419 requests the CPU bus 317 by asserting the Xmit_BR signal to the arbiter 414 and then proceeds to Step 807. In Step 807, the Xmit Cpu state machine 419 constantly checks if it has been granted the bus by the arbiter. When the CPU bus 317 is granted, the Xmit CPU state machine proceeds to Step 808. In Step 808, the Xmit Cpu state machine 419 generates an address and a data cycle which essentially reads 32-bytes of the descriptor from the send queue 312S in the RAM 312 into its transmit buffer register 421. The Xmit Cpu state machine 419 now proceeds to step 810. In Step 810, the Xmit Cpu state machine 419 loads the descriptor that was read into its buffer registers 421 and proceeds to Step 811. In Step 811, a check is made for any local cache memory 319 coherency errors (i.e., checks whether the most recent data is in the cache memory 319 and not in the RAM 312) on these 32-bytes of data. If this data is detected to be resident in the local CPU cache memory 319, then the Xmit Cpu state machine 419 discards this data and proceeds to Step 806. The Xmit Cpu state machine 419 now requests for the CPU bus 317 again and when granted, transfers another 32-bytes of data into the transmit buffer register 421, by which time the CPU has already transferred the latest copy of the descriptor into the RAM 312. In cases when the 32-bytes of the descriptor initially fetched from the RAM 312 was not resident in the local CPU cache memory 319 (i.e., if no cache coherency errors were detected), the Xmit Cpu state machine 419 proceeds to Step 812. In Step 812, the Xmit CPU state machine 419 decrements counters 431 and increments the address register 413 so that such address register 413 points to the next address. The Xmit Cpu state machine then proceeds to step 814. When in Step 814, the Transmit CPU state machine 419 checks to see if the transfer counter 431 has expired, i.e., counted to zero, if the count was found to be non-zero, it then, proceeds to Step 804 to start the transfer of the next 32-bytes of the descriptor. In case the counter 431 is zero, the process goes to Step 816 to complete the transfer. The successful transfer of the second 32-bytes of descriptor from the RAM 312 into the transmit DMA buffer register 421 completes the transfer over the CPU bus 317. The message interface 409 is 32 bits wide and sixteen, 32 bit words constitute a 64-byte descriptor. The 64-byte descriptor is transferred in batches of 32 bytes each. The Xmit msg state machine 427 controls and manages the interface 409. The Xmit Cpu state machine asserts the signal 433 to indicate that the first 32 bytes have been successfully transferred over the CPU bus 317 (Step 818, FIG. 14B), this puts the Xmit msg state machine into Step 818 and starts the transfer on the message interface. In step 820, the Xmit msg machine 427 resets burst/transfer counters 439 and initiates the transfer over the message interface 409. In Step 820, the transfer is initiated over the message interface 409 by asserting the "transfer valid" (TX_DATA_Vaild) signal indicating to the message engine 315 state machine 410 that valid data is available on the bus 409. The transmit msg machine 427 transfers 32 bits of data on every subsequent clock until its burst counter in burst/transfer counter 439 reaches a value equal to eight, Step 822. The burst counter in burst/transfer counter 439 is incremented with each 32-bit word put on the message bus 409 by a signal on line 433. When the burst count is eight, a check is made by the state machine 427 as to whether the transmit counter 431 has expired, i.e., is zero, Step 824. The expiry of the transfer counter in burst/transfer counter 439 indicates the 64 byte descriptor has been transferred to the transmit buffer 422 in message engine 315. If it has expired, the transmit message state machine 427 proceeds to Step 826. In step 826, the Xmit msg state machine asserts the output End of Transfer (Tx_EOT) indicating the end of transfer over the message bus 409 .In this state, after the assertion of the Tx_EOT signal the status of the transfer captured in the status register 404 is sent to the message engine 315 state machine 410. The DMA operation is complete with the descriptor being stored in the transmit buffer 422 (FIG. 7). On the other hand, if the transfer counter in burst/transfer counter 439 has not expired, the process goes to Step 800 and repeats the above described procedure to transfer the 2nd 32 bytes of descriptor data, at which time the transfer will be complete.
Referring now to FIG. 10, the receive DMA 420 is shown. Here, a message received from another director is to be written into the RAM 312 (FIG. 7). The receive DMA 420 is adapted to handle three types of information: error information which is 8 bytes in size; acknowledgement information which is 16 bytes in size; and receive message payload and/or fabric management information which is 32 byes in size. Referring also to FIG. 7, the message engine 315 state machine 410 asserts the Rx_WE signal, indicating to the Receive DMA 420 that it is ready transfer the Data in its Rec buffer 426 FIG. 7. The data in the Receive buffer could be the 8-byte error information, the 16-byte Acknowledgment information or the 32-byte Fabric management/ Receive message payload information. It places a 2 bit encoded receive transfer count, on the Rx_transfer count signal indicating the type of information and an address which is the address where this information is to be stored in the receive queue of RAM 312. In response to the receive write enable signal Rx_WE, the Receive message machine 450 (FIG. 10) loads the address into the address register 452 and the transfer count indicating the type of information, into the receive transfer counter 454. The address loaded into the address register 452 is checked by the address check circuitry 456 to see if it is with in the range of the Receive queue addresses, in the RAM 312. This is done by checking the address against the values loaded into the memory registers 457 (i.e., a base address register and an offset register therein). The base address register contains the start address of the receive queue 312R residing in the RAM 312 and the offset register contains the size of this receive queue 312R in RAM 312. Therefore the additive sum of, the values stored in the base address register and the offset register specifies the range of addresses of the receive queue in the RAM 312R. The memory registers 457 are loaded during initialization. On the subsequent clock after the assertion of the Rx_WE signal, the message engine 315 state machine 410 the proceeds to place the data on a 32-bit message engine 315 data bus 407, FIG. 10. A Rx_data_valid signal accompanies each 32 bits of data, indicating that the data on the message engine data bus 407 is valid. In response to this Rx_data_valid signal the receive message state machine 450 loads the data on the data bus into the receive buffer register 460. The end of the transfer over the message engine data bus 407d is indicated by the assertion of the Rx_EOT signal at which time the Receive message state machine 450 loads the last 32 bits of data on the message engine data bus 407D of bus 407, into the receive buffer registers 460. This signals the end of the transfer over the message engine data bus 407D portion of bus 407. At the end of such transfer is conveyed to the Rx_Cpu state machine 462 by the assertion of the signal 464. The Receive CPU machine 462 now, requests for the CPU bus 317 by asserting the signal REC_Br. After an arbitration by CPU bus arbiter 414 (FIG. 7) the receive DMA 420 (FIG. 10) is given access to the CPU bus 317. The Receive CPU state machine 462 proceeds to transfer the data in its buffer registers 424 over the CPU bus 317 into the Receive queue 312R in the RAM 312. Simultaneously, this data is also transferred into a duplicate buffer register 466. The data at the output of the receive buffer register 460 passes to one input of a selector 470 and also passes to a duplicate data receive buffer register 460. The output of the duplicate receive buffer register 466 is fed to a second input of the selector 470. As the data is being transferred by the Receive CPU state machine 462, it is also checked for cache coherency errors. If the data corresponding to the address being written into the RAM 312, is located in the CPU's local cache memory 319 (FIG. 7), the receive DMA machine 420 waits for the CPU 310 to copy the old data in its local cache memory 319 back to the receive queue 312R in the RAM 312 and then overwrites this old data with a copy of the new data from the duplicate buffer register 466.
More particularly, if central processing unit 310 indicates to the DMA receiver 420 that the data the receive buffer register 460 is available in the local cache memory 319, the receive CPU state machine 462 produces a select signal on line 463 which couples the data in the duplicate buffer register 466 to the output of selector 470 and then to the bus 317 for store in the random access memory 312. The successful write into the RAM 312 completes the DMA transfer. The receive DMA 420 then signals the message engine 315 state machine 410 on the status of the transfer. The status of the transfer is captured in the status register 459.
Thus, with both the receive DMA and the transmit DMA, there is a checking of the local cache memory 319 to determine whether it has "old" data, in the case of the receive DMA or whether it has "new data" in the case of the transmit DMA.
Referring now to FIG. 15 A, the operation of the receive DMA 420 is shown. Thus, in Step 830 the Receive message machine 450 checks if the write enable signal Rx_WE is asserted. If found asserted, the receive DMA 420 proceeds to load the address register 452 and the transfer counter 454. The value loaded into the transfer counter 454 determines the type of DMA transfer requested by the Message engine state machme 310 in FIG 7. The assertion of the Rx_WE signal starts the DMA receive transfer operation. This puts the Rx msg state machine 450 in Step 832. In Step 832 the Rec msg state machine 450 loads the address register 452, the transfer counter 454 and then proceeds to Step 834. In Step 834, it checks to see if the Rx_DATA_VALID signal is asserted. If asserted it proceeds to step 836. The Rx msg state machine loads the buffer register 460 (FIG. 10) in Step 836 with the data on the message engine data bus 407D of bus 407 fig 10. The Rx_DATA_VALID signal accompanies each piece of data put on the bus 407. The data is sequentially loaded into the buffer registers 460 (FIG.10). The End of the transfer on the message engine data bus 407D of bus 407 is indicated by the assertion of the Rx_EOT signal. When the Receive message state machine 450 is in the End of transfer state Step 840 it signals the Receive CPU state machine 462 and this starts the transfer on the CPU bus 317 side.
The flow for the Receive CPU state machine is explained below. Thus, referring to FIG. 15B, the End of the transfer on the Message engine data bus 407D portion of bus 407 starts the Receive CPU state machine 462 and puts it in Step 842. The Receive CPU state machine 462 checks for validity of the address in this state (Step 844). This is done by the address check circuitry 456. If the address loaded in the address register 452 is outside the range of the receive queue 312R in the RAM 312, the transfer is aborted and the status is captured in the Receive status register 459 and the Rec Cpu state machine 462 proceeds to Step 845. On a valid address the Receive CPU state machine 462 goes to Step 846. In Step 846 the Receive Cpu state machine 462 requests for access of the CPU bus 317. It then proceeds to Step 848. In step 848 it checks for a grant on the bus 317. On a qualified grant it proceeds to step 850. In Step 850, The Rec Cpu state machine 462 performs an address and a data cycle, which essentially writes the data in the buffer registers 460 into the receive queue 312R in RAM 312. Simultaneously with the write to the RAM 312, the data put on the CPU bus 317 is also loaded into the duplicate buffer register 466. At same time, the CPU 310 also indicates on one of the control lines, if the data corresponding to the address written to in the RAM 312 is available in its local cache memory 319. At the end of the address and data cycle the Rec Cpu state machine 462 proceeds to Step 850. In this step it checks for cache coherency errors of the type described above in connection with the transmit DMA 418 (FIG. 9). If cache coherency error is detected and the receive CPU state machine 462 proceeds to Step 846 and retries the transaction more particularly, the Receive CPU state machine 462 now generates another address and data cycle to the previous address and this time the data from the duplicate buffer 466 is put on to the CPU data bus 317. If there were no cache coherency errors the Receive CPU state machine 462 proceeds to Step 852 where it decrements the transfer counter 454 and increment the address in the address register 452. The Receive Cpu state machine 462 then proceeds to Step 854. In Step 854, the state machine 462 checks if the transfer counter has expired, i.e., is zero. On a non zero transfer count the receive Cpu state machine 462 proceeds to Step 844 and repeats the above described procedure until the transfer becomes zero. A zero transfer count when in step 854 completes the write into the receive queue 312R in RAM 312 and the Rec Cpu state machine proceeds to 845. In step 845, it conveys status stored in the status register back to status is conveyed to the message engine 315 state machine 410. Referring again to FIG. 7, the interrupt control status register 412 will be described in more detail. As described above, a packet is sent by the pocketsize portion of the packetizer/de- packetizer 428 to the crossbar switch 320 for transmission to one or more of the directors. It is to be noted that the packet sent by the packetizer portion of the packetizer/de-packetizer 428 passes through a parity generator PG in the message engine 315 prior to passing to the crossbar switch 320. When such packet is sent by the message engine 315 in exemplary director 180ι, to the crossbar switch 320, a parity bit is added to the packet by parity bit generator PG prior to passing to the crossbar switch 320. The parity of the packet is checked in the parity checker portion of a parity checker/generator (PG/C) in the crossbar switch 320. The result of the check is sent by the PG/C in the crossbar switch 320 to the interrupt control status register 412 in the director 180ι. Likewise, when a packet is transmitted from the crossbar switch 320 to the message engine 315 of exemplary director 180ι, the packet passes through a parity generator portion of the parity checker/generator (PG/C) in the crossbar switch 320 prior to being transmitted to the message engine 315 in director 180ι. The parity of the packet is then checked in the parity checker portion of the parity checker (PC) in director 180ι and is the result (i.e., status) is transmitted to the status register 412.
A number of embodiments of the invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. Accordingly, other embodiments are within the scope of the following claims.
WHAT IS CLAIMED IS:

Claims

1 1. A data storage system for transferring data between a host computer/server and a bank of disk
2 drives through a system interface, such system interface comprising:
3 a plurality of first directors coupled to the host computer/server;
4 a plurality of second directors coupled to the bank of disk drives;
5 a cache memory; and
6 wherein there are separate point-to-point data paths between each one of the directors and
7 the global cache memory.
1 2. The system recited in claim 1 including a backplane and wherein the cache memory and the
2 directors are interconnected through the backplane.
l 3. The system recited in claim 2 wherein the backplane is a printed circuit board.
1 4. The system recited in claim 1 further comprising:
2 a data transfer section coupled to the plurality of first directors, the second directors, and
3 the cache memory; and
4 a messaging network coupled to the plurality of first directors and the plurality of second
5 directors, such first and second directors controlling data transfer between the host computer and
6 the bank of disk drives in response to messages passing between the directors through the
7 messaging network as such data passes through the memory via the data transfer section.
1 5. The system recited in claim 4 including a backplane and wherein the cache memory and the
2 directors are interconnected through the backplane,
l 6. The system recited in claim 5 wherein the backplane is a printed circuit board.
1 7. A method for transferring data between a host computer/server and a bank of disk drives
2 through a system interface, such system interface comprising: a plurality of first directors
3 coupled to the host computer/server; a plurality of second directors coupled to the bank of disk
4 drives; a data transfer section coupled to the plurality of first directors and second directors; and
5 a messaging network coupled to the plurality of first directors and the plurality of second
6 directors, such first and second directors controlling data transfer between the host computer and 7 the bank of disk drives in response to messages passing between the directors through the
8 messaging network as such data passes through the data transfer section, such method comprising:
9 preparing in a transmitting one of the directors, a message to be sent to a receiving one, or 0 ones, of the directors; 1 building in such transmitting one of the directors, a descriptor, such descriptor comprising a 2 command field indicating the receiving one, or ones, of directors to receive the built message; 3 encapsulating the message payload of the descriptor into a packet, such packet comprising: 4 a header, such header including: s a source portion indicating the transmitting one of the directors; 6 a destination portion indicating the receiving one, or ones, of the directors; and 7 the message payload; s transmitting such packet to said receiving one, or ones, of the directors through the 9 messaging network; 0 decoding the destination portion of the packet to route such packet to such receiving one, or i ones of the directors.
1 8. The method recited in claim 7 including:
2 receiving in one of the receiving one or ones of the directors the transmitted packet;
3 determining in such receiving one, or ones, the receiving directors whether the received
4 packet is from a proper, or an improper transmitting one of the directors;
5 de-encapsulating in such receiving one, or ones, of the directors the received packet after
6 determining in such receiving one, or ones, of the directors that the packet is from a proper
7 transmitting one of the directors.
1 9. The method recited in claim 8 including having the receiving, one or ones, of the directors
2 send an acknowledge receipt of the packet to said transmitting one of the transmitting such
3 packet.
i
10. The method of claim 7 wherein in building in the descriptor in such transmitting one of the
2 directors, the descriptor further comprises a message payload indicating an address in the bank of
3 disk drives having the requested data.
1 11. A method for transferring data between a host computer/server and a bank of through a
2 system interface, such system interface comprising: a plurality of first directors coupled to the
3 host computer/server; a plurality of second directors coupled to the bank of disk drives; a data
4 transfer section coupled to the plurality of first directors and second directors; and a messaging
5 network coupled to the plurality of first directors and the plurality of second directors, such first
6 and second directors controlling data transfer between the host computer and the bank of disk
7 drives in response to messages passing between the directors through the messaging network as
8 such data passes through the data transfer section, such method comprising:
9 determining, in a transmitting one of the directors, that action is requested by a receiving one,
10 or ones, of the directors; π preparing in such transmitting one of the directors, in response to such determination, a
12 message to be sent to the receiving one, or ones, of the directors;
13 building in such transmitting one of the directors, a descriptor, such descriptor comprising a
14 command field indicating receiving one, or ones, of the directors to receive the built message; is storing the built descriptor in a memory within such transmitting one of the directors;
16 incrementing a pointer or counter in the transmitting one of the directors each time a
17 descriptor is ready to be retrieved from the memory in such transmitting one of the directors; is retrieving such stored descriptor from the memory in such transmitting one of the directors;
19 encapsulating the message payload of the retrieved descriptor into a MAC type packet, such 0 packet comprising: i a header, such header including: 2 a source portion indicating the one of the transmitting directors; 3 a destination portion indicating the another one, or ones of the receiving directors; 4 and 5 the message payload; 6 transmitting such packet to said receiving one, or ones, of the directors through the 7 messaging network; 8 decoding the destination portion of the packet to route such packet to such receiving one, or 9 ones, of the directors.
1 12. The method recited in claim 11 including:
2 receiving in one of the receiving one, or ones, of the directors the transmitted packet; determining in such one of the receiving one, or ones, of the directors whether the received packet is from a proper, or an improper transmitting one of the directors; de-encapsulating in such one of the receiving one, or ones, of the directors the received packet after determining in such one of the receiving one, or ones, of the directors that the packet is from a proper transmitting one of the directors.
13. The method recited in claim 12 including having the receiving one, or ones, of the directors send an acknowledge receipt of the packet to said transmitting one of the directors transmitting such packet.
14. The method recited in claim 13 including incrementing in said transmitting one of the directors pointer or counter as a consequence of receiving such acknowledge receipt.
15. The method recited in claim 11 wherein the descriptor includes a message payload indicating an address in the bank of disk drives having the requested data.
16. The method recited in claim 15 including: receiving in one of the receiving one, or ones, of the directors the transmitted packet; determining in such one of the receiving one, or ones, of the directors whether the received packet is from a proper, or an improper transmitting one of the directors; de-encapsulating in such one of the receiving one, or ones, of the directors the received packet after determining in such one of the receiving one, or ones, of the directors that the packet is from a proper transmitting one of the directors.
17. The method recited in claim 16 including having the receiving one, or ones, of the directors send an acknowledge receipt of the packet to said transmitting one of the directors transmitting such packet.
18. The method recited in claim 17 including incrementing in said transmitting one of the directors a pointer or counter as a consequence of receiving such acknowledge receipt.
19. A method for transferring data between a host computer/server and a bank of through a system interface, such system interface comprising: a plurality of first directors coupled to the 3 host computer/server; a plurality of second directors coupled to the bank of disk drives; a data
4 transfer section coupled to the plurality of first directors and the second directors; and a
5 messaging network coupled to the plurality of first directors and the plurality of second directors,
6 such first and second directors controlling data transfer between the host computer and the bank of
7 disk drives in response to messages passing between the directors through the messaging network as
8 such data passes through the data transfer section, such method comprising: determining, in a transmitting one of the directors, that data requested for transfer by such o transmitting one of the directors is unavailable in the cache memory; i preparing in such transmitting one of the directors, in response to such determination, a 2 message to be sent to a receiving one, or ones, of the directors; 3 building in such transmitting one of the transmitting directors, a descriptor, such descriptor 4 comprising: s a message payload indicating an address in the bank of disk drives having the requested 6 data; and 7 a command field indicating receiving one, or ones, of the directors to receive the built s message; storing the built descriptor in a memory within such transmitting one of the directors; 0 incrementing a pointer or counter in the transmitting one of the directors each time a i descriptor is ready to be retrieved from the memory in such transmitting one of the directors; 2 retrieving such stored descriptor from the memory in such transmitting one of the directors; 3 decrementing the pointer or counter when the descriptor is retrieved from the memory in 4 such transmitting one of the directors; 5 encapsulating the message payload of the retrieved descriptor into a MAC type packet, such 6 packet comprising: 7 a header, such header including: 8 a source portion indicating the one of the transmitting directors; 9 a destination portion indicating the another one, or ones of the receiving directors; 0 and i the message payload; 2 transmitting such packet to said receiving one, or ones, of the directors through the 3 messaging network; decoding the destination portion of the packet to route such packet to such receiving one, or 5 ones, of the directors.
20. The method recited in claim 19 including: receiving in one of the receiving one, or ones, of the directors the transmitted packet; determining in such one of the receiving one, or ones, of the directors whether the received packet is from a proper, or an improper transmitting one of the directors; de-encapsulating in such one of the receiving one, or ones, of the directors the received packet after determining in such one of the receiving one, or ones, of the directors that the packet is from a proper transmitting one of the directors.
21. The method recited in claim 20 including having the receiving one, or ones, of the directors send an acknowledge receipt of the packet to said transmitting one of the directors transmitting such packet.
22. The method recited in claim 21 including incrementing in said transmitting one of the directors a pointer or counter as a consequence of receiveing such acknowledge receipt.
23. The method recited in claim 19 wherein the descriptor includes a message payload indicating an address in the bank of disk drives having the requested data.
24. The method recited in claim 23 including: receiving in one of the receiving one, or ones, of the directors the transmitted packet; determining in such one of the receiving one, or ones, of the directors whether the received packet is from a proper, or an improper transmitting one of the directors; de-encapsulating in such one of the receiving one, or ones, of the directors the received packet after determining in such one of the receiving one, or ones, of the directors that the packet is from a proper transmitting one of the directors.
25. The method recited in claim 24 including having the receiving one, or ones, of the directors send an acknowledge receipt of the packet to said transmitting one of the directors transmitting such packet.
26. The method recited in claim 25 including incrementing in said transmitting one of the directors a pointer or counter as a consequence of receiving such acknowledge receipt.
27. The method recited in claim 7 wherein the messages pass between the directors through the messaging network as said data passes through the cache memory via the data transfer section.
PCT/US2001/041175 2000-06-29 2001-06-27 Data storage system having point-to-point configuration WO2002003387A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
GB0207148A GB2374697A (en) 2000-06-29 2001-06-27 Data storage system having point-to-point configuration
JP2002507375A JP4885406B2 (en) 2000-06-29 2001-06-27 Data storage system with point-to-point configuration

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US09/606,730 2000-06-29
US09/606,730 US7007194B1 (en) 2000-06-29 2000-06-29 Data storage system having point-to-point configuration
US09/606,734 US7073020B1 (en) 1999-01-04 2000-06-29 Method for message transfer in computer storage system
US09/606,734 2000-06-29

Publications (2)

Publication Number Publication Date
WO2002003387A2 true WO2002003387A2 (en) 2002-01-10
WO2002003387A3 WO2002003387A3 (en) 2003-06-05

Family

ID=27085321

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2001/041175 WO2002003387A2 (en) 2000-06-29 2001-06-27 Data storage system having point-to-point configuration

Country Status (3)

Country Link
JP (1) JP4885406B2 (en)
GB (1) GB2374697A (en)
WO (1) WO2002003387A2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2411987A (en) * 2004-03-09 2005-09-14 Hewlett Packard Development Co Suppressing production of bus transactions
WO2006113087A2 (en) * 2005-04-13 2006-10-26 Emc Corporation Data storage system having memory controller with embedded cpu
EP3716219A1 (en) * 2004-03-23 2020-09-30 Google LLC Generating and serving tiles in a digital mapping system

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2560560C (en) * 2004-03-23 2011-08-02 Google Inc. Generating and serving tiles in a digital mapping system
US8010407B1 (en) 2006-11-14 2011-08-30 Google Inc. Business finder for locating local businesses to contact

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5689728A (en) * 1988-08-30 1997-11-18 Hitachi, Ltd. Controller of paths for data transfer in system having external memory device
US5805821A (en) * 1994-09-08 1998-09-08 International Business Machines Corporation Video optimized media streamer user interface employing non-blocking switching to achieve isochronous data transfers
US5819054A (en) * 1993-06-30 1998-10-06 Hitachi, Ltd. Storage system realizing scalability and fault tolerance
US6061274A (en) * 1999-01-04 2000-05-09 Emc Corporation Methods and apparatus for message transfer in computer storage system

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4484262A (en) * 1979-01-09 1984-11-20 Sullivan Herbert W Shared memory computer method and apparatus
JPS59100964A (en) * 1982-12-01 1984-06-11 Hitachi Ltd Parallel transfer type director device
JP2872251B2 (en) * 1988-10-12 1999-03-17 株式会社日立製作所 Information processing system
JP3195489B2 (en) * 1994-03-03 2001-08-06 株式会社日立製作所 External storage control device and bus switching control method
JPH10333836A (en) * 1997-05-29 1998-12-18 Hitachi Ltd Disk array controller
JPH11177589A (en) * 1997-12-12 1999-07-02 Canon Inc Data transfer device, data processing method therefor and computer readable storage medium stored with program
JP3657428B2 (en) * 1998-04-27 2005-06-08 株式会社日立製作所 Storage controller
US6134624A (en) * 1998-06-08 2000-10-17 Storage Technology Corporation High bandwidth cache system
US7010575B1 (en) * 2000-03-31 2006-03-07 Emc Corporation Data storage system having separate data transfer section and message network having bus arbitration

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5689728A (en) * 1988-08-30 1997-11-18 Hitachi, Ltd. Controller of paths for data transfer in system having external memory device
US5819054A (en) * 1993-06-30 1998-10-06 Hitachi, Ltd. Storage system realizing scalability and fault tolerance
US5805821A (en) * 1994-09-08 1998-09-08 International Business Machines Corporation Video optimized media streamer user interface employing non-blocking switching to achieve isochronous data transfers
US6061274A (en) * 1999-01-04 2000-05-09 Emc Corporation Methods and apparatus for message transfer in computer storage system

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2411987A (en) * 2004-03-09 2005-09-14 Hewlett Packard Development Co Suppressing production of bus transactions
EP3716219A1 (en) * 2004-03-23 2020-09-30 Google LLC Generating and serving tiles in a digital mapping system
WO2006113087A2 (en) * 2005-04-13 2006-10-26 Emc Corporation Data storage system having memory controller with embedded cpu
WO2006113087A3 (en) * 2005-04-13 2006-12-14 Emc Corp Data storage system having memory controller with embedded cpu

Also Published As

Publication number Publication date
GB0207148D0 (en) 2002-05-08
GB2374697A (en) 2002-10-23
WO2002003387A3 (en) 2003-06-05
JP4885406B2 (en) 2012-02-29
JP2004502263A (en) 2004-01-22

Similar Documents

Publication Publication Date Title
US6779071B1 (en) Data storage system having separate data transfer section and message network with status register
US6651130B1 (en) Data storage system having separate data transfer section and message network with bus arbitration
US6611879B1 (en) Data storage system having separate data transfer section and message network with trace buffer
US7010575B1 (en) Data storage system having separate data transfer section and message network having bus arbitration
US7117275B1 (en) Data storage system having separate data transfer section and message network
US6684268B1 (en) Data storage system having separate data transfer section and message network having CPU bus selector
US6779150B1 (en) CRC error detection system and method
US6594712B1 (en) Inifiniband channel adapter for performing direct DMA between PCI bus and inifiniband link
US6636933B1 (en) Data storage system having crossbar switch with multi-staged routing
US6993621B1 (en) Data storage system having separate data transfer section and message network with plural directors on a common printed circuit board and redundant switching networks
US6763402B2 (en) Data storage subsystem
US6061274A (en) Methods and apparatus for message transfer in computer storage system
US6609164B1 (en) Data storage system having separate data transfer section and message network with data pipe DMA
US6816916B1 (en) Data storage system having multi-cast/unicast
US6631433B1 (en) Bus arbiter for a data storage system
US7073020B1 (en) Method for message transfer in computer storage system
US6742017B1 (en) Data storage system having separate data transfer section and message network with pointer or counters
JP2006190327A (en) Data storage system having separate data transfer section and message network with plural directors on common printed circuit board
US6901468B1 (en) Data storage system having separate data transfer section and message network having bus arbitration
US6581112B1 (en) Direct memory access (DMA) receiver
US6687797B1 (en) Arbitration system and method
US6889301B1 (en) Data storage system
US6584513B1 (en) Direct memory access (DMA) transmitter
US7007194B1 (en) Data storage system having point-to-point configuration
WO2002003387A2 (en) Data storage system having point-to-point configuration

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): DE GB JP

ENP Entry into the national phase in:

Ref country code: GB

Ref document number: 200207148

Kind code of ref document: A

Format of ref document f/p: F

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642