EP1256198A1 - Method and apparatus for transferring data between different network devices over an ip network - Google Patents

Method and apparatus for transferring data between different network devices over an ip network

Info

Publication number
EP1256198A1
EP1256198A1 EP00917881A EP00917881A EP1256198A1 EP 1256198 A1 EP1256198 A1 EP 1256198A1 EP 00917881 A EP00917881 A EP 00917881A EP 00917881 A EP00917881 A EP 00917881A EP 1256198 A1 EP1256198 A1 EP 1256198A1
Authority
EP
European Patent Office
Prior art keywords
packet
port interface
network
format
data packets
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP00917881A
Other languages
German (de)
French (fr)
Inventor
Aamer Latif
Rodney N. Mullendore
Joseph L. White
Brian Y. Uchino
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nishan Systems Inc
Original Assignee
Nishan Systems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/500,119 external-priority patent/US6400730B1/en
Application filed by Nishan Systems Inc filed Critical Nishan Systems Inc
Publication of EP1256198A1 publication Critical patent/EP1256198A1/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/06Answer-back mechanisms or circuits
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/08Protocols for interworking; Protocol conversion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/16Implementation or adaptation of Internet protocol [IP], of transmission control protocol [TCP] or of user datagram protocol [UDP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/16Implementation or adaptation of Internet protocol [IP], of transmission control protocol [TCP] or of user datagram protocol [UDP]
    • H04L69/161Implementation details of TCP/IP or UDP/IP stack architecture; Specification of modified or new header fields
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/16Implementation or adaptation of Internet protocol [IP], of transmission control protocol [TCP] or of user datagram protocol [UDP]
    • H04L69/169Special adaptations of TCP, UDP or IP for interworking of IP based networks with other networks 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04JMULTIPLEX COMMUNICATION
    • H04J3/00Time-division multiplex systems
    • H04J3/02Details
    • H04J3/06Synchronising arrangements
    • H04J3/0635Clock or time synchronisation in a network
    • H04J3/0685Clock or time synchronisation in a node; Intranode synchronisation
    • H04J3/0688Change of the master or reference, e.g. take-over or failure of the master
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/30Definitions, standards or architectural aspects of layered protocol stacks
    • H04L69/32Architecture of open systems interconnection [OSI] 7-layer type protocol stacks, e.g. the interfaces between the data link level and the physical level
    • H04L69/322Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions
    • H04L69/329Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions in the application layer [OSI layer 7]

Definitions

  • the present invention relates to transferring information between storage devices and a network via a switched, packetized communications system.
  • the present invention relates to methods and apparatus for receiving, translating, and routing data-packets between SCSI (Small Computer Systems Interface), Fibre Channel 0 and Ethernet devices in a flexible, programmable manner.
  • SCSI Small Computer Systems Interface
  • Fibre Channel 0 Fibre Channel 0
  • Ethernet devices in a flexible, programmable manner.
  • SANs Storage Area Networks
  • the current SAN paradigm assumes that the entire network is constructed using Fibre Channel switches. Therefore, most solutions involving SANs require implementation of separate networks: one to support the normal LAN and another to support the SAN.
  • New equipment and technology such as new equipment at the storage device level (Fibre Channel interfaces), the host/server level (Fibre Channel adapter cards) and the transport level (Fibre Channel hubs, switches and routers), into a mission-critical enterprise computing environment could be described as less than desirable for data center managers, as it involves replication of network infrastructure, new technologies (i.e., Fibre Channel), and new training for personnel.
  • Most companies have already invested significant amounts of money constructing and maintaining their network (e.g., based on Ethernet and/or ATM). Construction of a second high-speed network based on a different technology is a significant impediment to the proliferation of SANs. Therefore, a need exists for a method and apparatus that can alleviate problems with access to storage devices by multiple hosts, while retaining current equipment and network infrastructures, and minimizing the need for new training for data center personnel.
  • SCSI, Fibre Channel and Ethernet are protocols for data transfer, each of which uses a different individual format for data transfer.
  • SCSI commands were designed to be implemented over a parallel bus architecture and therefore are not packetized.
  • Fibre Channel like Ethernet, uses a serial interface with data transferred in packets-
  • the physical interface and frame formats between Fibre Channel and Ethernet are not compatible.
  • Gigabit Ethernet was designed to be compatible with existing Ethernet infrastructures and is therefore based on an Ethernet packet architecture. Because of these differences there is a need for new methods and apparatus to allow efficient communication between these protocols.
  • the present invention solves the above and other problems, thereby advancing the state of the useful arts, by providing methods and apparatus for transferring data between storage device interfaces and network interfaces.
  • IP Internet Protocol
  • the present invention brings sophisticated SAN capabilities to existing enterprise computing configurations, without the installation of costly Fibre Channel switches and hubs, by providing the means for Internet Protocol (IP) devices to transparently communicate with SCSI and Fibre Channel devices over an IP network.
  • IP Internet Protocol
  • the present invention accomplishes this through the use of Fibre Channel Protocol (FCP), an industry standard developed for implementation of SCSI commands over a Fibre Channel network.
  • FCP Fibre Channel Protocol
  • the invention allows the storage devices to retain the use of standard SCSI and Fibre Channel storage interfaces and construct a SAN using a company's existing network infrastructure. Therefore, no changes are required in host bus adapters (HBA) or storage devices (e.g. disk drives, tape drives, etc).
  • HBA host bus adapters
  • storage devices e.g. disk drives, tape drives, etc.
  • the device interfaces may be either SCSI, Fibre Channel or IP interfaces such as Gigabit Ethernet. Data is switched between SCSI and IP, Fibre Channel and IP, or between SCSI and Fibre Channel. Data can also be switched from SCSI to SCSI, Fibre Channel to Fibre Channel and IP to IP.
  • the port interfaces provide the conversion from the input frame format to an internal frame format, which can be routed within the apparatus.
  • the apparatus may include any number of total ports. The amount of processing performed by each port interface is dependent on the interface type.
  • the processing capabilities of the present invention permit rapid transfer of information packets between multiple interfaces at latency levels meeting the stringent requirements for storage protocols.
  • the configuration control can be applied to each port on a switch and, in turn, each switch on the network, via an SNMP or Web-based interface, providing a flexible, programmable control for the apparatus.
  • a method is provided for routing data packets in a switch device in a network such as a SAN.
  • the method typically comprises the steps of receiving a packet from a first network device at a first port interface of the switch device, wherein the packet is one of a SCSI formatted packet (i.e., SCSI formatted data stream converted into a packet), a Fibre Channel (FC) formatted packet and an Internet protocol (IP) formatted packet, wherein the first port interface is communicably coupled to the first network device, and converting the received packet into a packet having an internal format.
  • the method also typically includes the steps of routing the internal format packet to a second port interface of the switch device, reconverting the internal format packet to one of a SCSI formatted packet, an FC formatted packet or an IP formatted packet, and transmitting the reconverted packet to a second network device communicably coupled to the second port interface.
  • a network switch device typically comprises a first port interface including a means for receiving data packets from a network device, wherein the receiving means receives one of a SCSI formatted packet and a Fibre Channel (FC) formatted packet from a first network device, and a means for converting received packets into packets having an internal format, wherein the received data packet is converted into a first packet having the internal format.
  • the switch device also typically comprises a second port interface including a means for reconverting packets from the internal format to an IP format, wherein the first packet is converted into a packet having an IP format, and a means for transmitting IP packets to a network, wherein the IP formatted packet is transmitted to an IP network.
  • a means for routing the first packet to the second port interface is also provided.
  • a network switch device typically comprises a first port interface including a means for receiving data packets from an IP network, wherein the first interface means receives a packet in an IP format, and a means for converting received packets into packets having an internal format, wherein the received packet is converted into a first packet having an internal format.
  • the switch device also typically comprises a second port interface including a means for reconverting packets having the internal format to packets having the SCSI format, and a means for transmitting reconverted packets to a SCSI network device.
  • the switch device further typically includes a third port interface having a means for reconverting packets having the internal format to packets having the FC format, and a means for transmitting reconverted packets to a FC network device.
  • a means for routing packets between the first, second and third port interfaces is also typically provided. In operation, wherein if the first packet is routed to the second port interface, the first packet is converted to the SCSI format and transmitted to the SCSI network device, and wherein if the first packet is routed to the third port interface, the first packet is converted to the FC format and transmitted to the FC network device.
  • a network switch device for use in a storage area network (SAN).
  • the switch device typically comprises a first port interface communicably coupled to a SCSI device, wherein the first port interface converts SCSI formatted data packets received from the SCSI device into data packets having an internal format, and wherein the first port interface converts data packets having the internal format into SCSI formatted data packets.
  • the switch device also typically comprises a second port interface communicably coupled to a FC device, wherein the second port interface converts FC formatted data packets received from the FC device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format into FC formatted data packets.
  • the switch device further typically includes a third port interface communicably coupled to a IP device, wherein the third port interface converts IP formatted data packets received from the IP device into data packets having the internal format, and wherein the third port interface converts data packets having the internal format into IP formatted data packets, and a switch fabric for routing data packets having the internal format between the first, second and third port interfaces.
  • a third port interface communicably coupled to a IP device, wherein the third port interface converts IP formatted data packets received from the IP device into data packets having the internal format, and wherein the third port interface converts data packets having the internal format into IP formatted data packets, and a switch fabric for routing data packets having the internal format between the first, second and third port interfaces.
  • a network switch device for use in a storage area network (SAN) is provided.
  • the switch may comprise any combination of Fibre Channel, SCSI, Ethernet and Infiniband ports, and may comprise any number of total ports.
  • the switch device typically comprises a first port interface communicably coupled to one of a SCSI device(s), an FC device, or an IP device, a second port interface, wherein the second port interface is configurable to communicate with either a FC device or an Ethernet device, and a switch fabric for routing data packets having the internal format between the first and second port interfaces.
  • the second port interface when the second port interface is configured to communicate with a FC device, the second port interface converts FC formatted data packets received from the FC device into data packets having an internal format, and wherein the second port interface converts data packets having the internal format received from the switch fabric into FC formatted data packets, and wherein when the second port interface is configured to communicate with an Ethernet device, the second port interface converts Ethernet formatted data packets received from the Ethernet device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format received from the switch fabric into Ethernet formatted data packets.
  • the second port interface can be either self-configurable or user configurable.
  • Figure 1 illustrates an example of a SAN constructed according to the present invention
  • FIG. 2 is a block diagram of an overview of the Storage over Internet Protocol (SoIP) implementation;
  • Figure 3 illustrates the required protocol conversion steps between Fibre
  • Channek- ⁇ SCSI and IP devices in the apparatus switch fabric according to an embodiment of the present invention
  • Figure 4 is an overview of the legacy storage protocol conversion method by which the functionality of the invention is achieved;
  • Figure 5 is a high level switch diagram outlining the basic architecture of the physical apparatus according to an embodiment of the present invention.
  • FIGS. 6a-c illustrate FCP packet encapsulation according to an embodiment of the present invention
  • Figure 7 shows the frame flow for the "session" initialization for Fibre Channel devices connected to an SoIP network
  • Figures 8 and 9 show the flow of data frames for a node login initiated by FC port A of switch 1 to FC Port B of switch 2 located remotely according to an embodiment of the present invention
  • Figure 10 shows the routing of Port Login Request and Response frames for local FC ports according to an embodiment of the present invention
  • Figure 11 shows an example of the address domains which exist in a network according to one embodiment of the present invention.
  • Figures 12a-d illustrate a network architecture and address tables for a Third Party Command example
  • Figure 13 illustrates layer 2 FCP packet encapsulation according to an embodiment of the present invention
  • Figures 14a-c illustrate examples of UDP Frame demultiplexing according to embodiments of the present invention
  • Figure 15 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet according to an embodiment of the present invention
  • Figure 16 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet, where two routing blocks are combined into a single block according to an embodiment of the present invention
  • Figure 17 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet wherein low- level port interface logic blocks are combined according to an embodiment of the present invention
  • Figure 18 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet using a Field Programmable Gate Array (FPGA) according to an embodiment of the present invention
  • FPGA Field Programmable Gate Array
  • Figure 19 shows a block diagram of a common FC/Gigabit Ethernet port combined with a GBIC interface according to an embodiment of the present invention.
  • Figure 20 illustrates the architecture of an intelligent network interface card (NIC) according to an embodiment of the present invention.
  • NIC network interface card
  • FIG 1 illustrates an example of a storage area network (SAN) 10 according to an embodiment of the present invention.
  • network 10 includes numerous storage devices, such as tape libraries 15, RAID drives 20 and optical drives 25 (e.g., CD, DVD, etc.) and servers 30.
  • the storage devices can be either storage targets (e.g., tape libraries 15, RAID drives 20, etc.) or initiators (e.g., servers 30). Note that a device could be both an initiator and a target.
  • the invention is implemented in a switching device 35 within network 10.
  • each switching device 35 is an "edge" switch which provides the connectivity between nodes (i.e., one or more storage devices) and a network 40.
  • the switch resides on the "edge" of the network where the devices are located.
  • Each edge switch 35 allows connected storage elements to communicate through the edge switch with no traffic being sent to network 40.
  • Each edge switch 35 also allows storage elements connected to different edge switches to communicate with each other through network 40.
  • network 40 is an Ethernet network, but other networks may be used, for example, Asynchronous Transfer Mode (ATM)-based or FDDI-based networks, or the like.
  • ATM Asynchronous Transfer Mode
  • a switching device 35 is implemented in an SoIP (Storage over Internet Protocol) storage area network (SAN) as shown in Figure 2.
  • SoIP is a framework for transporting SCSI commands and data-ever IP networks using the Fibre Channel Protocol for SCSI (FCP) for communication between IP networked storage devices.
  • FCP Fibre Channel Protocol for SCSI
  • FCP is an FC-4 Upper Layer Protocol for sending SCSI commands and data over a Fibre Channel network yielding a "serial" SCSI network.
  • the SoIP framework enables FCP for use on an IP network by defining the SoIP protocol.
  • each SoIP device 50 converts SCSI commands and data into FCP data frames in FCP block 52.
  • the SoIP protocol layer block 54 then encapsulates these FCP frames in multiple IP packets using either the User Datagram Protocol (UDP) or Transport Control Protocol (TCP).
  • IP port 56 forwards the packet to IP network 60, which routes the IP packets between the devices 50 or to switch 35.
  • IP network 60 is preferably an Ethernet network, but may be based on any IP-compatible media including ATM, FDDI, SONET and the like.
  • the storage name server 65 serves as a database where devices store their own information and retrieve information on other devices in the SoIP network.
  • the SoIP proxy 70 performs protocol conversion between SoIP based on UDP and SoIP based on TCP.
  • FIG. 3 illustrates data exchange between storage devices using a switch 135 according to an embodiment of the present invention.
  • switch 135 is configured to receive data from different interfaces, each of which has a different data or frame format.
  • SCSI device 105 transmits data using a "parallel" SCSI interface 106
  • Fibre Channel (FC) device 1 10 transmits data using Fibre Channel interface 1 11
  • Ethernet device 1 15 transmits data using Ethernet interface 116.
  • Switch 135 translates data received from a source port in one of the three different formats into an internal format and transfers the data in the internal format through switch fabric 140 to a destination port. The destination port translates the data back into the native format appropriate for the connection thereto.
  • FC Fibre Channel
  • each device e.g., SCSI device 105, FC device 110, Ethernet device 115, or generic IP device 120 (e.g., disk drive, tape drive, server), performs storage operations based on the SCSI Command Set.
  • SCSI device 105 SCSI device 105
  • FC device 110 FC device 110
  • Ethernet device 115 Ethernet device 115
  • generic IP device 120 e.g., disk drive, tape drive, server
  • SCSI device 105 For SCSI device 105 the SCSI commands and data are transferred directly using a "parallel" bus 106.
  • the SCSI port interface 125 of switch 135 acts like a SCSI to FC bridge so that the SCSI port looks like an FC port from the point of view of switch fabric 140.
  • the SCSI data is preferably converted to FCP, and is not actually transmitted using a Fibre Channel interface.
  • Ethernet device 115 SCSI commands and data are converted to FCP and then encapsulated in an IP packet using UDP or TCP. The IP packet is then encapsulated in an Ethernet frame and transmitted using Ethernet interface 1 16.
  • SCSI device implies a device with a "parallel SCSI bus” while the term “Fibre Channel device” implies a device with a Fibre Channel interface. Both devices operate as SCSI devices at the command level.
  • SCSI device 105 does not convert the SCSI commands and data to an FCP format. Therefore, it is not possible to transfer data between FC device 110 and SCSI device 105 directly.
  • switch 135 it is possible for all devices connected to switch 135 to exchange data frames because the data format of all interfaces into switch fabric 140 are FCP compatible frames.
  • Fibre Channel with another interface.
  • Figure 3 shows a storage device constructed using Ethernet in the same manner as a device is constructed with FC. Ethernet simply replaced Fibre Channel as the media for transport.
  • Infimband may also be implemented, for example in generic IP device 120.
  • Infiniband is an I/O interface that merges the work of NGIO (Next Generation I/O) and Future I O.
  • Figure 4 illustrates data exchange between Fibre Channel, SCSI and IP devices in switch apparatus 135 according to an embodiment of the present invention.
  • the example in Figure 4 is for an Ethernet based IP network 160, however any other IP networks based on other protocols such as ATM, FDDI, etc. may be used. Similar to the embodiment in Figure 3, Figure 4 shows the protocol translations which occur for each device.
  • SCSI device 105 communicates with switch 135 using SCSI commands directly with no encapsulation of data or commands in data frames.
  • FC device 1 10 uses the FCP protocol to send SCSI commands and data to switch 135.
  • Switch 135 converts the received data to a common protocol based on FCP to allow the devices to communicate with each other. In addition, switch 135 performs address translation between the Fibre Channel and SCSI addressing schemes to the IP addressing method as will be discussed in more detail below. This is done transparently so that no changes are required in Fibre Channel device 110 or SCSI device 105, or in any host bus adapters, driver software or application software.
  • FIG. 5 is a high-level switch diagram outlining the basic architecture of a physical switch apparatus 235 according to an embodiment of the present invention.
  • switch 235 includes three main elements: switch fabric 240, management processor 250 and port interfaces 270.
  • Switch fabric 240 provides a high bandwidth mechanism for transferring data between the various port interfaces 270 as well as between port interfaces 270 and management processor 250.
  • Management processor 250 performs management related functions for switch 235 (e.g. switch initialization, configuration, SNMP, Fibre Channel services, etc ) p ⁇ ma ⁇ ly through management bus 255
  • Port interfaces 270 convert data packets from the input frame format (e g , parallel SCSI, FC, or Ethernet) to an internal frame format
  • the internal frame format data packets are then routed within switch fab ⁇ c 240 to the approp ⁇ ate destination port interface Port interfaces 270 also determine how packets are routed within the switch
  • the amount of processing performed by each port interface 270 is dependent on the interface type SCSI ports 270 ⁇ and 270 2 provide the most processing because the SCSI interface is half-duplex and it is not frame o ⁇ ented
  • the SCSI port interfaces 270 ⁇ and 270 2 also emulate the functionality of a SCSI host and or target Fibre Channel ports 270 3 and 270 4 require the least amount of processing because the internal frame format is most compatible with Fibre Channel
  • IP ports 270 5 and 270 6 e g , Ethernet ports
  • SCSI ports 270 ⁇ and 270 2 convert data received into an internal frame format before sending the packets through switch fab ⁇ c 240 Because FCP frames are not directly compatible with an Ethernet
  • Figure 6a illustrates FCP packet encapsulation in an IP frame earned over an Ethernet frame according to an embodiment of the present invention
  • Field Definitions for Figure 6a include the following
  • Ethernet packet type CHECKSUM PAD An optional 2-byte field which may be used to guarantee that the UDP checksum is correct even when a data frame begins transmission before all of the contents are known.
  • the SoIP Header field contain the following parameters CLASS: This 4-bit field indicates the class of service. In one embodiment, orly the values 2 or 3 are used.
  • NERS This 4-bit field indicates the protocol version of SoIP.
  • SoIP FLAGS This 8-bit field contains bits that indicate various parameters for a data frame as shown in Figure 6b.
  • the User Datagram Protocol (UDP) Header is the protocol used within the IP packet. TCP may also be used.
  • the UDP header defined in RFC 768, is 8 bytes in length consisting of four 16-bit fields as shown in Figure 6c, with the following field definitions:
  • SOURCE PORT An optional field. When meaningful, it indicates the port of the sending process, and may be assumed to be the port to which a reply should be addressed in the absence of other information. If not used, a value of zero is inserted.
  • DESTINATION PORT has a meaning within the context of a particular internet destination address.
  • the UDP Length is the sum of the UDP Header Length, FCP Header length, and FCP Payload length and optionally the checksum pad.
  • CHECKSUM the 16-bit one's complement of the one's complement sum of a pseudo header of information from the IP header, the UDP header, and the data, padded with bytes of zero at the end (if necessary) to make a multiple of 2 bytes.
  • a switch 235 encapsulates FC packets into an Ethernet Frame with a "wrapper" around the FC information.
  • the encapsulation of an FCP data frame in an Ethernet packet may require that the FCP data frame be limited in size because the maximum FCP data frame size is 2136 bytes (24 byte header + 21 12 byte payload) whereas an Ethernet packet has a maximum size of 1518 bytes.
  • a node's maximum receive data field size is provided to switch fab ⁇ c 240 du ⁇ ng "Fab ⁇ c Login” and to each destination node du ⁇ ng "Port Login " The fab ⁇ c or node being "logged into” generates a login response which indicates the maximum receive data field size for data frames it is capable of receiving Note that these values may not be the same For example, a fab ⁇ c may have
  • the upper limit value is set by determining or discove ⁇ ng the maximum IP datagram size and subtracting 60 bytes to account for the va ⁇ ous headers and trailers For example, for an Ethernet Frame, the upper limit value equals 1440 bytes That is, the payload for an FCP Frame cannot exceed 1440 bytes in size This limit is established because an FCP Frame being transported across an IP network will not be allowed to fragment Allowing IP datagrams to fragment degrades network performance and so most networks rarely fragment An IP header's Do Not
  • Fragment Flag can be used to prevent the IP layer from fragmenting the datagram Even with node login setting an approp ⁇ ate size for the FCP payload, this bit is set to ensure that fragmentation does not occur
  • the payload is padded to a multiple of 4 bytes to make it easy to convert frames being sent to legacy FC devices
  • Each switch 235 preferably makes use of the Buffer to Buffer Receive
  • one method for enforcing the maximum frame size is to intercept Node Login and Node Response frames which are redirected to the management processor 250, which adjusts the Buffer to Buffer Receive Data Field Size in the frame as necessary and then routes the modified packet to the original destination.
  • Name server 65 is preferably a node attached to the network, but may be implemented within a switch and can be distributed across the fabric for redundancy and to ensure quick access.
  • FIG. 7 is a flow diagram that illustrates an example of "session" initialization for Fibre Channel devices connected to an SoIP network according to an embodiment of the present invention.
  • switch 235 assigns an IP address to the device from a block of IP addresses assigned to the switch.
  • Switch 235 uses the IP address assigned to an FC device when packets are transmitted to or received from an IP compatible port.
  • An FC device now has 2 addresses assigned by the SoIP network: the FC address and the IP address.
  • the FC address is used when FC devices communicate in a local Fibre Channel "Island" while the IP address is used when the device communicates across an IP network.
  • FIGs 8 and 9 show the flow of data frames for a Node login initiated by FC port A of switch 1 to FC Port B of switch 2 located remotely from switch 1 according to an embodiment of the present invention.
  • the Login request data frame is redirected by FC port A to the Switch 1 Management Processor which makes any changes required in the Buffer-to-Buffer Receive Data Field Size parameter and then forwards the packet to the original destination port.
  • the Switch 1 Management Processor which makes any changes required in the Buffer-to-Buffer Receive Data Field Size parameter and then forwards the packet to the original destination port.
  • switch 2 there is no redirection of the Login Request packet.
  • the Login Response frame from FC Port B is redirected to the Switch 2 Management Processor which changes the Buffer-to- Buffer Receive Data Field Size in the response if necessary.
  • each Management Processor adjusts its Buffer to Buffer Receive Data Field Size to a value to allow the FCP data frames to fit into an IP packet which can be transmitted across an Ethernet network without being fragmented. Therefore, each Management Processor may need to perform MTU (Maximum Transmission Unit) discovery to determine a size which does not result in fragmentation of IP packets in the network.
  • MTU Maximum Transmission Unit
  • the switch (i.e. connected to the same switch), it is not necessary to change the Buffer to Buffer Receive Data Field Size of the Login request or response.
  • the switch supports the maximum frame size for transfers between FC ports (on the same switch).
  • the FC port interface logic will always redirect the Port Login packets to the switch's Management Processor to simplify the port interface logic.
  • the switch looks and acts like an FC switch from the point of view of any FC devices connected thereto.
  • An example of the routing of Port Login Request and Response frames for local FC ports is shown in Figure 10.
  • routing FC Port Login Request/Response packets to the Management Processor allows the Port Login for SCSI ports to be handled by the Management Processor.
  • the Management Processor always handles login for SCSI.
  • an SoIP device is uniquely identified using two parameters: an IP address and an SoIP socket number. Therefore, it is possible for a device to have a unique IP address or for multiple devices to share an IP address. For example, all of the devices on a Fibre Channel arbitrated loop may share an IP address while a server Host Bus Adapter may have a dedicated IP address.
  • a single SoIP device connected directly to an IP network must have a unique IP address in order for the network to be able to route data frames to the device. An IP network will not route traffic based on the SoIP socket number.
  • devices connected to a switch e.g., switch 235
  • an SoIP network SAN with "legacy" Fibre Channel devices attached has different address domains due to the two different address methods used: IP and Fibre Channel.
  • Figure 11 shows an example of the address domains which exist in a network according to one embodiment of the present invention.
  • SoIP devices communicate using IP addresses and the SoIP socket numbers while the Fibre Channel devices (CSI devices are treated as Fibre Charmel devices by a switch) use Fibre Channel addresses.
  • Each switch 235 performs address translation between the IP and Fibre Channel address domains.
  • Switch 235 1 performs address translation between the IP address domain and FC address domain 1
  • Switch 235 performs address translation between the IP address domain and FC address domain 2.
  • Each switch 235 assigns an IP address, SoIP socket number and Fibre Channel address to each Fibre Channel device when the device performs a fabric login.
  • a Fibre Channel device only learns about its assigned Fibre Channel address.
  • the assigned IP address, SoIP socket number and Fibre Channel Address are maintained within a translation table (not shown) in the switch.
  • Parallel SCSI devices are assigned their addresses by the switch during initialization of the SCSI port.
  • the Fibre Channel ports direct all Name server requests by a Fibre Channel device to the management processor for processing.
  • the management processor converts Fibre Channel Name Server requests into SoIP Name Server requests that are then forwarded to the SoIP Name Server, e.g., implemented in server 280.
  • SoIP name server functionality is distributed and thus handled directly by the management processor.
  • Responses from the name server are returned to the management processor where they are converted into Fibre Channel Name Server responses before being forwarded to the port that originated the name server request.
  • switch 235 converts the packet into an SoIP compatible packet.
  • the conversion encapsulates the FCP data frame in an IP data frame as described above.
  • the IP addresses and SoIP socket numbers are derived by using the Fibre Channel source address (S_ID) and the destination address (D_ID) as "keys" into the IP/Fibre Channel address conversion table on the name server.
  • the Fibre Channel address fields are replaced by the SoIP socket numbers when translating a Fibre Channel data frame to an SoIP data frame.
  • the packet is then transmitted on the IP network and routed using the destination IP address. If the destination device is an SoIP compatible device, the packet is processed directly (i.e., de- encapsulated and processed as an FCP packet) by the destination device.
  • the packet is routed to a switch 235, which receives the packet, de-encapsulates the SoIP packet and replaces the SoIP socket numbers with the appropriate source and destination Fibre Channel addresses based on the source and destination IP addresses and SoIP socket numbers.
  • a switch 235 which receives the packet, de-encapsulates the SoIP packet and replaces the SoIP socket numbers with the appropriate source and destination Fibre Channel addresses based on the source and destination IP addresses and SoIP socket numbers.
  • local assignment is the preferred method for assigning SoIP socket numbers.
  • native SoIP devices select their SoIP socket numbers while an SoIP switch (e.g., switch 235) assigns the SoIP socket number for Fibre Channel and SCSI devices attached to the switch-
  • the SoIP socket number is assigned locally, the value chosen may be any value that results in a unique IP Address/SoIP socket number combination.
  • Devices that share an IP address must be assigned unique SoIP socket numbers in order to create a unique IP Address/SoIP socket number pair.
  • Devices that have a unique IP address may have any desired SoIP socket number.
  • an SoIP switch assigns the SoIP socket numbers in such a manner as to simplify the routing of received data frames.
  • a switch must also assign a locally significant Fibre Channel address to each "remote" device for use by the local devices in addressing the "remote” devices. These locally assigned addresses are only known by a switch within its Fibre Channel address domain- Thus each switch maintains a set of locally assigned Fibre Channel addresses which correspond to the globally known IP Address/SoIP Port Number pairs defined in the SoIP Name Server.
  • each switch 235 intercepts Fibre Channel Extended Link Service requests and responses which have Fibre Channel address information embedded in the payload.
  • Extended Link Service requests and responses which have Fibre Channel address information embedded in the payload.
  • the IP Address and SoIP socket number assigned to a Fibre Channel or SCSI device are determined by the switch. The assignment of these addresses is implementation dependent.
  • the SoIP socket number is assigned the device's local Fibre Channel address.
  • the switch obtains the local Fibre Channel address directly from the received data frame.
  • IP addresses are assigned such that at least a subset of a switch's attached devices share an IP address. For example, an IP address can be assigned to each switch port. Each device attached to that switch port then shares the port's IP address. Thus, an attached Fibre Channel N_Port would have a unique IP address while the devices on a Fibre Channel arbitrated loop attached thereto would share an IP address.
  • Fibre Channel addresses are assigned globally. Globally assigned Fibre Channel addresses provide the maximum compatibility for "legacy" Fibre Channel devices.
  • the SoIP name server is responsible for managing the allocation of Global Fibre Channel Addresses.
  • a global Fibre Channel address space may need to be supported because in some cases Fibre Channel addresses may be embedded within "third-party" SCSI commands.
  • An example of such a third-party command is COPY.
  • the COPY command instructs another device to copy data.
  • the use of "third-party" commands is rare but when used, either the command would need to be modified for address compatibility or the Fibre Channel addresses would need to be globally assigned.
  • FIG. 12a an example third party COPY command will be used to illustrate a problem that occurs with locally assigned Fibre Channel addresses and third-party commands.
  • locally assigned Fibre Channel Addresses are also used as the SoIP socket number.
  • Each device has a unique IP address in this example.
  • Figure 12b shows the IP Address and SoIP socket number each device has advertised to the Name Server which identifies how the device is addressed within the SoIP network.
  • Each device is uniquely identified by the combination of IP Address and SoIP socket number.
  • the switches 235 3 and 235 and Tape Library C are aware of every device in the system. Tape Library C would then have an address table that is the same as the name server's address table.
  • Switches 235 and 235 4 will have assigned local Fibre Channel addresses to each device.
  • Figure 12c illustrates the address table stored on switch 235
  • Figure 12d illustrates the address table stored on switch 235 4 . Because the Fibre Channel addresses are assigned locally, the address assignment is purely arbitrary.
  • Server A in local domain 1 sends a COPY command to Server B in local domain 3 indicating that data is to be copied from RAID drive B to Tape Library B, both of which are located in local domain 3.
  • the COPY command will contain the addresses from Server A's perspective. Therefore, referring to Figure 12c, the command received by Server B is COPY from Fibre Channel device 000500 (RAID drive B) to Fibre Channel device 000600 (Tape Library B). However, Server B will interpret the COPY command using the address table of switch 235 ( Figure 12d) and assume it should copy data from RAID drive A to Tape Library A and not RAID drive B to Tape Library B. Thus, the wrong operation will be performed.
  • Server B sends a command to Server A to copy from RAID drive A to Tape Library C.
  • the command will be COPY from Fibre Channel address 000500 to 009900 (the addresses are from the perspective of switch 235 ).
  • Server A will assume the command is to copy data from RAID drive B to a nonexistent device because 009900 is not in the address table of switch 235 4 .
  • the switch gets around this problem by intercepting each third party command and modifying the embedded Fibre Channel addresses to be compatible with the destination device.
  • this requires that the source switch know the assignment of local addresses in the destination switch- While it is possible for a switch to convert the third-party commands, alternative methods are preferred.
  • Fibre Channel addresses are globally assigned for devices that are referenced by Fibre Channel address in third-party commands.
  • the use of a Global Fibre Channel address allows third-party commands to be used with no modification, but sets the total number of devices possible in an SoIP network to the same maximum as a Fibre Channel network. Only those devices that are referenced in a third-party command require a global address, although all devices within an SoIP network can be assigned global addresses.
  • a Globally Assigned Fibre Channel address is preferably used as the device's SoIP socket number. This simplifies the conversion of "legacy" Fibre Channel data frames to SoIP compatible data frames. Therefore, globally assigning Fibre Channel addresses is equivalent to globally assigning SoIP socket numbers.
  • Global Sc IP socket number allocation is managed by the SoIP Name Server, which allocates Global SoIP socket numbers as requested from a pool of free socket numbers, and deallocates socket numbers (returns them to the free pool) when they are no longer used.
  • the assignment of Global SoIP socket numbers for all devices in an SoIP network is the simplest solution from a management standpoint because it does not require specifying the subset of devices that require a Global SoIP socket number (or alternatively, the devices that can use a local SoIP socket number).
  • SoIP network either have a locally assigned SoIP socket number or a globally assigned SoIP socket number. All SoIP compatible devices and switches support both modes. Each device or switch determines from the SoIP Name Server which mode is to be supported when it logs into the network.
  • An SoIP Name Server configuration parameter indicates the SoIP socket number allocation mode-
  • This new third-party command is the EXTENDED COPY command.
  • Native SoIP devices preferably use the version of third-party commands that embed World Wide Names in the command when SoIP socket numbers are locally assigned.
  • the requester when SoIP socket numbers are assigned globally, the requester indicates the minimum number of socket numbers requested and a 24-bit mask defining the boundary. For example, a 16-port switch may request 4096 socket numbers with a bit mask of FFFOOO (hex) indicating that the socket numbers should be allocated on a boundary where the lower 12 bits are 0. The switch would then allocate 256 socket numbers to each port (for support of an arbitrated loop). Allocation of socket numbers on a specified boundary allows the switch to allocate socket numbers that directly co ⁇ elate to port numbers. In the above example, bits 11 :8 would identify the port. Native SoIP devices preferably allocate only one global SoIP socket number from the SoIP Name Server.
  • the SoIP Name Server also includes a configuration parameter that selects "Maximum Fibre Channel Compatibility" mode which only has meaning for Global assignment of SoIP socket numbers- Devices are able to query the Name server for the value of this parameter.
  • this mode specifies that global SoIP socket numbers are to be allocated in blocks of 65536 (on boundaries of 65536) to switches.
  • This mode is compatible with the existing Fibre Channel modes of address allocation where the lower 8 bits identify the device, the middle 8 bits identify the port and the upper 8 bits identify the switch. SoIP switches check for this mode and, if enabled, request 65536 socket numbers when requesting global SoIP socket numbers.
  • Native SoIP devices preferably allocate only one global SoIP socket number from the SoIP Name Server.
  • the frame format when operating in a Layer 2 network (e.g., no IP routers), the frame format is modified to simplify the encapsulation logic.
  • a Layer 2 network does not require the IP Header or the UDP header. All frames are forwarded using the physical address (e.g. Ethernet MAC address).
  • a switch then routes frames internally based on the Layer 2 physical address (e.g. Ethernet MAC address) combined with the SoIP socket number.
  • the Layer 2 physical address replaces the IP address as a parameter in uniquely identifying an SoIP device.
  • Figure 13 shows the frame format for an FCP frame transmitted on Ethernet.
  • An Ethernet Type value 290 is defined specifically for SoIP to allow a station receiving the frame to distinguish the frame from other frame types (e.g., IP).
  • the IP and UDP headers have been removed which reduces the frame overhead-
  • An advantage is that the length and checksum fields in the UDP header no longer need to be generated.
  • the generation of the IP and UDP headers introduces additional latency for the frame transmission because the length and checksum are located at the beginning of the frame. Therefore, it is necessary to buffer the entire frame to determine the length and checksum and write them into the header.
  • For an Ethernet Layer 2 SoIP frame it is only necessary to determine the amount of padding, if any, added at the end of the frame.
  • the number of PAD bytes must be included in the SoIP Header to allow the PAD bytes to be removed at the receiving station. Since the padding is only required to satisfy a minimum Ethernet frame size of 64 bytes, it is possible to complete the header generation after 64 bytes of the frame (or the entire frame) have been received.
  • the Layer 2 frame format is similar to the Layer 3 frame format SoIP Frame conversion described above with reference to Figure 6 with the following differences: a. The IP and UDP headers are no longer present. b. The Ethernet Type value is different. c. The CHEKSUM PAD field is replaced by the FC CRC field- The FC CRC field is a 4-byte field containing the Fibre Channel CRC calculated over the FCP header and payload. This field may be inserted by a source when a Fibre
  • Channel data frame is encapsulated with no changes. Thus, the CRC received with the frame is still valid.
  • the CHECKSUM PAD flag is replaced by the FC CRC PRESENT flag. This bit indicates if the FC CRC field is present in the frame. Note that the CHECKSUM PAD field has no meaning since there is no need to calculate a UDP checksum.
  • the FRAME PAD LENGTH may have a non-zero value since the encapsulated frame length may be less than the Ethernet minimum of 64.
  • the UDP Header contains a Destination Port field and a Source Port field.
  • FIG. 14a illustrates "demultiplexing" of UDP datagrams as is typical in the industry.
  • Figures 14b and 14c illustrate ways to add an SoIP layer according to embodiments of the present invention.
  • Figure 14b illustrates frame demultiplexing when there is a single port number assigned to all SoIP devices. Further demultiplexing is then performed using the SoIP socket number to determine the device. Routing data frames to applications is then performed based on the FCP exchange numbers located in the FCP header.
  • Figure 14c illustrates a similar example, but with separate UDP port numbers assigned to each SoIP device. In this case, it is not necessary to examine the SoIP socket number in order to forward the UDP datagram. (The SoIP socket number and IP address must still uniquely identify the device). The choice of whether to use a single UDP port number for each SoIP device or one UDP port number for all devices is implementation dependent.
  • the UDP demultiplexing examples illustrated in Figures 14b and 14c are oriented toward a server with one or more host bus adapters (where the host bus adapters are the SoIP devices).
  • a switch is generally less complicated in the sense that data frames are forwarded to end devices and the application layer does not have to be handled.
  • the addressing mechanisms described above allow software applications to appear as SoIP devices by registering with the name server using a different address. This opens up the possibility for applications to advertise themselves in the name server for use by other applications.
  • An example is a COPY manager that could be used by a higher level backup application.
  • each storage device when it registers with the name server, must include the UDP port number to use when sending data frames to the device- In a normal UDP application, the destination port would save the source port number for use in sending a reply.
  • this mechanism is not feasible for use with "legacy" FC switches since it requires the switch to associate the source port numbers with the exchange ID's. It is much simpler to require a storage device to always use the same UDP port number.
  • a storage device is identified by 3 parameters in the name server database: IP Address, UDP Port Number, and SoIP socket number.
  • An additional parameter required is the physical address (e.g. Ethernet MAC address) which is determined in the normal manner for IP networks.
  • ARP address resolution protocol
  • ARP address resolution protocol
  • the physical address to use can also be learned when a frame is received from a device. For example, the physical address can be learned when a Port Login request is received.
  • the physical address may not be the physical address of the actual device but the address of an IP router.
  • the SoIP Name Server must have a UDP Port number that is known by all of the SoIP devices within an SoIP network since the port number cannot be learned from another source. This could be a "well-known" port number or a registered port number. This approach is similar to a Domain Name Server (DNS) that has a well- known port number of 53. The assignment of "well-known" port numbers is done by the IANA (Internet Assigned Numbers Authority). Routing within an IP network is affected by the choice of addressing mode which impacts the abilit/ of switches and routers to determine what constitutes a "conversation".
  • a conversation is a set of data frames that are related and which should arrive in order. However, it is assumed that conversations have no ordering relationship. In other words, the ordering of frames from different conversations can be changed with no effect. For example, assume that frames for 3 conversations (A, B and C) are transmitted in the following order (Al sent first):
  • the frames for a particular conversation arrive in order with respect to each other, but out of order with respect to frames from other conversations.
  • the ability to identify different conversations allows load balancing to be performed by allowing traffic to be routed on a conversation basis.
  • Switches and routers can determine conversations based on several parameters within a data frame including Destination/Source addresses, IP Protocol, UDP/TCP Port Numbers, etc. The parameters actually used are dependent on the switch/router implementation. Storage traffic between the same two devices should be treated as a single conversation. It is not acceptable for storage commands to be received out of order because there may be a relationship between the commands (e.g. ordered queuing).
  • FIG. 15 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Gigabit Ethernet according to an embodiment of the present invention.
  • the Fibre Channel and Gigabit Ethernet ports use the same encoding/decoding method (8B/10B) with each port requiring a serializer/deserializer (SERDES) block for converting to/from the high speed serial interface. Therefore, these two interfaces share the 8B/10B block 310 and SERDES block 315 in this embodiment as shown in Figure 15.
  • SERDES serializer/deserializer
  • a multiplexer 345 selects the clock used by the logic based on the port type.
  • these two interfaces share the switch fabric interface logic block 320 which interfaces with the switch fabric (including the management interface).
  • the MAC blocks (blocks 325 and 330) implement the appropriate protocol state machines for the interface (Fibre Channel or Gigabit Ethernet).
  • the MAC blocks 325 and 330 convert received data into frames which are forwarded to the routing logic blocks 335 and 340, respectively.
  • the MAC blocks 325 and 330 also receive data frames from the routing logic blocks 335 and 340, respectively, which are then transmitted according to the interface's (Fibre Channel or Gigabit Ethernet) protocol.
  • Routing logic blocks 335 and 340 determine where each received frame should be routed based on addressing information within the frame. Routing logic blocks 335 and 340 also perform any modifications to the frames that are required. For example, a routing logic block will remove the SoIP encapsulation from a frame being forwarded to a Fibre
  • the routing logic block then sends the frame to the switch fabric with an indication of the destination output ports.
  • Egress data frames (frames from the switch fabric to the output port) are received by a routing logic block and forwarded to the associated MAC. Additional processing may be performed on the frame by the routing logic block before the MAC receives the frame.
  • Ethernet port routing logic block 340 may convert a Fibre Channel frame into an SoIP frame.
  • routing logic block 350 includes logic blocks which are dependent on the port type and other blocks that are common to both port types. This optimization reduces the number of logic gates required on an ASIC. Routing block 350 determines where a frame is routed based on addressing information within the data frame. This function is known as address resolution and is performed for both Fibre Channel and Gigabit Ethernet data frames. Therefore, address resolution logic can be shared by these two port interfaces though it is necessary for the routing logic to select different data based on the port type.
  • Routing Logic block 350 can be implemented as hard coded logic or as a programmable method using a network processor, which is designed specifically for processing packets and which can be programmed to route either Fibre Channel frames or Ethernet frames. Therefore, the routing logic hardware can be shared by using different network processor software.
  • routing logic block 350 also includes an input and output FIFO memory which is shared by the two port interfaces. Additional logic which can be shared include statistics registers and control registers. Statistics registers are used to count the number of frames received, frames transmitted, bytes received, bytes transmitted, etc. A common set of statistics registers can be used. These registers are modified by control signals from each MAC. Control registers determine the operating mode of each MAC.
  • a common set of statistics and control registers reduces the logic required to implement the registers and for interfacing with an external control source such as a switch management CPU.
  • the low-level port interface logic e.g., FC MAC block 325 and Ethernet MAC block 330
  • FC MAC block 325 and Ethernet MAC block 330 is combined into a single MAC block 360.
  • FC MAC block 325 and Ethernet MAC block 330 are combined into a single MAC block 360.
  • a Field Programmable Gate Array (FPGA) 370 is used to select the interface protocol supported by the port.
  • the FPGA configuration loaded would be based on the port type.
  • separate FPGA code is developed for the Fibre Channel and Gigabit Ethernet interfaces.
  • the FPGA logic can be optimized for the particular interface.
  • a single hardware design supports both interfaces, with software determining the FPGA code to be downloaded based on the port type.
  • a common port must also deal with the physical interface external to an ASIC.
  • such an interface may include, for example, a copper, multi- mode fiber or single-mode fiber interface.
  • the components are not necessarily the same between Fibre Channel and Ethernet.
  • a Gigabit Interface Converter (GBIC) 380 is provided to allow a user to select the desired physical interface.
  • a GBIC is a standardized module which has a common form factor and electrical interface and allows any of the many physical interfaces to be installed.
  • GBIC modules are available from many vendors (e.g. HP, AMP, Molex, etc.) and support all of the standard Fibre Channel and Gigabit Ethernet physical interfaces.
  • Figure 14 shows a block diagram of a common FC/Gigabit Ethernet port interface (e.g., as shown in Figures 15, 16, 17 and 18) combined with a GBIC interface according to this embodiment.
  • the ASIC connects to a GBIC connector 385 which allows the user to change GBIC modules.
  • the user can select the media type by installing the appropriate GBIC 380.
  • GBIC modules typically contain a serial EEROM whose contents can be read to determine the type of module (e.g. Fibre Channel, Gigabit Ethernet, Infiniband, Copper, Multi-mode, Single-mode, etc.).
  • the GBIC can thus indicate the type of interface, e.g., FC or GE or Infiniband, to use.
  • FC or GE Fibre Channel
  • the port interface type is user switchable/configurable, and in another embodiment the type of the link interface is automatically determined through added intelligence, for example, through a "handshake".
  • an SoIP intelligent network interface card (NIC) 400 is provided as shown in Figure 20.
  • NIC card 400 is able to send and receive both IP and SoIP traffic. In either case, NIC card 400 has the intelligence to determine the type of traffic and direct it accordingly.
  • the host 410 may issue both storage commands and network commands to NIC card 400 through the PCI interface 420. These commands are sent with a specified address which is used to direct the commands to either the Direct Path or the Storage Traffic Engine. Storage commands are issued via the SCSI Command Set, and Network commands are issued via Winsock and/or TCP/IP. NIC card 4C0 directs storage commands to the Storage Traffic Engine 430 based on the specified addiess. Storage Traffic Engine 430 handles the exchange management and sequence management for the duration of the SCSI operation. SCSI operations are then carried out via SoIP and transmitted to the network 470 via a media access controller (MAC) block 450, which in one embodiment is a Gigabit Ethernet MAC.
  • MAC media access controller
  • NIC card 400 directs non-SoIP traffic to the Direct Path 440 based on the specified address.
  • the Direct Path 440 processes the commands and transmits the specified packets to network 470 via block 450.
  • NIC 400 demultiplexes the traffic and directs it accordingly.
  • Storage traffic received as SoIP is sent to storage traffic block 430.
  • Non-SoIP traffic is sent directly to the host via direct path 440.
  • the multiplexer block 460 handles arbitration for the output path when both Direct Path 440 and Storage Traffic Engine 430 simultaneously send traffic to MAC 450.
  • Mux block 460 demultiplexes the traffic and sends it accordingly to either Direct Path 440 or Storage Traffic Engine 430.

Abstract

A method and apparatus for transferring data between IP devices (115) and SCSI (105) or Fibre Channel FC (110) devices over an IP network (160). The device interfaces may be either SCSI, FC or IP interface such as Gigavit Ethernet. Data is switched between SCSI and IP, FC and IP, or between SCSI and FC. Data can also be switched from SCSI to SCSI, IP to IP, or FC to FC. The port interfaces provide the conversion from the input frame format to an internal frame format, which can be routed within the apparatus. The amount of processing performed by each port interface is dependent on the interface type. The apparatus permits rapid transfer of information packets between interfaces at latency levels meeting the stringent requirements for storage protocols. The configuration control can be applied to each port on a switch, and each switch on the network, via an SNMP or Web-based interface.

Description

METHOD AND APPARATUS FOR TRANSFERRING DATA BETWEEN DIFFERENT NETWORK DEVICES OVER AN IP NETWORK
CROSS REFERENCES TO RELATED APPLICATIONS
This application is related to U.S. Provisional Patent Application Serial No. 60/123,606 (Atty. Docket No. 019678-000100), filed March 10, 1999, entitled 0 "METHOD AND APPARATUS FOR TRANSFERRING DATA BETWEEN IP
NETWORK DEVICES AND SCSI AND FIBRE CHANNEL DEVICES OVER AN IP NETWORK," the disclosure of which is hereby incorporated by reference in its entirety.
5 BACKGROUND OF THE INVENTION
The present invention relates to transferring information between storage devices and a network via a switched, packetized communications system. In particular, the present invention relates to methods and apparatus for receiving, translating, and routing data-packets between SCSI (Small Computer Systems Interface), Fibre Channel 0 and Ethernet devices in a flexible, programmable manner.
In enterprise computing environments, it is desirable and beneficial to have multiple servers able to directly access multiple storage devices to support high- bandwidth data transfers, system expansion, modularity, configuration flexibility and optimization of resources. In conventional computing environments, such access is 5 typically provided via file system level Local Area Network (LAN) connections, which operate at a fraction of the speed of direct storage connections. As such, access to storage systems is highly susceptible to bottlenecks.
Storage Area Networks (SANs) have been proposed as one method of solving this storage access bottleneck problem. By applying the networking paradigm to 0 storage devices, SANs enable increased connectivity and bandwidth, sharing of resources, and configuration flexibility. The current SAN paradigm assumes that the entire network is constructed using Fibre Channel switches. Therefore, most solutions involving SANs require implementation of separate networks: one to support the normal LAN and another to support the SAN. The installation of new equipment and technology, such as new equipment at the storage device level (Fibre Channel interfaces), the host/server level (Fibre Channel adapter cards) and the transport level (Fibre Channel hubs, switches and routers), into a mission-critical enterprise computing environment could be described as less than desirable for data center managers, as it involves replication of network infrastructure, new technologies (i.e., Fibre Channel), and new training for personnel. Most companies have already invested significant amounts of money constructing and maintaining their network (e.g., based on Ethernet and/or ATM). Construction of a second high-speed network based on a different technology is a significant impediment to the proliferation of SANs. Therefore, a need exists for a method and apparatus that can alleviate problems with access to storage devices by multiple hosts, while retaining current equipment and network infrastructures, and minimizing the need for new training for data center personnel.
In general, a majority of storage devices currently use "parallel" SCSI or Fibre Channel data transfer protocols whereas most LANs use an Ethernet protocol, such as Gigabit Ethernet. SCSI, Fibre Channel and Ethernet are protocols for data transfer, each of which uses a different individual format for data transfer. For example, SCSI commands were designed to be implemented over a parallel bus architecture and therefore are not packetized. Fibre Channel, like Ethernet, uses a serial interface with data transferred in packets- However, the physical interface and frame formats between Fibre Channel and Ethernet are not compatible. Gigabit Ethernet was designed to be compatible with existing Ethernet infrastructures and is therefore based on an Ethernet packet architecture. Because of these differences there is a need for new methods and apparatus to allow efficient communication between these protocols.
SUMMARY OF THE INVENTION The present invention solves the above and other problems, thereby advancing the state of the useful arts, by providing methods and apparatus for transferring data between storage device interfaces and network interfaces. In particular, the present invention brings sophisticated SAN capabilities to existing enterprise computing configurations, without the installation of costly Fibre Channel switches and hubs, by providing the means for Internet Protocol (IP) devices to transparently communicate with SCSI and Fibre Channel devices over an IP network. The present invention accomplishes this through the use of Fibre Channel Protocol (FCP), an industry standard developed for implementation of SCSI commands over a Fibre Channel network. The invention allows the storage devices to retain the use of standard SCSI and Fibre Channel storage interfaces and construct a SAN using a company's existing network infrastructure. Therefore, no changes are required in host bus adapters (HBA) or storage devices (e.g. disk drives, tape drives, etc).
According to the present invention, methods and apparatus are provided for transferring data between IP devices (including, but not limited to, Gigabit Ethernet devices) and SCSI or Fibre Channel devices. The device interfaces may be either SCSI, Fibre Channel or IP interfaces such as Gigabit Ethernet. Data is switched between SCSI and IP, Fibre Channel and IP, or between SCSI and Fibre Channel. Data can also be switched from SCSI to SCSI, Fibre Channel to Fibre Channel and IP to IP. The port interfaces provide the conversion from the input frame format to an internal frame format, which can be routed within the apparatus. The apparatus may include any number of total ports. The amount of processing performed by each port interface is dependent on the interface type. The processing capabilities of the present invention permit rapid transfer of information packets between multiple interfaces at latency levels meeting the stringent requirements for storage protocols. The configuration control can be applied to each port on a switch and, in turn, each switch on the network, via an SNMP or Web-based interface, providing a flexible, programmable control for the apparatus. According to one aspect of the present invention, a method is provided for routing data packets in a switch device in a network such as a SAN. The method typically comprises the steps of receiving a packet from a first network device at a first port interface of the switch device, wherein the packet is one of a SCSI formatted packet (i.e., SCSI formatted data stream converted into a packet), a Fibre Channel (FC) formatted packet and an Internet protocol (IP) formatted packet, wherein the first port interface is communicably coupled to the first network device, and converting the received packet into a packet having an internal format. The method also typically includes the steps of routing the internal format packet to a second port interface of the switch device, reconverting the internal format packet to one of a SCSI formatted packet, an FC formatted packet or an IP formatted packet, and transmitting the reconverted packet to a second network device communicably coupled to the second port interface. According to another aspect of the present invention, a network switch device is provided which typically comprises a first port interface including a means for receiving data packets from a network device, wherein the receiving means receives one of a SCSI formatted packet and a Fibre Channel (FC) formatted packet from a first network device, and a means for converting received packets into packets having an internal format, wherein the received data packet is converted into a first packet having the internal format. The switch device also typically comprises a second port interface including a means for reconverting packets from the internal format to an IP format, wherein the first packet is converted into a packet having an IP format, and a means for transmitting IP packets to a network, wherein the IP formatted packet is transmitted to an IP network. A means for routing the first packet to the second port interface is also provided.
According to yet another aspect of the present invention, a network switch device is provided which typically comprises a first port interface including a means for receiving data packets from an IP network, wherein the first interface means receives a packet in an IP format, and a means for converting received packets into packets having an internal format, wherein the received packet is converted into a first packet having an internal format. The switch device also typically comprises a second port interface including a means for reconverting packets having the internal format to packets having the SCSI format, and a means for transmitting reconverted packets to a SCSI network device. The switch device further typically includes a third port interface having a means for reconverting packets having the internal format to packets having the FC format, and a means for transmitting reconverted packets to a FC network device. A means for routing packets between the first, second and third port interfaces is also typically provided. In operation, wherein if the first packet is routed to the second port interface, the first packet is converted to the SCSI format and transmitted to the SCSI network device, and wherein if the first packet is routed to the third port interface, the first packet is converted to the FC format and transmitted to the FC network device.
According to a further aspect of the present invention, a network switch device is provided for use in a storage area network (SAN). The switch device typically comprises a first port interface communicably coupled to a SCSI device, wherein the first port interface converts SCSI formatted data packets received from the SCSI device into data packets having an internal format, and wherein the first port interface converts data packets having the internal format into SCSI formatted data packets. The switch device also typically comprises a second port interface communicably coupled to a FC device, wherein the second port interface converts FC formatted data packets received from the FC device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format into FC formatted data packets. The switch device further typically includes a third port interface communicably coupled to a IP device, wherein the third port interface converts IP formatted data packets received from the IP device into data packets having the internal format, and wherein the third port interface converts data packets having the internal format into IP formatted data packets, and a switch fabric for routing data packets having the internal format between the first, second and third port interfaces. In typical operation, when a first one of the SCSI, FC and IP devices sends a first data packet to a second one of the SCSI, FC and IP devices, the port interface coupled to the first device converts the first data packet to a packet having the internal format and routes the internal format packet through the switch fabric to the port interface coupled to the second device, wherein the port interface coupled to the second device reconverts the internal format packet into the format associated with the second device and sends the reconverted packet to the second device. According to yet a further aspect of the present invention, a network switch device for use in a storage area network (SAN) is provided. The switch may comprise any combination of Fibre Channel, SCSI, Ethernet and Infiniband ports, and may comprise any number of total ports. The switch device typically comprises a first port interface communicably coupled to one of a SCSI device(s), an FC device, or an IP device, a second port interface, wherein the second port interface is configurable to communicate with either a FC device or an Ethernet device, and a switch fabric for routing data packets having the internal format between the first and second port interfaces. In typical operation, when the second port interface is configured to communicate with a FC device, the second port interface converts FC formatted data packets received from the FC device into data packets having an internal format, and wherein the second port interface converts data packets having the internal format received from the switch fabric into FC formatted data packets, and wherein when the second port interface is configured to communicate with an Ethernet device, the second port interface converts Ethernet formatted data packets received from the Ethernet device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format received from the switch fabric into Ethernet formatted data packets. The second port interface can be either self-configurable or user configurable.
Reference to the remaining portions of the specification, including the drawings and claims, will realize other features and advantages of the present invention. Further features and advantages of the present invention, as well as the structure and operation of various embodiments of the present invention, are described in detail below with respect to the accompanying drawings. In the drawings, like reference numbers indicate identical or functionally similar elements.
BRIEF DESCRIPTION OF THE DRAWINGS
Figure 1 illustrates an example of a SAN constructed according to the present invention;
Figure 2 is a block diagram of an overview of the Storage over Internet Protocol (SoIP) implementation; Figure 3 illustrates the required protocol conversion steps between Fibre
Channek-^SCSI and IP devices in the apparatus switch fabric according to an embodiment of the present invention;
Figure 4 is an overview of the legacy storage protocol conversion method by which the functionality of the invention is achieved; Figure 5 is a high level switch diagram outlining the basic architecture of the physical apparatus according to an embodiment of the present invention;
Figures 6a-c illustrate FCP packet encapsulation according to an embodiment of the present invention;
Figure 7 shows the frame flow for the "session" initialization for Fibre Channel devices connected to an SoIP network; Figures 8 and 9 show the flow of data frames for a node login initiated by FC port A of switch 1 to FC Port B of switch 2 located remotely according to an embodiment of the present invention;
Figure 10 shows the routing of Port Login Request and Response frames for local FC ports according to an embodiment of the present invention;
Figure 11 shows an example of the address domains which exist in a network according to one embodiment of the present invention;
Figures 12a-d illustrate a network architecture and address tables for a Third Party Command example; Figure 13 illustrates layer 2 FCP packet encapsulation according to an embodiment of the present invention;
Figures 14a-c illustrate examples of UDP Frame demultiplexing according to embodiments of the present invention;
Figure 15 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet according to an embodiment of the present invention;
Figure 16 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet, where two routing blocks are combined into a single block according to an embodiment of the present invention;
Figure 17 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet wherein low- level port interface logic blocks are combined according to an embodiment of the present invention; Figure 18 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Ethernet using a Field Programmable Gate Array (FPGA) according to an embodiment of the present invention;
Figure 19 shows a block diagram of a common FC/Gigabit Ethernet port combined with a GBIC interface according to an embodiment of the present invention; and
Figure 20 illustrates the architecture of an intelligent network interface card (NIC) according to an embodiment of the present invention. DESCRIPTION OF THE SPECIFIC EMBODIMENTS Figure 1 illustrates an example of a storage area network (SAN) 10 according to an embodiment of the present invention. As shown, network 10 includes numerous storage devices, such as tape libraries 15, RAID drives 20 and optical drives 25 (e.g., CD, DVD, etc.) and servers 30. The storage devices can be either storage targets (e.g., tape libraries 15, RAID drives 20, etc.) or initiators (e.g., servers 30). Note that a device could be both an initiator and a target. In a preferred embodiment, the invention is implemented in a switching device 35 within network 10. For example, as shown in Figure 1, each switching device 35 is an "edge" switch which provides the connectivity between nodes (i.e., one or more storage devices) and a network 40. In other words, the switch resides on the "edge" of the network where the devices are located. Each edge switch 35 allows connected storage elements to communicate through the edge switch with no traffic being sent to network 40. Each edge switch 35 also allows storage elements connected to different edge switches to communicate with each other through network 40. In a preferred embodiment, network 40 is an Ethernet network, but other networks may be used, for example, Asynchronous Transfer Mode (ATM)-based or FDDI-based networks, or the like.
In one embodiment, a switching device 35 is implemented in an SoIP (Storage over Internet Protocol) storage area network (SAN) as shown in Figure 2. According to the present invention, SoIP is a framework for transporting SCSI commands and data-ever IP networks using the Fibre Channel Protocol for SCSI (FCP) for communication between IP networked storage devices. A majority of storage devices currently communicate using either a "parallel" SCSI bus or a Fibre Channel serial interface. FCP is an FC-4 Upper Layer Protocol for sending SCSI commands and data over a Fibre Channel network yielding a "serial" SCSI network. The SoIP framework enables FCP for use on an IP network by defining the SoIP protocol. Storage devices and host bus adapters operating the SoIP protocol form a storage area network (SAN) directly on an IP network. This framework offers an enormous advantage in the installation and utility of SANs. As shown in Figure 2, each SoIP device 50 converts SCSI commands and data into FCP data frames in FCP block 52. The SoIP protocol layer block 54 then encapsulates these FCP frames in multiple IP packets using either the User Datagram Protocol (UDP) or Transport Control Protocol (TCP). IP port 56 forwards the packet to IP network 60, which routes the IP packets between the devices 50 or to switch 35. IP network 60 is preferably an Ethernet network, but may be based on any IP-compatible media including ATM, FDDI, SONET and the like. The storage name server 65 serves as a database where devices store their own information and retrieve information on other devices in the SoIP network. The SoIP proxy 70 performs protocol conversion between SoIP based on UDP and SoIP based on TCP.
Because the majority of storage devices currently use "parallel" SCSI or Fibre Channel protocols, the transition to SoIP-based SANs may be hampered unless such "legacy" devices can be connected to an SoIP network. For these "legacy" devices, a switch as shown in Figure 3 is provided for connection into an SoIP SAN.
Figure 3 illustrates data exchange between storage devices using a switch 135 according to an embodiment of the present invention. In this embodiment, switch 135 is configured to receive data from different interfaces, each of which has a different data or frame format. SCSI device 105 transmits data using a "parallel" SCSI interface 106, Fibre Channel (FC) device 1 10 transmits data using Fibre Channel interface 1 11 and Ethernet device 1 15 transmits data using Ethernet interface 116. Switch 135 translates data received from a source port in one of the three different formats into an internal format and transfers the data in the internal format through switch fabric 140 to a destination port. The destination port translates the data back into the native format appropriate for the connection thereto.
In this embodiment, each device, e.g., SCSI device 105, FC device 110, Ethernet device 115, or generic IP device 120 (e.g., disk drive, tape drive, server), performs storage operations based on the SCSI Command Set. For Fibre Channel device 110, the SCSI commands and data are converted to FCP and transmitted using Fibre
Channel interface 111. For SCSI device 105 the SCSI commands and data are transferred directly using a "parallel" bus 106. In this embodiment, the SCSI port interface 125 of switch 135 acts like a SCSI to FC bridge so that the SCSI port looks like an FC port from the point of view of switch fabric 140. As shown, the SCSI data is preferably converted to FCP, and is not actually transmitted using a Fibre Channel interface. For Ethernet device 115, SCSI commands and data are converted to FCP and then encapsulated in an IP packet using UDP or TCP. The IP packet is then encapsulated in an Ethernet frame and transmitted using Ethernet interface 1 16. Note that the term "SCSI device" implies a device with a "parallel SCSI bus" while the term "Fibre Channel device" implies a device with a Fibre Channel interface. Both devices operate as SCSI devices at the command level. Note that SCSI device 105 does not convert the SCSI commands and data to an FCP format. Therefore, it is not possible to transfer data between FC device 110 and SCSI device 105 directly. As shown in Figure 3, it is possible for all devices connected to switch 135 to exchange data frames because the data format of all interfaces into switch fabric 140 are FCP compatible frames. Also note that it is possible to replace Fibre Channel with another interface. For example, Figure 3 shows a storage device constructed using Ethernet in the same manner as a device is constructed with FC. Ethernet simply replaced Fibre Channel as the media for transport. Infimband may also be implemented, for example in generic IP device 120. As is well known, Infiniband is an I/O interface that merges the work of NGIO (Next Generation I/O) and Future I O. Figure 4 illustrates data exchange between Fibre Channel, SCSI and IP devices in switch apparatus 135 according to an embodiment of the present invention. The example in Figure 4 is for an Ethernet based IP network 160, however any other IP networks based on other protocols such as ATM, FDDI, etc. may be used. Similar to the embodiment in Figure 3, Figure 4 shows the protocol translations which occur for each device. SCSI device 105 communicates with switch 135 using SCSI commands directly with no encapsulation of data or commands in data frames. FC device 1 10 uses the FCP protocol to send SCSI commands and data to switch 135. Switch 135 converts the received data to a common protocol based on FCP to allow the devices to communicate with each other. In addition, switch 135 performs address translation between the Fibre Channel and SCSI addressing schemes to the IP addressing method as will be discussed in more detail below. This is done transparently so that no changes are required in Fibre Channel device 110 or SCSI device 105, or in any host bus adapters, driver software or application software.
Figure 5 is a high-level switch diagram outlining the basic architecture of a physical switch apparatus 235 according to an embodiment of the present invention. In this embodiment, switch 235 includes three main elements: switch fabric 240, management processor 250 and port interfaces 270. Switch fabric 240 provides a high bandwidth mechanism for transferring data between the various port interfaces 270 as well as between port interfaces 270 and management processor 250. Management processor 250 performs management related functions for switch 235 (e.g. switch initialization, configuration, SNMP, Fibre Channel services, etc ) pπmaπly through management bus 255
Port interfaces 270 convert data packets from the input frame format (e g , parallel SCSI, FC, or Ethernet) to an internal frame format The internal frame format data packets are then routed within switch fabπc 240 to the appropπate destination port interface Port interfaces 270 also determine how packets are routed within the switch The amount of processing performed by each port interface 270 is dependent on the interface type SCSI ports 270ι and 2702 provide the most processing because the SCSI interface is half-duplex and it is not frame oπented The SCSI port interfaces 270ι and 2702 also emulate the functionality of a SCSI host and or target Fibre Channel ports 2703 and 2704 require the least amount of processing because the internal frame format is most compatible with Fibre Channel In essence, IP ports 2705 and 2706 (e g , Ethernet ports) and SCSI ports 270ι and 2702 convert data received into an internal frame format before sending the packets through switch fabπc 240 Because FCP frames are not directly compatible with an Ethernet interface as they are with a Fibre Channel interface, the transmission of FCP packets on an Ethernet interface requires that an FCP frame be encapsulated in an Ethernet frame as shown in Figure 6a
Figure 6a illustrates FCP packet encapsulation in an IP frame earned over an Ethernet frame according to an embodiment of the present invention Field Definitions for Figure 6a include the following
DA Ethernet Destination Address (6 bytes) SA Ethernet Source Address (6 Bytes) TYPE The Ethernet packet type CHECKSUM PAD An optional 2-byte field which may be used to guarantee that the UDP checksum is correct even when a data frame begins transmission before all of the contents are known The CHECKSUM PAD bit m the SoIP Header indicates if this field is present
ETHERNET CRC Cyclic Redundancy Check (4 bytes)
As shown in Figure 6a, the SoIP Header field contain the following parameters CLASS: This 4-bit field indicates the class of service. In one embodiment, orly the values 2 or 3 are used.
NERS: This 4-bit field indicates the protocol version of SoIP.
SoIP FLAGS: This 8-bit field contains bits that indicate various parameters for a data frame as shown in Figure 6b.
In Figure 6a, the User Datagram Protocol (UDP) Header is the protocol used within the IP packet. TCP may also be used. The UDP header, defined in RFC 768, is 8 bytes in length consisting of four 16-bit fields as shown in Figure 6c, with the following field definitions:
SOURCE PORT: An optional field. When meaningful, it indicates the port of the sending process, and may be assumed to be the port to which a reply should be addressed in the absence of other information. If not used, a value of zero is inserted. DESTINATION PORT: has a meaning within the context of a particular internet destination address.
LENGTH: the length, in bytes, of the user datagram including the UDP header and data (thus, if there were no data in the datagram, the length would be 8). For an encapsulated FCP packet, the UDP Length is the sum of the UDP Header Length, FCP Header length, and FCP Payload length and optionally the checksum pad.
CHECKSUM: the 16-bit one's complement of the one's complement sum of a pseudo header of information from the IP header, the UDP header, and the data, padded with bytes of zero at the end (if necessary) to make a multiple of 2 bytes.
In one embodiement, a switch 235 encapsulates FC packets into an Ethernet Frame with a "wrapper" around the FC information. The encapsulation of an FCP data frame in an Ethernet packet may require that the FCP data frame be limited in size because the maximum FCP data frame size is 2136 bytes (24 byte header + 21 12 byte payload) whereas an Ethernet packet has a maximum size of 1518 bytes. The use of Ethernet Jumbo Frames, which permit packet sizes up to 9 Kbytes to be used, eliminates the need to limit the Fibre Channel frame size. However, support for Ethernet jumbo frames is limited within the existing network infrastructure Therefore, FCP data frames need to be limited otherwise a large FCP data frame may need to be "fragmented" into 2 separate Ethernet frames The Login procedures defined in the Fibre Channel standard allows devices to negotiate the maximum payload with the switch fabπc 240 Thus, the switch fabπc 240 can respond to a login with a smaller payload size than the maximum (e g , 1024 bytes) Switch 235 makes use of this fact to limit FC packets to a size which can be encapsulated in an Ethernet packet to eliminate the need for fragmenting FC packets According to one embodiment, a node's maximum receive data field size is provided to switch fabπc 240 duπng "Fabπc Login" and to each destination node duπng "Port Login " The fabπc or node being "logged into" generates a login response which indicates the maximum receive data field size for data frames it is capable of receiving Note that these values may not be the same For example, a fabπc may have the maximum allowed size of 2112 bytes while a node may limit the maximum size to 1024 bytes (e g the Hewlett-Packard Tachyon-Lite Fibre Channel Controller) A source node may not transmit a data frame larger than the maximum frame size as determined for the login response
Since an encapsulated FCP data frame cannot be larger than the maximum Ethernet packet size, an upper limit is placed on the frame payload size duπng login by a device According to one embodiment, the upper limit value is set by determining or discoveπng the maximum IP datagram size and subtracting 60 bytes to account for the vaπous headers and trailers For example, for an Ethernet Frame, the upper limit value equals 1440 bytes That is, the payload for an FCP Frame cannot exceed 1440 bytes in size This limit is established because an FCP Frame being transported across an IP network will not be allowed to fragment Allowing IP datagrams to fragment degrades network performance and so most networks rarely fragment An IP header's Do Not
Fragment Flag can be used to prevent the IP layer from fragmenting the datagram Even with node login setting an appropπate size for the FCP payload, this bit is set to ensure that fragmentation does not occur According to one embodiment, the payload is padded to a multiple of 4 bytes to make it easy to convert frames being sent to legacy FC devices Each switch 235 preferably makes use of the Buffer to Buffer Receive
Data Field size to force end nodes to communicate with data frames that will fit within an IP packet earned over an Ethernet link According to an embodiment of the present invention, one method for enforcing the maximum frame size is to intercept Node Login and Node Response frames which are redirected to the management processor 250, which adjusts the Buffer to Buffer Receive Data Field Size in the frame as necessary and then routes the modified packet to the original destination.
Following fabric login, a device logs into the name server 65. The login to name server 65 is done to establish the parameters used for communication between the device and name server 65 (e.g. maximum payload size). The devices "register" with name server 65 to provide information for a database, which describes parameters for devices on the network. An initiator can then query the database to determine information about devices in the system thereby eliminating the need to "probe" the system to determine which devices exist. Probing of the network would not be feasible since there are 16 million Fibre Channel addresses. Name server 65 is preferably a node attached to the network, but may be implemented within a switch and can be distributed across the fabric for redundancy and to ensure quick access.
Figure 7 is a flow diagram that illustrates an example of "session" initialization for Fibre Channel devices connected to an SoIP network according to an embodiment of the present invention. During fabric login, switch 235 assigns an IP address to the device from a block of IP addresses assigned to the switch.
Switch 235 uses the IP address assigned to an FC device when packets are transmitted to or received from an IP compatible port. An FC device now has 2 addresses assigned by the SoIP network: the FC address and the IP address. The FC address is used when FC devices communicate in a local Fibre Channel "Island" while the IP address is used when the device communicates across an IP network.
Figures 8 and 9 show the flow of data frames for a Node login initiated by FC port A of switch 1 to FC Port B of switch 2 located remotely from switch 1 according to an embodiment of the present invention. The Login request data frame is redirected by FC port A to the Switch 1 Management Processor which makes any changes required in the Buffer-to-Buffer Receive Data Field Size parameter and then forwards the packet to the original destination port. In the destination switch, (switch 2) there is no redirection of the Login Request packet. As shown in Figure 9, the Login Response frame from FC Port B is redirected to the Switch 2 Management Processor which changes the Buffer-to- Buffer Receive Data Field Size in the response if necessary.
In one embodiment, each Management Processor adjusts its Buffer to Buffer Receive Data Field Size to a value to allow the FCP data frames to fit into an IP packet which can be transmitted across an Ethernet network without being fragmented. Therefore, each Management Processor may need to perform MTU (Maximum Transmission Unit) discovery to determine a size which does not result in fragmentation of IP packets in the network. When an FC port performs a Port Login with an FC port which is local
(i.e. connected to the same switch), it is not necessary to change the Buffer to Buffer Receive Data Field Size of the Login request or response. This is because, in one embodiment, the switch supports the maximum frame size for transfers between FC ports (on the same switch). However, the FC port interface logic will always redirect the Port Login packets to the switch's Management Processor to simplify the port interface logic. Thus, in this embodiment, the switch looks and acts like an FC switch from the point of view of any FC devices connected thereto. An example of the routing of Port Login Request and Response frames for local FC ports is shown in Figure 10.
According to one embodiment, routing FC Port Login Request/Response packets to the Management Processor allows the Port Login for SCSI ports to be handled by the Management Processor. The Management Processor always handles login for SCSI.
According to one embodiment, an SoIP device is uniquely identified using two parameters: an IP address and an SoIP socket number. Therefore, it is possible for a device to have a unique IP address or for multiple devices to share an IP address. For example, all of the devices on a Fibre Channel arbitrated loop may share an IP address while a server Host Bus Adapter may have a dedicated IP address. In one: embodiment, there are two possible modes for assignment of the SoIP socket number: local or global. A single SoIP device connected directly to an IP network must have a unique IP address in order for the network to be able to route data frames to the device. An IP network will not route traffic based on the SoIP socket number. However, devices connected to a switch (e.g., switch 235) may share an IP address if the switch uses both the IP address and the SoIP socket number when switching data frames.
According to the present invention, an SoIP network SAN with "legacy" Fibre Channel devices attached has different address domains due to the two different address methods used: IP and Fibre Channel. Figure 11 shows an example of the address domains which exist in a network according to one embodiment of the present invention. SoIP devices communicate using IP addresses and the SoIP socket numbers while the Fibre Channel devices ( CSI devices are treated as Fibre Charmel devices by a switch) use Fibre Channel addresses. Each switch 235 performs address translation between the IP and Fibre Channel address domains. Switch 2351 performs address translation between the IP address domain and FC address domain 1, and Switch 235 performs address translation between the IP address domain and FC address domain 2. Each switch 235 assigns an IP address, SoIP socket number and Fibre Channel address to each Fibre Channel device when the device performs a fabric login. A Fibre Channel device only learns about its assigned Fibre Channel address. The assigned IP address, SoIP socket number and Fibre Channel Address are maintained within a translation table (not shown) in the switch. Parallel SCSI devices are assigned their addresses by the switch during initialization of the SCSI port. The Fibre Channel ports direct all Name server requests by a Fibre Channel device to the management processor for processing.
According to one embodiment of the present invention, the management processor converts Fibre Channel Name Server requests into SoIP Name Server requests that are then forwarded to the SoIP Name Server, e.g., implemented in server 280. In one embodiment, the SoIP name server functionality is distributed and thus handled directly by the management processor. Responses from the name server are returned to the management processor where they are converted into Fibre Channel Name Server responses before being forwarded to the port that originated the name server request. When a Fibre Channel device sends data frames to a device not located in its Fibre
Channel address domain, switch 235 converts the packet into an SoIP compatible packet. The conversion encapsulates the FCP data frame in an IP data frame as described above. Referring back to Figure 6a, in one embodiment, the IP addresses and SoIP socket numbers are derived by using the Fibre Channel source address (S_ID) and the destination address (D_ID) as "keys" into the IP/Fibre Channel address conversion table on the name server. The Fibre Channel address fields are replaced by the SoIP socket numbers when translating a Fibre Channel data frame to an SoIP data frame. The packet is then transmitted on the IP network and routed using the destination IP address. If the destination device is an SoIP compatible device, the packet is processed directly (i.e., de- encapsulated and processed as an FCP packet) by the destination device. However, if the destination is a Fibre Channel (or parallel SCSI) device, the packet is routed to a switch 235, which receives the packet, de-encapsulates the SoIP packet and replaces the SoIP socket numbers with the appropriate source and destination Fibre Channel addresses based on the source and destination IP addresses and SoIP socket numbers.
According to one embodiment, local assignment is the preferred method for assigning SoIP socket numbers. In this embodiment, native SoIP devices select their SoIP socket numbers while an SoIP switch (e.g., switch 235) assigns the SoIP socket number for Fibre Channel and SCSI devices attached to the switch- When the SoIP socket number is assigned locally, the value chosen may be any value that results in a unique IP Address/SoIP socket number combination. Devices that share an IP address must be assigned unique SoIP socket numbers in order to create a unique IP Address/SoIP socket number pair. Devices that have a unique IP address may have any desired SoIP socket number. In one embodiment, an SoIP switch assigns the SoIP socket numbers in such a manner as to simplify the routing of received data frames. A switch must also assign a locally significant Fibre Channel address to each "remote" device for use by the local devices in addressing the "remote" devices. These locally assigned addresses are only known by a switch within its Fibre Channel address domain- Thus each switch maintains a set of locally assigned Fibre Channel addresses which correspond to the globally known IP Address/SoIP Port Number pairs defined in the SoIP Name Server.
According to one embodiment, due to the different address domains, each switch 235 intercepts Fibre Channel Extended Link Service requests and responses which have Fibre Channel address information embedded in the payload. Extended Link
Service requests and responses are generated infrequently. Therefore, it is acceptable to redirect the Extended Link Service requests to the switch's management processor which makes any necessary changes to the data frame. If an Extended Link Service request/response has no addressing information embedded in the payload, the Management Processor simply retransmits the packet with no modifications.
The IP Address and SoIP socket number assigned to a Fibre Channel or SCSI device are determined by the switch. The assignment of these addresses is implementation dependent. In a preferred embodiment, the SoIP socket number is assigned the device's local Fibre Channel address. In this embodiment, the switch obtains the local Fibre Channel address directly from the received data frame.
Alternatively, assignment of the SoIP socket number is based on an incrementing number that can be used as an index into an address table. In one embodiment, each device is assigned a unique IP address. However, this type of assignment may result in the use of a large number of IP addresses. The use of a single IP address for each device also has implications for routing in the IP network. Therefore, in a preferred embodiment, IP addresses are assigned such that at least a subset of a switch's attached devices share an IP address. For example, an IP address can be assigned to each switch port. Each device attached to that switch port then shares the port's IP address. Thus, an attached Fibre Channel N_Port would have a unique IP address while the devices on a Fibre Channel arbitrated loop attached thereto would share an IP address. According to one embodiment, Fibre Channel addresses are assigned globally. Globally assigned Fibre Channel addresses provide the maximum compatibility for "legacy" Fibre Channel devices. In this embodiment, the SoIP name server is responsible for managing the allocation of Global Fibre Channel Addresses. A global Fibre Channel address space may need to be supported because in some cases Fibre Channel addresses may be embedded within "third-party" SCSI commands. An example of such a third-party command is COPY. The COPY command instructs another device to copy data. The use of "third-party" commands is rare but when used, either the command would need to be modified for address compatibility or the Fibre Channel addresses would need to be globally assigned. With reference to the SoIP network shown in Figure 12a, an example third party COPY command will be used to illustrate a problem that occurs with locally assigned Fibre Channel addresses and third-party commands. In this example locally assigned Fibre Channel Addresses are also used as the SoIP socket number. Each device has a unique IP address in this example. Figure 12b shows the IP Address and SoIP socket number each device has advertised to the Name Server which identifies how the device is addressed within the SoIP network. Each device is uniquely identified by the combination of IP Address and SoIP socket number. Assume that the switches 2353 and 235 and Tape Library C are aware of every device in the system. Tape Library C would then have an address table that is the same as the name server's address table. Switches 235 and 2354 will have assigned local Fibre Channel addresses to each device. Figure 12c illustrates the address table stored on switch 235 and Figure 12d illustrates the address table stored on switch 2354. Because the Fibre Channel addresses are assigned locally, the address assignment is purely arbitrary.
Assume that Server A in local domain 1 sends a COPY command to Server B in local domain 3 indicating that data is to be copied from RAID drive B to Tape Library B, both of which are located in local domain 3. The COPY command will contain the addresses from Server A's perspective. Therefore, referring to Figure 12c, the command received by Server B is COPY from Fibre Channel device 000500 (RAID drive B) to Fibre Channel device 000600 (Tape Library B). However, Server B will interpret the COPY command using the address table of switch 235 (Figure 12d) and assume it should copy data from RAID drive A to Tape Library A and not RAID drive B to Tape Library B. Thus, the wrong operation will be performed. As another example, assume that Server B sends a command to Server A to copy from RAID drive A to Tape Library C. The command will be COPY from Fibre Channel address 000500 to 009900 (the addresses are from the perspective of switch 235 ). Server A will assume the command is to copy data from RAID drive B to a nonexistent device because 009900 is not in the address table of switch 2354.
According to one embodiment, the switch gets around this problem by intercepting each third party command and modifying the embedded Fibre Channel addresses to be compatible with the destination device. However, this requires that the source switch know the assignment of local addresses in the destination switch- While it is possible for a switch to convert the third-party commands, alternative methods are preferred.
According to one alternative method, Fibre Channel addresses are globally assigned for devices that are referenced by Fibre Channel address in third-party commands. The use of a Global Fibre Channel address allows third-party commands to be used with no modification, but sets the total number of devices possible in an SoIP network to the same maximum as a Fibre Channel network. Only those devices that are referenced in a third-party command require a global address, although all devices within an SoIP network can be assigned global addresses. A Globally Assigned Fibre Channel address is preferably used as the device's SoIP socket number. This simplifies the conversion of "legacy" Fibre Channel data frames to SoIP compatible data frames. Therefore, globally assigning Fibre Channel addresses is equivalent to globally assigning SoIP socket numbers. Global Sc IP socket number allocation is managed by the SoIP Name Server, which allocates Global SoIP socket numbers as requested from a pool of free socket numbers, and deallocates socket numbers (returns them to the free pool) when they are no longer used. The assignment of Global SoIP socket numbers for all devices in an SoIP network is the simplest solution from a management standpoint because it does not require specifying the subset of devices that require a Global SoIP socket number (or alternatively, the devices that can use a local SoIP socket number).
Thus, all devices in an SoIP network either have a locally assigned SoIP socket number or a globally assigned SoIP socket number. All SoIP compatible devices and switches support both modes. Each device or switch determines from the SoIP Name Server which mode is to be supported when it logs into the network. An SoIP Name Server configuration parameter indicates the SoIP socket number allocation mode-
An environment that supports both local and global SoIP socket numbers is not required because it is expected that the need for global SoIP socket numbers will be eliminated due to a new form of Third-Party command format, which embeds World
Wide Names in the command instead of the Fibre Channel address. Because World Wide Names are unique, the device receiving the command is able to determine the appropriate address(es) to use from its point of view. One implementation of this new third-party command is the EXTENDED COPY command. Native SoIP devices preferably use the version of third-party commands that embed World Wide Names in the command when SoIP socket numbers are locally assigned.
In one embodiment, when SoIP socket numbers are assigned globally, the requester indicates the minimum number of socket numbers requested and a 24-bit mask defining the boundary. For example, a 16-port switch may request 4096 socket numbers with a bit mask of FFFOOO (hex) indicating that the socket numbers should be allocated on a boundary where the lower 12 bits are 0. The switch would then allocate 256 socket numbers to each port (for support of an arbitrated loop). Allocation of socket numbers on a specified boundary allows the switch to allocate socket numbers that directly coπelate to port numbers. In the above example, bits 11 :8 would identify the port. Native SoIP devices preferably allocate only one global SoIP socket number from the SoIP Name Server.
In one embodiment, the SoIP Name Server also includes a configuration parameter that selects "Maximum Fibre Channel Compatibility" mode which only has meaning for Global assignment of SoIP socket numbers- Devices are able to query the Name server for the value of this parameter. When enabled, this mode specifies that global SoIP socket numbers are to be allocated in blocks of 65536 (on boundaries of 65536) to switches. This mode is compatible with the existing Fibre Channel modes of address allocation where the lower 8 bits identify the device, the middle 8 bits identify the port and the upper 8 bits identify the switch. SoIP switches check for this mode and, if enabled, request 65536 socket numbers when requesting global SoIP socket numbers. In this mode, Native SoIP devices preferably allocate only one global SoIP socket number from the SoIP Name Server. According to one embodiment, when operating in a Layer 2 network (e.g., no IP routers), the frame format is modified to simplify the encapsulation logic. A Layer 2 network does not require the IP Header or the UDP header. All frames are forwarded using the physical address (e.g. Ethernet MAC address). A switch then routes frames internally based on the Layer 2 physical address (e.g. Ethernet MAC address) combined with the SoIP socket number. In essence, the Layer 2 physical address replaces the IP address as a parameter in uniquely identifying an SoIP device. Figure 13 shows the frame format for an FCP frame transmitted on Ethernet. An Ethernet Type value 290 is defined specifically for SoIP to allow a station receiving the frame to distinguish the frame from other frame types (e.g., IP). The IP and UDP headers have been removed which reduces the frame overhead- An advantage is that the length and checksum fields in the UDP header no longer need to be generated. The generation of the IP and UDP headers introduces additional latency for the frame transmission because the length and checksum are located at the beginning of the frame. Therefore, it is necessary to buffer the entire frame to determine the length and checksum and write them into the header. For an Ethernet Layer 2 SoIP frame, it is only necessary to determine the amount of padding, if any, added at the end of the frame. The number of PAD bytes must be included in the SoIP Header to allow the PAD bytes to be removed at the receiving station. Since the padding is only required to satisfy a minimum Ethernet frame size of 64 bytes, it is possible to complete the header generation after 64 bytes of the frame (or the entire frame) have been received.
The Layer 2 frame format is similar to the Layer 3 frame format SoIP Frame conversion described above with reference to Figure 6 with the following differences: a. The IP and UDP headers are no longer present. b. The Ethernet Type value is different. c. The CHEKSUM PAD field is replaced by the FC CRC field- The FC CRC field is a 4-byte field containing the Fibre Channel CRC calculated over the FCP header and payload. This field may be inserted by a source when a Fibre
Channel data frame is encapsulated with no changes. Thus, the CRC received with the frame is still valid. d. The CHECKSUM PAD flag is replaced by the FC CRC PRESENT flag. This bit indicates if the FC CRC field is present in the frame. Note that the CHECKSUM PAD field has no meaning since there is no need to calculate a UDP checksum. e. The FRAME PAD LENGTH may have a non-zero value since the encapsulated frame length may be less than the Ethernet minimum of 64.
The UDP Header contains a Destination Port field and a Source Port field.
The normal usage of these fields is to identify the software applications that are communicating with each other. An application requests a port number for use when sending a UDP "datagram". This port number becomes the source port number for each UDP datagram sent by the application. When a UDP datagram is received, the destination port number is used by the UDP layer to determine the application to which the datagram will be forwarded. Figure 14a illustrates "demultiplexing" of UDP datagrams as is typical in the industry.
Figures 14b and 14c illustrate ways to add an SoIP layer according to embodiments of the present invention. Figure 14b illustrates frame demultiplexing when there is a single port number assigned to all SoIP devices. Further demultiplexing is then performed using the SoIP socket number to determine the device. Routing data frames to applications is then performed based on the FCP exchange numbers located in the FCP header. Figure 14c illustrates a similar example, but with separate UDP port numbers assigned to each SoIP device. In this case, it is not necessary to examine the SoIP socket number in order to forward the UDP datagram. (The SoIP socket number and IP address must still uniquely identify the device). The choice of whether to use a single UDP port number for each SoIP device or one UDP port number for all devices is implementation dependent. The UDP demultiplexing examples illustrated in Figures 14b and 14c are oriented toward a server with one or more host bus adapters (where the host bus adapters are the SoIP devices). A switch is generally less complicated in the sense that data frames are forwarded to end devices and the application layer does not have to be handled.
The addressing mechanisms described above allow software applications to appear as SoIP devices by registering with the name server using a different address. This opens up the possibility for applications to advertise themselves in the name server for use by other applications. An example is a COPY manager that could be used by a higher level backup application.
According to one embodiment, each storage device, when it registers with the name server, must include the UDP port number to use when sending data frames to the device- In a normal UDP application, the destination port would save the source port number for use in sending a reply. However, this mechanism is not feasible for use with "legacy" FC switches since it requires the switch to associate the source port numbers with the exchange ID's. It is much simpler to require a storage device to always use the same UDP port number.
As a result, according to this embodiment, a storage device is identified by 3 parameters in the name server database: IP Address, UDP Port Number, and SoIP socket number. An additional parameter required is the physical address (e.g. Ethernet MAC address) which is determined in the normal manner for IP networks. ARP (address resolution protocol) is preferably used to learn the physical address to use fo. an IP address. The physical address to use can also be learned when a frame is received from a device. For example, the physical address can be learned when a Port Login request is received. The physical address may not be the physical address of the actual device but the address of an IP router.
The SoIP Name Server (SNS) must have a UDP Port number that is known by all of the SoIP devices within an SoIP network since the port number cannot be learned from another source. This could be a "well-known" port number or a registered port number. This approach is similar to a Domain Name Server (DNS) that has a well- known port number of 53. The assignment of "well-known" port numbers is done by the IANA (Internet Assigned Numbers Authority). Routing within an IP network is affected by the choice of addressing mode which impacts the abilit/ of switches and routers to determine what constitutes a "conversation". A conversation is a set of data frames that are related and which should arrive in order. However, it is assumed that conversations have no ordering relationship. In other words, the ordering of frames from different conversations can be changed with no effect. For example, assume that frames for 3 conversations (A, B and C) are transmitted in the following order (Al sent first):
Al A2 Bl B2 B3 A3 B4 A4 A5 A6 A7 B5 B6 B7 Cl C2 C3 A8.
It is permissible for the frames to be received in any of the following sequences (note that there are many more possible sequences that are acceptable):
Al A2 A3 A4 A5 A6 A7 A8 Bl B2 B3 B4 B5 B6 B7 Cl C2 C3; Al A2 A3 A4 A5 A6 Cl C2 Bl B2 B3 B4 B5 B6 B7 C3 A7 A8; and
Cl C2 Al C3 A2 Bl B2 A3 A4 A5 A6 B3 B4 B5 A7 A8 B6 B7.
In each of the above sequences, the frames for a particular conversation arrive in order with respect to each other, but out of order with respect to frames from other conversations. The ability to identify different conversations allows load balancing to be performed by allowing traffic to be routed on a conversation basis. Switches and routers can determine conversations based on several parameters within a data frame including Destination/Source addresses, IP Protocol, UDP/TCP Port Numbers, etc. The parameters actually used are dependent on the switch/router implementation. Storage traffic between the same two devices should be treated as a single conversation. It is not acceptable for storage commands to be received out of order because there may be a relationship between the commands (e.g. ordered queuing). Therefore, it is preferable to select an addressing mechanism that makes a device unique to a switch/router but does not attempt to distinguish commands. Different IP addresses are an ideal choice for distinguishing devices since this method works with all switches and routers. When an IP address is shared, it is preferred that the UDP Port Numbers be unique for the devices sharing the IP address. Thus, devices that share an IP address have the possibility to be treated separately by switches and routers that classify conversations based on UDP port numbers. It is understood that the discussion of UDP Port Numbers above also applies to TCP Header Port Numbers when SoIP is implemented using TCP instead of UDP.
Figure 15 is a high level block diagram which illustrates the basic architecture for a switch port that supports both Fibre Channel and Gigabit Ethernet according to an embodiment of the present invention. The Fibre Channel and Gigabit Ethernet ports use the same encoding/decoding method (8B/10B) with each port requiring a serializer/deserializer (SERDES) block for converting to/from the high speed serial interface. Therefore, these two interfaces share the 8B/10B block 310 and SERDES block 315 in this embodiment as shown in Figure 15. These two interface types differ in clock speed with Fibre Channel operating at 1062.5 MHz and Gigabit Ethernet operating at 1250 MHz. Higher speed versions of these interfaces are being developed which will also have a different clock speed- Therefore, a multiplexer 345 selects the clock used by the logic based on the port type. In addition, these two interfaces share the switch fabric interface logic block 320 which interfaces with the switch fabric (including the management interface). The MAC blocks (blocks 325 and 330) implement the appropriate protocol state machines for the interface (Fibre Channel or Gigabit Ethernet). The MAC blocks 325 and 330 convert received data into frames which are forwarded to the routing logic blocks 335 and 340, respectively. The MAC blocks 325 and 330 also receive data frames from the routing logic blocks 335 and 340, respectively, which are then transmitted according to the interface's (Fibre Channel or Gigabit Ethernet) protocol. Routing logic blocks 335 and 340 determine where each received frame should be routed based on addressing information within the frame. Routing logic blocks 335 and 340 also perform any modifications to the frames that are required. For example, a routing logic block will remove the SoIP encapsulation from a frame being forwarded to a Fibre
Channel port. The routing logic block then sends the frame to the switch fabric with an indication of the destination output ports. Egress data frames (frames from the switch fabric to the output port) are received by a routing logic block and forwarded to the associated MAC. Additional processing may be performed on the frame by the routing logic block before the MAC receives the frame. For example, Ethernet port routing logic block 340 may convert a Fibre Channel frame into an SoIP frame.
According to another embodiment of the present invention as shown in Figure 16, the two routing blocks of Figure 15 are combined into a single routing logic block 350. This optimization is possible because the routing logic used by these two interfaces is very similar. In one embodiment, routing logic block 350 includes logic blocks which are dependent on the port type and other blocks that are common to both port types. This optimization reduces the number of logic gates required on an ASIC. Routing block 350 determines where a frame is routed based on addressing information within the data frame. This function is known as address resolution and is performed for both Fibre Channel and Gigabit Ethernet data frames. Therefore, address resolution logic can be shared by these two port interfaces though it is necessary for the routing logic to select different data based on the port type. The logic within Routing Logic block 350 can be implemented as hard coded logic or as a programmable method using a network processor, which is designed specifically for processing packets and which can be programmed to route either Fibre Channel frames or Ethernet frames. Therefore, the routing logic hardware can be shared by using different network processor software. In one embodiment, routing logic block 350 also includes an input and output FIFO memory which is shared by the two port interfaces. Additional logic which can be shared include statistics registers and control registers. Statistics registers are used to count the number of frames received, frames transmitted, bytes received, bytes transmitted, etc. A common set of statistics registers can be used. These registers are modified by control signals from each MAC. Control registers determine the operating mode of each MAC. A common set of statistics and control registers reduces the logic required to implement the registers and for interfacing with an external control source such as a switch management CPU. In another embodiment as shown in Figure 17, the low-level port interface logic (e.g., FC MAC block 325 and Ethernet MAC block 330) is combined into a single MAC block 360. One problem with this approach, however, is that these two logic blocks have little in common. In addition, it is possible to purchase proprietary blocks which implement Gigabit Ethernet MAC and Fibre Channel Port Interface logic. Combining these two blocks would severely hinder the use of these proprietary blocks.
According to another embodiment of the present invention as shown in Figure 18, a Field Programmable Gate Array (FPGA) 370 is used to select the interface protocol supported by the port. The FPGA configuration loaded would be based on the port type. In this embodiment, separate FPGA code is developed for the Fibre Channel and Gigabit Ethernet interfaces. Thus, the FPGA logic can be optimized for the particular interface. A single hardware design supports both interfaces, with software determining the FPGA code to be downloaded based on the port type.
A common port must also deal with the physical interface external to an ASIC. As is well known, such an interface may include, for example, a copper, multi- mode fiber or single-mode fiber interface. Also, the components are not necessarily the same between Fibre Channel and Ethernet. According to an embodiment of the present invention as shown in Figure 19, a Gigabit Interface Converter (GBIC) 380 is provided to allow a user to select the desired physical interface. A GBIC is a standardized module which has a common form factor and electrical interface and allows any of the many physical interfaces to be installed. GBIC modules are available from many vendors (e.g. HP, AMP, Molex, etc.) and support all of the standard Fibre Channel and Gigabit Ethernet physical interfaces. Figure 14 shows a block diagram of a common FC/Gigabit Ethernet port interface (e.g., as shown in Figures 15, 16, 17 and 18) combined with a GBIC interface according to this embodiment. The ASIC connects to a GBIC connector 385 which allows the user to change GBIC modules. Thus, the user can select the media type by installing the appropriate GBIC 380.
GBIC modules typically contain a serial EEROM whose contents can be read to determine the type of module (e.g. Fibre Channel, Gigabit Ethernet, Infiniband, Copper, Multi-mode, Single-mode, etc.). The GBIC can thus indicate the type of interface, e.g., FC or GE or Infiniband, to use. However, it is possible for the GBIC to support multiple interfaces, for example both FC and GE. Therefore, in one embodiment, the port interface type is user switchable/configurable, and in another embodiment the type of the link interface is automatically determined through added intelligence, for example, through a "handshake". According to another embodiment of the present invention, an SoIP intelligent network interface card (NIC) 400 is provided as shown in Figure 20. NIC card 400 is able to send and receive both IP and SoIP traffic. In either case, NIC card 400 has the intelligence to determine the type of traffic and direct it accordingly.
The host 410 may issue both storage commands and network commands to NIC card 400 through the PCI interface 420. These commands are sent with a specified address which is used to direct the commands to either the Direct Path or the Storage Traffic Engine. Storage commands are issued via the SCSI Command Set, and Network commands are issued via Winsock and/or TCP/IP. NIC card 4C0 directs storage commands to the Storage Traffic Engine 430 based on the specified addiess. Storage Traffic Engine 430 handles the exchange management and sequence management for the duration of the SCSI operation. SCSI operations are then carried out via SoIP and transmitted to the network 470 via a media access controller (MAC) block 450, which in one embodiment is a Gigabit Ethernet MAC. NIC card 400 directs non-SoIP traffic to the Direct Path 440 based on the specified address. The Direct Path 440 processes the commands and transmits the specified packets to network 470 via block 450. When receiving data from network 470 via MAC 450, NIC 400 demultiplexes the traffic and directs it accordingly. Storage traffic received as SoIP is sent to storage traffic block 430. Non-SoIP traffic is sent directly to the host via direct path 440.
The multiplexer block 460 handles arbitration for the output path when both Direct Path 440 and Storage Traffic Engine 430 simultaneously send traffic to MAC 450. For traffic received from network 470 by MAC 450, Mux block 460 demultiplexes the traffic and sends it accordingly to either Direct Path 440 or Storage Traffic Engine 430.
While the invention has been described by way of example and in terms of the specific embodiments, it is to be understood that the invention is not limited to the disclosed embodiments. To the contrary, it is intended to cover various modifications and similar arrangements as would be apparent to those skilled in the art. Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.

Claims

WHAT IS CLAIMED IS:
1. In a network, a method of routing data packets in a switch device, the method comprising the steps of: receiving a packet from a first network device at a first port interface of the switch device, wherein the packet is one of a SCSI formatted packet, a Fibre Channel (FC) formatted packet and an Internet protocol (IP) formatted packet, wherein the first port interface is communicably coupled to the first network device; converting the received packet into a packet having an internal format; routing the internal format packet to a second port interface of the switch device; reconverting the internal format packet to one of a SCSI formatted packet, an FC formatted packet and an IP formatted packet; and transmitting the reconverted packet to a second network device communicably coupled to the second port interface.
2. The method of claim 1 , wherein the IP formatted packet is transported over one of an Ethernet protocol, and ATM protocol and a FDDI protocol.
3. The method of claim 1 , wherein the second port interface couples the switch device to a network, wherein the step of transmitting includes sending the reconverted packet to the second network device over the network, wherein the reconverted packet is in the IP format.
4. The method of claim 3, wherein the network is an Ethernet network, wherein the IP format is an Ethernet format, and wherein the step of reconverting includes the step of encapsulating the internal format packet in an Ethernet frame.
5. The method of claim 1 , wherein the first port interface couples the switch device to a network, wherein the step of receiving includes receiving the packet from the first network device over the network, wherein the received packet is in the IP format.
6. The method of claim 5, wherein the network is an Ethernet network, and wherein the IP format is an Ethernet format.
7. The method of claim 1 , wherein the reconverted packet is in a different format than the received packet.
8. The method of claim 1 , wherein the first network device is one of a server and a storage device, and wherein the second network device is one of a server and a storage device.
9. The method of claim 1, wherein the internal format is an FCP- based format.
10. A network switch device, comprising: a) a first port interface including: means for receiving data packets from a network device, wherein the receiving means receives one of a SCSI formatted packet and a Fibre Channel (FC) formatted packet from a first network device; and means for converting received packets into packets having an internal format, wherein the received data packet is converted into a first packet having the internal format; b) a second port interface including: means for reconverting packets from the internal format to an IP format, wherein the first packet is converted into a packet having an IP format; and means for transmitting IP packets to a network, wherein the IP formatted packet is transmitted to an IP network; and c) means for routing the first packet to the second port interface.
11. The switch device of claim 10, wherein the IP network is an
Ethernet network and wherein the IP formatted packet is encapsulated in an Ethernet frame.
12. The switch device of claim 10, wherein the internal format is an FCP-based format.
13. A network switch device, comprising: a) a first port interface including: means for receiving data packets from an IP network, wherein the first interface means receives a packet in an IP format; and means for converting received packets into packets having an internal format, wherein the received packet is converted into a first packet having an internal format; b) a second port interface including: means for reconverting packets having the internal format to packets having the SCSI format; and means for transmitting reconverted packets to a SCSI network device; c) a third port interface including: means for reconverting packets having the internal format to packets having the FC format; and means for transmitting reconverted packets to a FC network device; and d) means for routing packets between the first, second, and third port interfaces, wherein the first packet is routed to one of the second and third port interfaces; wherein if the first packet is routed to the second port interface, the first packet is converted to the SCSI format and transmitted to the SCSI network device, and wherein if the first packet is routed to the third port interface, the first packet is converted to the FC format and transmitted to the FC network device.
14. The switch device of claim 13, wherein the IP network is an Ethernet network and wherein the IP format is an Ethernet format.
15. The switch device of claim 13, wherein the internal format is one of an FCP-based format andan IP format.
16. A storage area network (SAN) comprising: a SCSI device capable of receiving and transmitting SCSI formatted data packets; a fibre channel (FC) device capable of receiving and transmitting FC formatted data packets; an IP device capable of receiving and transmitting IP formatted data packets; and a switch device including: a first port interface communicably coupled to the SCSI device, wherein the first port interface converts SCSI formatted data packets received from the SCSI device into data packets having an internal format, and wherein the first port interface converts data packets having the internal format into SCSI formatted data packets; a second port interface communicably coupled to the FC device, wherein the second port interface converts FC formatted data packets received from the FC device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format into FC formatted data packets; a third port interface communicably coupled to the IP device, wherein the third port interface converts IP formatted data packets received from the IP device into data packets having the internal format, and wherein the third port interface converts data packets having the internal format into IP formatted data packets; and a switch fabric for routing data packets having the internal format between the first, second and third port interfaces; wherein when a first one of the SCSI, FC and IP devices sends a first data packet to a second one of the SCSI, FC and IP devices, the port interface coupled to the first device converts the first data packet to a packet having the internal format and routes the internal format packet through the switch fabric to the port interface coupled to the second device, wherein the port interface coupled to the second device reconverts the internal format packet into the format associated with the second device and sends the reconverted packet to the second device.
17. The SAN of claim 16, wherein the IP formatted data packets include one of Ethernet formatted data packets, ATM formatted data packets, FDDI formatted data packets, and Infiniband formatted data packets.
18. The SAN of claim 16, wherein the internal format is an FCP-based format.
19. The SAN of claim 16, further including an IP network coupling the IP device to the third port interface.
20. A network switch device for use in a storage area network (SAN), the switch device comprising: a first port interface communicably coupled to a SCSI device, wherein the first port interface converts SCSI formatted data packets received from the SCSI device into data packets having an internal format, and wherein the first port interface converts data packets having the internal format into SCSI formatted data packets; a second port interface communicably coupled to a FC device, wherein the second port interface converts FC formatted data packets received from the FC device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format into FC formatted data packets; a third port interface communicably coupled to a IP device, wherein the third port interface converts IP formatted data packets received from the IP device into data packets having the internal format, and wherein the third port interface converts data packets having the internal format into IP formatted data packets; and a switch fabric for routing data packets having the internal format between the first, second and third port interfaces; wherein when a first one of the SCSI, FC and IP devices sends a first data packet to a second one of the SCSI, FC and IP devices, the port interface coupled to the first device converts the first data packet to a packet having the internal format and routes the internal format packet through the switch fabric to the port interface coupled to the second device, wherein the port interface coupled to the second device reconverts the internal format packet into the format associated with the second device and sends the reconverted packet to the second device.
21. The switch device of claim 20, wherein the internal format is an FCP-based format.
22. The switch device of claim 16, wherein an IP network couples the IP device to the third port interface.
23. A network switch device for use in a storage area network (SAN), the switch device comprising: a first port interface communicably coupled to one of a SCSI device a FC device and an IP device; a second port interface, wherein the second port interface is configurable to communicate with either a FC device or an Ethernet device; and a switch fabric for routing data packets having the internal format between the first and second port interfaces; wherein when the second port interface is configured to communicate with a FC device, the second port interface converts FC formatted data packets received from the FC device into data packets having an internal format, and wherein the second port interface converts data packets having the internal format received from the switch fabric into FC formatted data packets, and wherein when the second port interface is configured to communicate with an Ethernet device, the second port interface converts Ethernet formatted data packets received from the Ethernet device into data packets having the internal format, and wherein the second port interface converts data packets having the internal format received from the switch fabric into Ethernet formatted data packets.
24. The switch device of claim 23, wherein the second port interface is self-configurable based on whether a FC device or an Ethernet device is coupled to the second port interface.
25. The switch device of claim 24, wherein the second port interface includes a means for determining whether an attached device is a FC device or an Ethernet device.
26. The switch device of claim 23, wherein the second port interface is configured by a user to communicate with one of a FC device and an Ethernet device.
27. The switch device of claim 23, wherein the internal format is a FCP format.
EP00917881A 2000-02-08 2000-03-09 Method and apparatus for transferring data between different network devices over an ip network Withdrawn EP1256198A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US500119 1983-06-01
US09/500,119 US6400730B1 (en) 1999-03-10 2000-02-08 Method and apparatus for transferring data between IP network devices and SCSI and fibre channel devices over an IP network
PCT/US2000/006475 WO2001059966A1 (en) 2000-02-08 2000-03-09 Method and apparatus for transferring data between different network devices over an ip network

Publications (1)

Publication Number Publication Date
EP1256198A1 true EP1256198A1 (en) 2002-11-13

Family

ID=23988112

Family Applications (1)

Application Number Title Priority Date Filing Date
EP00917881A Withdrawn EP1256198A1 (en) 2000-02-08 2000-03-09 Method and apparatus for transferring data between different network devices over an ip network

Country Status (4)

Country Link
EP (1) EP1256198A1 (en)
JP (1) JP2004503122A (en)
AU (1) AU2000238787A1 (en)
WO (1) WO2001059966A1 (en)

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6434620B1 (en) 1998-08-27 2002-08-13 Alacritech, Inc. TCP/IP offload network interface device
AU2002248158A1 (en) 2000-11-02 2002-08-12 Pirus Networks Tcp/udp acceleration
US7864758B1 (en) * 2001-09-28 2011-01-04 Emc Corporation Virtualization in a storage system
US6976134B1 (en) * 2001-09-28 2005-12-13 Emc Corporation Pooling and provisioning storage resources in a storage network
US7707304B1 (en) * 2001-09-28 2010-04-27 Emc Corporation Storage switch for storage area network
US7185062B2 (en) * 2001-09-28 2007-02-27 Emc Corporation Switch-based storage services
US7404000B2 (en) * 2001-09-28 2008-07-22 Emc Corporation Protocol translation in a storage system
JP2005505819A (en) * 2001-09-28 2005-02-24 マランティ ネットワークス インコーポレイテッド Packet classification in storage systems
US7543087B2 (en) 2002-04-22 2009-06-02 Alacritech, Inc. Freeing transmit memory on a network interface device prior to receiving an acknowledgement that transmit data has been received by a remote device
US7398326B2 (en) 2002-04-25 2008-07-08 International Business Machines Corporation Methods for management of mixed protocol storage area networks
US7478169B2 (en) 2003-10-16 2009-01-13 International Business Machines Corporation Accessing data processing systems behind a NAT enabled network
JP2005339323A (en) 2004-05-28 2005-12-08 Hitachi Ltd Storage system, computing system, and interface module
US9137175B2 (en) * 2007-12-19 2015-09-15 Emulex Corporation High performance ethernet networking utilizing existing fibre channel fabric HBA technology
US8539513B1 (en) 2008-04-01 2013-09-17 Alacritech, Inc. Accelerating data transfer in a virtual computer system with tightly coupled TCP connections
US8341286B1 (en) 2008-07-31 2012-12-25 Alacritech, Inc. TCP offload send optimization
US9306793B1 (en) 2008-10-22 2016-04-05 Alacritech, Inc. TCP offload device that batches session layer headers to reduce interrupts as well as CPU copies
EP2432164B1 (en) 2009-05-11 2017-03-29 Fujitsu Limited Node device and communication method
WO2013049710A1 (en) * 2011-09-29 2013-04-04 Agiledelta, Inc. Interface-adaptive data exchange
JP5035480B1 (en) 2012-01-27 2012-09-26 オムロン株式会社 Data relay device, data transmission device, network system

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2694605B2 (en) * 1995-03-15 1997-12-24 株式会社超高速ネットワーク・コンピュータ技術研究所 ATM cell transmission method and network system
JP3402996B2 (en) * 1997-03-05 2003-05-06 ブラザー工業株式会社 Network connection device
US6000020A (en) * 1997-04-01 1999-12-07 Gadzoox Networks, Inc. Hierarchical storage management from a mirrored file system on a storage network segmented by a bridge
US5941972A (en) * 1997-12-31 1999-08-24 Crossroads Systems, Inc. Storage router and method for providing virtual local storage

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO0159966A1 *

Also Published As

Publication number Publication date
WO2001059966A1 (en) 2001-08-16
JP2004503122A (en) 2004-01-29
AU2000238787A1 (en) 2001-08-20

Similar Documents

Publication Publication Date Title
US6400730B1 (en) Method and apparatus for transferring data between IP network devices and SCSI and fibre channel devices over an IP network
TWI446755B (en) A method for interfacing a fibre channel network with an ethernet based network
US7233570B2 (en) Long distance repeater for digital information
US7921240B2 (en) Method and system for supporting hardware acceleration for iSCSI read and write operations and iSCSI chimney
WO2001059966A1 (en) Method and apparatus for transferring data between different network devices over an ip network
US7599289B2 (en) Electronic communication control
US7640364B2 (en) Port aggregation for network connections that are offloaded to network interface devices
US6988161B2 (en) Multiple port allocation and configurations for different port operation modes on a host
US7145866B1 (en) Virtual network devices
US7020715B2 (en) Protocol stack for linking storage area networks over an existing LAN, MAN, or WAN
US9219683B2 (en) Unified infrastructure over ethernet
US8180928B2 (en) Method and system for supporting read operations with CRC for iSCSI and iSCSI chimney
US20150039792A1 (en) ACCELERATING INTERNET SMALL COMPUTER SYSTEM INTERFACE (iSCSI) Proxy Input/Output (I/O)
EP1759317B1 (en) Method and system for supporting read operations for iscsi and iscsi chimney
US7099955B1 (en) End node partitioning using LMC for a system area network
Krueger et al. Small computer systems interface protocol over the Internet (iSCSI) requirements and design considerations
US20030016669A1 (en) Full transmission control protocol off-load
US20050283545A1 (en) Method and system for supporting write operations with CRC for iSCSI and iSCSI chimney
US20050281261A1 (en) Method and system for supporting write operations for iSCSI and iSCSI chimney
US7155537B1 (en) Infiniband isolation bridge merged with architecture of an infiniband translation bridge
EP1158750B1 (en) Systems and method for peer-level communications with a network interface card
Munson Introduction To Fibre Channel Connectivity.
Krueger et al. RFC3347: Small Computer Systems Interface protocol over the Internet (iSCSI) Requirements and Design Considerations
Guendert Fibre Channel Standard
Bakke Network Working Group M. Krueger Request for Comments: 3347 R. Haagens Category: Informational Hewlett-Packard Corporation C. Sapuntzakis Stanford

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20020903

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Free format text: AL PAYMENT 20020903;LT PAYMENT 20020903;LV PAYMENT 20020903;MK PAYMENT 20020903;RO PAYMENT 20020903;SI PAYMENT 20020903

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20080401