EP1502179A2 - Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure - Google Patents

Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure

Info

Publication number
EP1502179A2
EP1502179A2 EP02797469A EP02797469A EP1502179A2 EP 1502179 A2 EP1502179 A2 EP 1502179A2 EP 02797469 A EP02797469 A EP 02797469A EP 02797469 A EP02797469 A EP 02797469A EP 1502179 A2 EP1502179 A2 EP 1502179A2
Authority
EP
European Patent Office
Prior art keywords
virtual
port
virtual enclosure
virtualization
area network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
EP02797469A
Other languages
German (de)
French (fr)
Inventor
Sanjaya Kumar
Manas Barooah
John L. Burnett
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Cisco Technology Inc
Original Assignee
Cisco Technology Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cisco Technology Inc filed Critical Cisco Technology Inc
Publication of EP1502179A2 publication Critical patent/EP1502179A2/en
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F2003/0697Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers device management, e.g. handlers, drivers, I/O schedulers

Definitions

  • the present invention relates to network technology. More particularly, the present invention relates to methods and apparatus for supporting virtualization of storage within a storage area network.
  • SAN storage area network
  • a storage area network is a high-speed special-purpose network that interconnects different data storage devices and associated data hosts on behalf of a larger network of users.
  • a SAN enables a storage device to be configured for use by various network devices and/or entities within a network, data storage needs are often dynamic rather than static.
  • FIG. 1 A illustrates an exemplary conventional storage area network. More specifically, within a storage area network 102, it is possible to couple a set of hosts (e.g., servers or workstations) 104, 106, 108 to a pool of storage devices (e.g., disks). In SCSI parlance, the hosts may be viewed as "initiators" and the storage devices may be viewed as "targets.”
  • a storage pool may be implemented, for example, through a set of storage arrays or disk arrays 110, 112, 114. Each disk array 110, 112, 114 further corresponds to a set of disks.
  • first disk array 110 corresponds to disks 116, 118
  • second disk array 112 corresponds to disk 120
  • third disk array 114 corresponds to disks 122, 124.
  • storage e.g., disks
  • physical memory e.g., physical disks
  • virtual memory e.g., virtual disks
  • Virtual memory has traditionally been used to enable physical memory to be virtualized through the translation between physical addresses in physical memory and virtual addresses in virtual memory. Recently, the concept of "virtualization" has been implemented in storage area networks through various mechanisms. Virtualization interconverts physical storage and virtual storage on a storage network. The hosts (initiators) see virtual disks as targets. The virtual disks represent available physical storage in a defined but somewhat flexible manner. Virtualization provides hosts with a representation of available physical storage that is not constrained by certain physical arrangements/allocation of the storage.
  • Redundant Array of Independent Disks provides some limited features of virtualization.
  • Various RAID subtypes have been implemented.
  • RAEDl a virtual disk may correspond to two physical disks 116, 118 which both store the same data (or otherwise support recovery of the same data), thereby enabling redundancy to be supported within a storage area network.
  • RAIDO a single virtual disk is striped across multiple physical disks.
  • Some other types of virtualization include concatenation, sparing, etc.
  • Virtualization in the storage array is one of the most common storage virtualization solutions in use today. Through this approach, virtual volumes are created over the storage space of a specific storage subsystem (e.g., disk array). Creating virtual volumes at the storage subsystem level provides host independence, since virtualization of the storage pool is invisible to the hosts. In addition, virtualization at the storage system level enables optimization of memory access and therefore high performance. However, such a virtualization scheme typically will allow a uniform management structure only for a homogenous storage environment and even then only with limited flexibility. Further, since virtualization is performed at the storage subsystem level, the physical-virtual limitations set at the storage subsystem level are imposed on all hosts in the storage area network. Moreover, each storage subsystem (or disk array) is managed independently. Virtualization at the storage level therefore rarely allows a virtual volume to span over multiple storage subsystems (e.g., disk arrays), thus limiting the scalability of the storage-based approach.
  • a specific storage subsystem e.g., disk array
  • a host-based approach has an additional advantage, in that a limitation on one host does not impact the operation of other hosts in a storage area network.
  • virtualization at the host-level requires the existence of a software layer running on each host (e.g., server) that implements the virtualization function. Running this software therefore impacts the performance of the hosts running this software.
  • Another key difficulty with this method is that it assumes a prior partitioning of the available storage to the various hosts. Since such partitioning is supported at the host-level and the virtualization function of each host is performed independently of the other hosts in the storage area network, it is difficult to coordinate storage access across the hosts.
  • the host-based approach therefore fails to provide an adequate level of security. Due to this security limitation, it is difficult to implement a variety of redundancy schemes such as RAID which require the "locking" of memory during read and write operations. In addition, when mirroring is performed, the host must replicate the data multiple times, increasing its input-output and CPU load, and increasing the traffic over the SAN. Virtualization in a storage area network appliance placed between the hosts and the storage solves some of the difficulties of the host-based and storage-based approaches.
  • the storage appliance globally manages the mapping and allocation of physical storage to virtual volumes. Typically, the storage appliance manages a central table that provides the current mapping of physical to virtual.
  • the storage appliance-based approach enables the virtual volumes to be implemented independently from both the hosts and the storage subsystems on the storage area network, thereby providing a higher level of security. Moreover, this approach supports virtualization across multiple storage subsystems.
  • the key drawback of many implementations of this architecture is that every input/output (I/O) of every host must be sent through the storage area network appliance, causing significant performance degradation and a storage area network bottleneck. This is particularly disadvantageous in systems supporting a redundancy scheme such as RAID, since data must be mirrored across multiple disks.
  • the appliance makes sure that all hosts receive the current version of the table.
  • Methods and apparatus for implementing virtualization of storage in a storage area network are disclosed. This is accomplished through the use of one or more network devices capable of being placed in a data path between the hosts and the storage devices. As a result, neither the storage devices nor the hosts require additional software or hardware to support storage virtualization. Thus, the present invention is superior to the host based approach, which requires that each host be burdened by additional software to implement virtualization functionality. Moreover, the present invention enables multiple network devices to simultaneously manage the virtualization of various storage devices.
  • switch-based virtualization may be implemented on a per port basis. Any number of ports on a switch can manage virtualization of its own traffic. This allows a network's virtualization capacity to scale with the number of ports. Since there are large numbers of ports in any network system, there will nearly always be sufficient bandwidth for virtualization. Accordingly, virtualization of storage may be achieved without many of the drawbacks present in conventional virtualization schemes.
  • a virtual enclosure is created that has one or more virtual enclosure ports and is adapted for representing one or more virtual storage units.
  • the virtual enclosure serves to "enclose" selected virtual storage units, which may be accessed via the virtual enclosure ports.
  • Each of the virtual storage units represents one or more physical storage locations on one or more physical storage units of the storage area network.
  • each of the virtual enclosure ports of the virtual enclosure is associated with a port of a network device within the storage area network.
  • An address or identifier is then assigned to each of the virtual enclosure ports.
  • the address or identifier may be a Fibre Channel identifier (FCID).
  • FCID Fibre Channel identifier
  • a virtual enclosure is implemented within a Fibre channel network.
  • NWWN Node World Wide Name
  • PWWN Port World Wide Name
  • a port of a network device within the storage area network is instructed to handle messages on behalf of a virtual enclosure port. This may be accomplished in two ways. First, the port may be instructed to "bind" itself to the virtual enclosure port. In other words, the port acts as the virtual enclosure port, and all messages directed to the virtual enclosure port and received by the port are handled by that port. Second, the port may be instructed to serve as a "trapping port.” More particularly, in addition to the port that is bound to the virtual enclosure port, one or more additional ports may also handle messages they receive that are directed to the virtual enclosure port.
  • a trapping port is preferably a port that is directly connected to a host, and therefore can track those requests received by it as well as the responses associated with those requests. Binding and trapping among multiple ports on behalf of a single virtual enclosure port is preferably coordinated at a central location such as a virtual enclosure server.
  • Various network devices may be configured or adapted for performing the disclosed virtualization processes. These network devices include, but are not limited to, servers (e.g., hosts), routers, and switches. Moreover, the functionality for the above-mentioned virtualization processes may be implemented in software as well as hardware.
  • Yet another aspect of the invention pertains to computer program products including machine-readable media on which are provided program instructions for implementing the methods and techniques described above, in whole or in part. Any of the methods of this invention may be represented, in whole or in part, as program instructions that can be provided on such machine-readable media.
  • the invention pertains to various combinations and arrangements of data generated and/or used as described herein. For example, packets and frames having the format described herein and provided on appropriate media are part of this invention.
  • FIG. 1 A is a block diagram illustrating an exemplary conventional storage area network capable of implementing various embodiments of prior art virtualization functions.
  • FIG. IB is a block diagram illustrating an exemplary storage area network in which various embodiments of the invention may be implemented.
  • FIG. 2 is a block diagram illustrating a virtualization model that may be implemented in accordance with various embodiments of the invention.
  • FIG. 3A is a block diagram illustrating an exemplary virtualization switch in which various embodiments of the present invention may be implemented.
  • FIG. 3B is a block diagram illustrating an exemplary standard switch in which various embodiments of the present invention may be implemented.
  • FIG. 4 is a transaction flow diagram illustrating a conventional method of implementing a node world wide name (NWWN) and port world wide name (PWWN) for each SCSI target port.
  • NWWN node world wide name
  • PWWN port world wide name
  • FIG. 5 is a block diagram illustrating an exemplary virtual enclosure in accordance with one embodiment of the invention.
  • FIG. 6 is a diagram illustrating an exemplary system in which a virtual enclosure is implemented through the binding of the virtual enclosure ports to various virtualization ports in accordance with various embodiments of the invention.
  • FIG. 7 is a process flow diagram illustrating a method of creating a virtual enclosure in accordance with various embodiments of the invention.
  • FIG. 8 is a diagram illustrating a conventional fibre channel identifier (FCID) that may be associated with a virtual enclosure port in accordance with various embodiments of the invention.
  • FCID fibre channel identifier
  • FIG. 9 is a transaction flow diagram illustrating one method of coordinating virtual enclosure binding and trapping functionality of virtualization ports in accordance with various embodiments of the invention.
  • FIG. 10 is a diagram illustrating an exemplary table that may be maintained by a virtualization port indicating FCIDs to be handled by the virtualization port in accordance with various embodiments of the invention.
  • FIG. 11 is a transaction flow diagram illustrating one method of establishing communication between a host and one or more virtualization ports (e.g., virtual enclosure ports, trapping ports) such that the host can access one or more LUNs in accordance with various embodiments of the invention.
  • virtualization ports e.g., virtual enclosure ports, trapping ports
  • FIG. 12 is a diagram illustrating an exemplary LUN mapping table that may be used at step 1128 of FIG. 11 to perform LUN mapping.
  • FIG. 13 is a diagram representing a virtual enclosure corresponding to the LUN mapping table of FIG. 12.
  • virtualization of storage within a storage area network may be implemented ⁇ through the creation of a virtual enclosure having one or more virtual enclosure ports.
  • the virtual enclosure is implemented, in part, by one or more network devices, which will be referred to herein as virtualization switches.
  • a virtualization switch or more specifically, a virtualization port within the virtualization switch, may handle messages such as packets or frames on behalf of one of the virtual enclosure ports.
  • embodiments of the invention may be applied to a packet or frame directed to a virtual enclosure port, as will be described in further detail below.
  • switches act on frames and use information about SANs to make switching decisions.
  • the frames being received and transmitted by a virtualization switch possess the frame format specified for a standard protocol such as Ethernet or fibre channel.
  • software and hardware conventionally used to generate such frames may be employed with this invention.
  • Additional hardware and/or software is employed to modify and/or generate frames compatible with the standard protocol in accordance with this invention.
  • the appropriate network devices should be configured with the appropriate software and/or hardware for performing virtualization functionality.
  • all network devices within the storage area network need not be configured with the virtualization functionality. Rather, selected switches and or ports may be configured with or adapted for virtualization functionality. Similarly, in various embodiments, such virtualization functionality may be enabled or disabled through the selection of various modes.
  • the standard protocol employed in the storage area network i.e., the protocol used to frame the data
  • the type of traffic is defined in some encapsulation formats. Examples of the type of traffic are typically layer 2 or corresponding layer formats such as Ethernet, Fibre channel, and InfiniBand.
  • a storage area network is a high-speed special- purpose network that interconnects different data storage devices with associated network hosts (e.g., data servers or end user machines) on behalf of a larger network of users.
  • a SAN is defined by the physical configuration of the system. In other words, those devices in a SAN must be physically interconnected.
  • various storage devices 132, 134, 136, 138, 140, and 142 may be implemented, which may be homogeneous (e.g., identical device types, sizes, or configurations) as well as heterogeneous (e.g., different device types, sizes or configurations). Data may be read from, as well as written to, various portions of the storage devices 132-142 in response to commands sent by hosts 144 and 146. Communication among the storage devices and hosts is accomplished by coupling the storage devices and hosts together via one or more switches, routers, or other network nodes configured to perform a switching function. In this example, switches 148, 150, and 152 communicate with one another via interswitch links 154 and 156.
  • this invention pertains to "virtualization" in storage networks.
  • virtualization in this invention is implemented through the creation and implementation of a virtual enclosure. This is accomplished, in part, through the use of switches or other "interior" network nodes of a storage area network to implement the virtual enclosure.
  • the virtualization of this invention typically is implemented on a per port basis. In other words, a multi-port virtualization switch will have virtualization separately implemented on one or more of its ports. Individual ports have dedicated logic for handing the virtualization functions for packets or frames handled by the individual ports. This allows virtualization processing to scale with the number of ports, and provides far greater bandwidth for virtualization than can be provided with host based or storage based virtualization schemes. In such prior art approaches the number of connections between hosts and the network fabric or between storage nodes and the network fabric are limited - at least in comparison to the number of ports in the network fabric.
  • the virtualization logic is separately implemented at individual ports of a given switch - rather than having centralized processing for all ports of a switch. This allows the virtualization processing capacity to be closely matched with the exact needs of the switch (and the virtual enclosure) on a per port basis. If a central processor is employed for the entire switch (serving numerous ports), the processor must be designed/selected to handle maximum traffic at all ports. For many applications, this represents extremely high processing requirements and a very large/expensive processor. If the central processor is too small, the switch will at times be unable to keep up with the switching/virtualization demands of the network.
  • Virtualization may take many forms. In general, it may be defined as logic or procedures that inter-relate physical storage and virtual storage on a storage network. Hosts see a representation of available physical storage that is not constrained by the physical arrangements or allocations inherent in that storage.
  • One example of a physical constraint that is transcended by virtualization includes the size and location of constituent physical storage blocks. For example, logical units as defined by the Small Computer System Interface (SCSI) standards come in precise physical sizes (e.g., 36GB and 72GB).
  • SCSI Small Computer System Interface
  • Virtualization can represent storage in virtual logical units that are smaller or larger than the defined size of a physical logical unit. Further, virtualization can present a virtual logical unit comprised of regions from two or more different physical logical units, sometimes provided on devices from different vendors.
  • the virtualization operations are transparent to at least some network entities (e.g., hosts).
  • network entities e.g., hosts.
  • virtualization on a storage area network is similar to virtual memory on a typical computer system.
  • Virtualization on a network brings far greater complexity and far greater flexibility. The complexity arises directly from the fact that there are a number of separately interconnected network nodes. Virtualization must span these nodes.
  • the nodes include hosts, storage subsystems, and switches (or comparable network traffic control devices such as routers).
  • the hosts and/or storage subsystems are heterogeneous, being provided by different vendors. The vendors may employ distinctly different protocols (standard protocols or proprietary protocols).
  • virtualization provides the ability to connect heterogeneous initiators (e.g., hosts or servers) to a distributed, heterogeneous set of targets (storage subsystems), enabling the dynamic and transparent allocation of storage.
  • Examples of network specific virtualization operations include the following: RAID 0 through RAID 5, concatenation of memory from two or more distinct logical units of physical memory, sparing (auto-replacement of failed physical media), remote mirroring of physical memory, logging information (e.g., errors and/or statistics), load balancing among multiple physical memory systems, striping (e.g., RAID 0), security measures such as access control algorithms for accessing physical memory, resizing of virtual memory blocks, Logical Unit (LUN) mapping to allow arbitrary LUNs to serve as boot devices, backup of physical memory (point in time copying), and the like.
  • LUN Logical Unit
  • FC-PH Fibre channel - Physical and Signaling Interface
  • an "initiator” is a device (usually a host system) that requests an operation to be performed by another device.
  • a host initiator will request a read or write operation be performed on a region of virtual or physical memory.
  • a target is a device that performs an operation requested by an initiator.
  • a target physical memory disk will obtain or write data as initially requested by a host initiator.
  • the host initiator may provide instructions to read from or write to a "virtual" target having a virtual address
  • a virtualization switch of this invention must first convert those instructions to a physical target address before instructing the target.
  • Targets may be divided into physical or virtual "logical units.” These are specific devices addressable through the target.
  • a physical storage subsystem may be organized in a number of distinct logical units.
  • hosts view virtual memory as distinct virtual logical units.
  • logical units will be referred to as "LUNs.”
  • LUN refers to a logical unit number. But in common parlance, LUN also refers to the logical unit itself.
  • a virtualization model This is the way in which physical storage provided on storage subsystems (such as disk arrays) is related to a virtual storage seen by hosts or other initiators on a network. While the relationship may take many forms and be characterized by various terms, a SCSI-based terminology will be used, as indicated above.
  • the physical side of the storage area network will be described as a physical LUN.
  • the host side sees one or more virtual LUNs, which are virtual representations of the physical LUNs.
  • the mapping of physical LUNs to virtual LUNs may logically take place over one, two, or more levels. In the end, there is a mapping function that can be used by switches of this invention to interconvert between physical LUN addresses and virtual LUN addresses.
  • FIG. 2 is a block diagram illustrating an example of a virtualization model that may be implemented within a storage area network in accordance with various embodiments of the invention.
  • the physical storage of the storage area network is made up of one or more physical LUNs, shown here as physical disks 202.
  • Each physical LUN is a device that is capable of containing data stored in one or more contiguous blocks which are individually and directly accessible.
  • each block of memory within a physical LUN may be represented as a block 204, which may be referred to as a disk unit (DUnit).
  • DUnit disk unit
  • mapping function 206 it is possible to convert physical LUN addresses associated with physical LUNs 202 to virtual LUN addresses, and vice versa. More specifically, as described above, the virtualization and therefore the mapping function may take place over one or more levels. For instance, as shown, at a first virtualization level, one or more virtual LUNs 208 each represents one or more physical LUNs 202, or portions thereof. The physical LUNs 202 that together make up a single virtual LUN 208 need not be contiguous. Similarly, the physical LUNs 202 that are mapped to a virtual LUN 208 need not be located within a single target. Thus, through virtualization, virtual LUNs 208 may be created that represent physical memory located in physically distinct targets, which may be from different vendors, and therefore may support different protocols and types of traffic.
  • a second virtualization level within the virtualization model of FIG. 2 is referred to as a high-level VLUN or volume 210.
  • the initiator device "sees" only
  • VLUN 210 when accessing data.
  • multiple VLUNs are "enclosed” within a virtual enclosure such that only the virtual enclosure may be “seen” by the initiator. In other words, the VLUNs enclosed by the virtual enclosure are not visible to the initiator.
  • VLUN 210 is implemented as a "logical" RAID array of virtual LUNs 208. Moreover, such a virtualization level may be further implemented, such as through the use of striping and/or mirroring. In addition, it is important to note that it is unnecessary to specify the number of virtualization levels to support the mapping function 206. Rather, an arbitrary number of levels of virtualization may be supported, for example, through a recursive mapping function. For instance, various levels of nodes may be built and maintained in a tree data structure, linked list, or other suitable data structure that can be traversed.
  • Each initiator may therefore access physical LUNs via nodes located at any of the levels of the hierarchical virtualization model.
  • Nodes within a given virtualization level of the hierarchical model implemented within a given storage area network may be both visible to and accessible to an allowed set of initiators (not shown). However, in accordance with various embodiments of the invention, these nodes are enclosed in a virtual enclosure, and are therefore no longer visible to the allowed set of initiators.
  • Nodes within a particular virtualization level e.g., VLUNs
  • functions e.g., read, write
  • various initiators may be assigned read and/or write privileges with respect to particular nodes (e.g., VLUNs) within a particular virtualization level. In this manner, a node within a particular virtualization level may be accessible by selected initiators.
  • VLUNs virtualization level
  • FIG. 3A is a block diagram illustrating an exemplary virtualization switch in which various embodiments of the present invention may be implemented.
  • data or messages are received by an intelligent, virtualization port via a bi-directional connector 302.
  • the virtualization port is adapted for handling messages on behalf of a virtual enclosure port, as will be described in further detail below.
  • Media Access Control (MAC) block 304 is provided, which enables frames of various protocols such as Ethernet or fibre channel to be received.
  • MAC Media Access Control
  • a virtualization intercept switch 306 determines whether an address specified in an incoming frame pertains to access of a virtual storage location of a virtual storage unit representing one or more physical storage locations on one or more physical storage units of the storage area network.
  • the virtual storage unit may be a virtual storage unit (e.g., VLUN) that is enclosed within a virtual enclosure.
  • the frame is processed by a virtualization processor 308 capable of performing a mapping function such as that described above. More particularly, the virtualization processor 308 obtains a virtual- physical mapping between the one or more physical storage locations and the virtual storage location. In this manner, the virtualization processor 308 may look up either a physical or virtual address, as appropriate. For instance, it may be necessary to perform a mapping from a physical address to a virtual address or, alternatively, from a virtual address to one or more physical addresses.
  • the virtualization processor 308 may then employ the obtained mapping to either generate a new frame or modify the existing frame, thereby enabling the frame to be sent to an initiator or a target specified by the virtual-physical mapping.
  • the mapping function may also specify that the frame needs to be replicated multiple times, such as in the case of a mirrored write. More particularly, the source address and/or destination addresses are modified as appropriate. For instance, for data from the target, the virtualization processor replaces the source address, which was originally the physical LUN address with the corresponding virtual LUN and address. In the destination address, the port replaces its own address with that of the initiator.
  • the port For data from the initiator, the port changes the source address from the initiator's address to the port's own address. It also changes the destination address from the virtual LUN/address to the corresponding physical LUN/address.
  • the new or modified frame may then be provided to the virtualization intercept switch 306 to enable the frame to be sent to its intended destination.
  • the frame or associated data may be stored in a temporary memory location (e.g., buffer) 310.
  • a temporary memory location e.g., buffer
  • this data may be received in an order that is inconsistent with the order in which the data should be transmitted to the initiator of the read command.
  • the new or modified frame is then received by a forwarding engine 312, which obtains information from various fields of the frame, such as source address and destination address.
  • the forwarding engine 312 then accesses a forwarding table 314 to determine whether the source address has access to the specified destination address. More specifically, the forwarding table 314 may include physical LUN addresses as well as virtual LUN addresses.
  • the forwarding engine 312 also determines the appropriate port of the switch via which to send the frame, and generates an appropriate routing tag for the frame.
  • the frame will be received by a buffer queuing block 316 prior to transmission. Rather than transmitting frames as they are received, it may be desirable to temporarily store the frame in a buffer or queue 318.
  • the frame is then transmitted via switch fabric 320 to the appropriate port.
  • the outgoing port has its own MAC block 322 and bi-directional connector 324 via which the frame may be transmitted.
  • FIG. 3B is a block diagram illustrating an exemplary standard switch in which various embodiments of the present invention may be implemented.
  • a standard port 326 has a MAC block 304.
  • a virtualization intercept switch and virtualization processor such as those illustrated in FIG. 3A are not implemented.
  • a frame that is received at the incoming port is merely processed by the forwarding engine 312 and its associated forwarding table 314.
  • a frame may be queued 316 in a buffer or queue 318. Frames are then forwarded via switch fabric 320 to an outgoing port.
  • the outgoing port also has an associated MAC block 322 and bi-directional connector 324.
  • each port may support a variey of protocols.
  • the outgoing port may be an iSCSI port (i.e. a port that supports SCSI over IP over Ethernet), which also supports virtualization, as well as parallel SCSI and serial SCSI.
  • network devices described above with reference to FIG. 3 A and 3B are described as switches, these network devices are merely illustrative. Thus, other network devices such as routers may be implemented to receive, process, modify and/or generate packets or frames with functionality such as that described above for transmission in a storage area network. Moreover, the above- described network devices are merely illustrative, and therefore other types of network devices may be implemented to perform the disclosed virtualization functionality.
  • SCSI targets are directly accessible by SCSI initiators (e.g., hosts).
  • SCSI targets such as PLUNs are visible to the hosts that are accessing those SCSI targets.
  • VLUNs are visible and accessible to the SCSI initiators.
  • each host must typically identify those VLUNs that are available to it. More specifically, the host typically determines which SCSI target ports are available to it. The host may then ask each of those SCSI target ports which VLUNs are available via those SCSI target ports.
  • Fibre channel devices have a World Wide Name (WWN). More specifically, a Node WWN (NWWN) is the WWN of the node that is connected to a particular port. In other words, the NWWN is the WWN of the system, storage device, or subsystem that is connected to the switch port. In addition to a Node WWN, a Port WWN (PWWN) serves as a given name for a particular port. A Fibre channel network ID (FCID) for the particular switch port is used to identify the physical location of a port. Each Fibre channel device may have multiple ports, each of which is uniquely identified by a NWWN and a PWWN.
  • FCID Fibre channel network ID
  • a NWWN and a PWWN must be allocated for each port of a Fibre channel device.
  • the number of available ports is limited to the number of ports in a particular storage device, subsystem or switch.
  • the storage device, subsystem or switch may have a greater number of ports than are needed in a particular storage virtualization scheme.
  • virtualization ports are expensive to implement, and it is therefore undesirable to "waste" these intelligent ports.
  • virtualization ports may be used on an as-needed basis. In other words, ports may be selected for implementing one or more virtual enclosures, either by binding those ports to virtual enclosure ports or by implementing these ports as trapping ports. As a result, those ports that are most expensive to produce may be used to their maximum capacity.
  • virtualization in a network such as a fibre channel network is implemented without limiting the number of accessible ports in a storage subsystem.
  • ports within a SAN may selected and allocated on an as-needed basis for a particular storage virtualization scheme.
  • FIG. 4 is a transaction flow diagram illustrating a conventional method of implementing a node world wide name (NWWN) and port world wide name (PWWN) for each SCSI target port.
  • NWWN node world wide name
  • PWWN port world wide name
  • interactions between a host 402, switch 404, Domain Name System (DNS) server 406 and SCSI target port 408 are represented by corresponding labeled vertical lines.
  • DNS Domain Name System
  • FLOGI fabric login
  • Fibre channel uses an address identifier, referred to as an FCID, which is dynamically assigned during login.
  • FCID address identifier
  • the switch 404 provides an FCID to the host 402 at 412.
  • the host may perform a DNS query of a DNS server 406 via a SCSI REPORT command, which requests the SCSI targets that are available, and therefore visible, to the host at block 414. More specifically, the DNS query determines those SCSI targets that are visible to the host FCID.
  • the FCID of a SCSI target port is then provided by the DNS server 406 to the host 402 at 416.
  • the host Once the host has the FCIDs of those SCSI target ports available to it, it sends a Fibre channel process login command to one of the available SCSI target ports (identified by its FCID) at 418. This process login implements a mapping layer, which "maps" Fibre channel to SCSI. Once completed, the SCSI target port 408 sends a Fibre channel accept message at 420. Since Fibre channel is mapped to SCSI, the host can send SCSI commands to the SCSI target.
  • the host 402 can send SCSI commands to the SCSI target port 408, it performs a SCSI process login at 422 by sending a SCSI process login command to the SCSI target port 408.
  • the SCSI target port 408 then sends a SCSI accept command 424. Communication between the host and the SCSI target via the SCSI protocol is therefore established. For instance, the host may determine which LUNs are available to it, as well as read and write to those LUNs available to it.
  • the host 402 Since the host 402 knows which SCSI target ports are available to it and can communicate with each of these ports, it can send a SCSI REPORT LUN command to the SCSI target port 408 to determine those LUNS that are visible to the host FCID at 426.
  • the SCSI target port 408 determines which LUNs are visible to the host FCID and sends a list of LUNs (e.g., PLUNs or VLUNs) at 428.
  • the host 402 may then send SCSI READ and WRITE commands to a particular PLUN or VLUN at 430 via the SCSI target port 408.
  • FIG. 5 is a block diagram of an exemplary virtual enclosure in accordance with one embodiment of the invention.
  • a virtual enclosure 502 is a virtual entity that is that is adapted for "enclosing" or representing one or more virtual storage units.
  • one or more of the virtual storage units may each comprise a VLUN or other virtual representation of storage on a SAN.
  • each virtual storage unit represents one or more physical storage locations on one or more physical storage units 504 of the storage area network.
  • various virtualization ports such as those described above may be implemented on an as-needed basis to support virtualization within a particular storage area network.
  • the virtual enclosure has one or more "virtual enclosure ports,” each of which is associated with a port (e.g., virtualization port) of a network device within the storage area network.
  • a port e.g., virtualization port
  • the port of the network device may be instructed, either statically or dynamically, to handle messages addressed to the virtual enclosure port.
  • An address or identifier is also assigned to each of the virtual enclosure ports.
  • a message e.g., packet or frame
  • a FCID may be assigned to each of the virtual enclosure ports for use in a fibre channel network, or with various fibre channel devices within a network such as a SAN. In this manner, a virtual enclosure is created for representing one or more virtual storage units.
  • VSAN virtual storage area networks
  • Each virtual enclosure may have any number of virtual enclosure ports. In other words, it may be desirable to enable the number of virtual enclosure ports to be selectable. For instance, a system administrator responsible for creating and maintaining the virtual enclosure definitions may select the appropriate number of virtual enclosure ports. It is important to note that the number of virtual enclosure ports within a virtual enclosure is not limited by the number of ports within a particular network device within the storage area network. More specifically, each virtual enclosure port is associated with a port of a network device within the storage area network. Li other words, the virtual enclosure ports of a single virtual enclosure may be simultaneously associated with ports (e.g., virtualization ports) of multiple, different network devices within the storage area network rather than a single network device. In addition, a single port of a network device such as a virtualization port may be simultaneously associated with or bound to multiple virtual enclosure ports. Accordingly, the number of ports that are available for virtualization of virtual storage units is virtually unlimited.
  • an address or identifier is assigned to each of the virtual enclosure ports.
  • this address or identifier may be an address or identifier that has been previously assigned to the port of the storage area network that is later associated with the virtual enclosure port.
  • an address or identifier previously assigned to the port e.g., virtualization port
  • the address or identifier is assigned to the virtual enclosure port to uniquely identify the virtual enclosure port.
  • the address or identifier is provided to the virtualization port that is associated with the virtual enclosure port.
  • the virtual enclosure is used in a fibre channel network, and therefore each fibre channel device has a NWWN.
  • both a NWWN and a PWWN together may be used as an address or identifier to uniquely identify a virtual enclosure port. More particularly, a NWWN is associated with the virtual enclosure. In addition, a PWWN is assigned to each virtual enclosure port. The NWWN and PWWN together identify the virtual enclosure port until it is assigned an address or identifier such as an FCID.
  • each of the virtual storage units may be a VLUN or other virtual representation of storage on the storage area network.
  • the virtual enclosure "encloses” those virtual storage units, thereby requiring a host to access those virtual storage units via the virtual enclosure.
  • the VLUNS may be "hidden” from the host.
  • the host merely has access to the virtual enclosure rather than a specific device or VLUN.
  • the VLUNs appear as logical units behind a set of virtual enclosure ports.
  • the PLUNs that are used to create the VLUNs are hidden from the host.
  • FIG. 6 is a diagram illustrating an exemplary system in which a virtual enclosure is implemented through the binding of the virtual enclosure ports to various virtualization ports in accordance with various embodiments of the invention.
  • a virtual enclosure 602 "encloses" two virtual storage units, VLUN1 604 and VLUN2 606.
  • the virtual enclosure 602 has three virtual enclosure ports, VE1 608, VE2 610, and VE3 612.
  • four network devices within the storage area network are illustrated.
  • Each of the virtualization switches has one or more virtualization ports as described above with reference to FIG. 3A. However, in order to simplify the illustration, each virtualization switch is shown to have a single virtualization port, labeled vl, v2, v3, and v4, respectively. Each virtualization port has an associated PWWN (not shown). Communication between the switches may be accomplished by an inter- switch link (not shown) between two switches (or two ports).
  • Each virtual enclosure port is uniquely identified. More specifically, within a fibre channel network, the virtual enclosure 602 is identified by a NWWN 621. In addition, once each virtual enclosure port is associated with a virtualization port, each of the virtual enclosure ports VE1 608, VE2 610, and
  • VE3 612 are further identified by a PWWN as well as its FCID, labeled PWWN1 622, PWWN2 624, and PWWN3 626, respectively. More particularly, the virtual enclosure ports may be identified by the NWWN and PWWN until an FCID is assigned to the virtual enclosure ports. An FCID will be further described below with reference to FIG. 8. As shown, the FCIDs that are assigned also identify the virtualization switch associated with the corresponding virtual enclosure port. This is because in Fibre Channel, a switch is seen as a "domain.” Thus, the first byte of the FCID of any port in that switch is the same as the domain ID.
  • the first byte of the FCID assigned to the first virtual enclosure port VE1 608 identifies the virtualization switch 616, shown as switch 50.
  • the first byte of the FCID assigned to the second virtual enclosure port VE2 610 identifies the virtualization switch 614, shown as switch 55, while the first byte of the FCID assigned to the third virtual enclosure port VE3 612 identifies the virtualization switch 618, shown as switch 45.
  • the creation of a virtual enclosure such as virtual enclosure 602 may be performed by a network device such as a virtual enclosure server 627.
  • the virtual enclosure 602 may be created, for example, by a system administrator.
  • various VLUNs such as VLUNs 604 and 606 may be assigned to the virtual enclosure 602.
  • Such an assignment may be subsequently modified as necessary to include additional VLUNs, remove VLUNs from the virtual enclosure, or otherwise modify the assignment of VLUNs to the virtual enclosure 602.
  • the virtual enclosure server 627 may also inform the virtualization ports of the virtualization switches of their association with the corresponding virtual enclosure ports. In this manner, the virtualization ports may be notified of their responsibility to intercept and handle packets (or frames) directed to those corresponding virtual enclosure ports.
  • one or more hosts 628, 630 may access data in the storage area network via the virtual enclosure 602. More specifically, the hosts 628, 630 may access various physical storage devices, referred to as physical storage units, corresponding to the virtual storage units, shown here as VLUN1 604 and VLUN2 606. For instance, a host may read data from or write data to various PLUNs 632, 634, 636, 638 within the storage area network by sending packets, frames or messages to a virtual address within a VLUN enclosed by a virtual enclosure (and available to the host via the virtual enclosure port). As described above, a virtual enclosure may be created for use in a variety of network environments. FIG.
  • FIG. 7 is a process flow diagram illustrating a method of creating a virtual enclosure in accordance with various embodiments of the invention.
  • the virtual enclosure is for use in a Fibre channel network.
  • a NWWN is associated with the virtual enclosure.
  • the number of virtual enclosure ports of the virtual enclosure is selected at block 704.
  • Each of these virtual enclosure ports may be identified by a PWWN until they are bound to a virtualization port at block 706 and assigned an FCID at block 708.
  • VLUNs are assigned to the virtual enclosure at block 710.
  • the host may access physical storage units via the virtual enclosure by sending SCSI read and write commands.
  • each SCSI target device is assigned a NWWN.
  • each SCSI target port is assigned a PWWN.
  • a Fibre channel identifier (FCID) identifies a physical location of each SCSI target port, and is therefore associated with the PWWN.
  • FIG. 8 is a diagram illustrating a conventional FCID that may be associated with a virtual enclosure port in accordance with various embodiments of the invention.
  • a FCID is a three byte address identifier, which identifies a switch 802 such as a virtualization switch, an area 804 within the network, and a port 806 of the switch 802 such as a virtualization port.
  • each virtual enclosure port is "bound" to a virtualization port.
  • coordination of binding virtual enclosure ports to virtualization ports is performed by a separate network device.
  • this network device may be the virtual enclosure server responsible for generating a virtual enclosure.
  • the virtual enclosure server instructs a virtualization port that it is "bound" to a virtual enclosure port and therefore should handle all messages addressed to that virtual enclosure port (that are received by the virtualization port).
  • it may be desirable to enable additional ports to "trap" messages addressed to the virtual enclosure port.
  • a virtualization port receiving a message addressed to a virtual enclosure port may be closer to the host than a virtualization port that is "bound" to the virtual enclosure port, and therefore may be better able to provide timely, efficient service.
  • it may be undesirable to route all messages to a single virtualization port that is bound to a virtual enclosure port, since this may overload a limited number of ports, resulting in ineffective service. Therefore, in accordance with various embodiments of the invention, the virtual enclosure server instructs various virtualization ports to "trap" messages addressed to one or more virtual enclosure ports. In this manner, the power of additional virtualization ports within the SAN to handle messages is leveraged. Accordingly, a virtualization port may service a virtual enclosure even though the virtualization port is not bound to a virtual enclosure port.
  • a "trapping" virtualization port is preferably an "external" node within the
  • the virtualization port should be directly connected to a host rather than indirectly connected via one or more other ports (e.g., standard or virtualization ports).
  • This direct connection is particularly important since an internal node may receive a message or request from a host, but the internal node may not receive messages from the target that are directed to the host, since the same return path is not guaranteed.
  • the virtualization port would not be able to complete this "communication loop," and therefore will not have knowledge of whether a host request has been serviced.
  • FIG. 9 One method of coordinating virtual enclosure binding and trapping functionality of virtualization ports in accordance with various embodiments of the invention is illustrated in FIG. 9. Steps performed at the virtual enclosure server are represented by vertical line 902, and steps performed at virtualization ports 1-3 are represented by vertical lines 904, 906, and 908, respectively.
  • a DNS server and associated switch are represented by vertical lines 912 and 910, respectively.
  • the DNS server 912 is often included in the switch 910.
  • the virtual enclosure server 902 may send various virtualization messages to various ports within the SAN to enable virtualization of storage within a SAN. For instance, as described above, a virtualization message may be a bind message or a trap message, which identifies the PWWN of a virtual enclosure port.
  • a bind message indicates that a FCID is to be dynamically assigned to a virtual enclosure port
  • a trap message indicates that the FCID that has previously been assigned to another virtual enclosure port is to be obtained so that messages directed to that FCID may be trapped by the receiving port.
  • the virtual enclosure server sends a bind message 914 instructing the first virtualization port, virtualization portl 904, to handle all messages directed to the first virtual enclosure port, VEP1 that are received by the virtualization portl 904.
  • Each virtual enclosure port such as VEP1 may be uniquely identified by its PWWN.
  • a virtualization message such as a bind or trap message may further indicate that the receiving virtualization port is to obtain an address or identifier assigned to the virtual enclosure port.
  • the virtualization message may indicate that the address or identifier is to be obtained from a DNS server, which may be specified or unspecified.
  • the virtualization portl 904 sends a FLOGI message to the switch 910 at 916 to establish a connection with the switch 910.
  • a FCID is dynamically assigned during login.
  • the switch 910 assigns a FCID, FCIDl, to the first virtualization port VEP1, which is provided to the DNS server 912 at 918, which then associates the virtual enclosure portl, VEP1 with the FCID FCIDl.
  • the switch 910 then sends an ACCEPT message at 920 indicating that the FCIDl is now associated by the DNS server 912 with the first virtual enclosure port, VEP1.
  • the ACCEPT message provides the FCID that has been assigned to the virtualization port 904.
  • a virtualization port keeps track of those virtual enclosure ports with which the virtualization port is bound or for which the virtualization port is responsible for trapping messages.
  • the virtualization port may then store the FCID, FCIDl, in a virtualization port table such as that described below with reference to FIG. 10.
  • a DNS REGISTRY is also performed at 922 to inform the DNS server 912 that VEPl understands the SCSI protocol.
  • the DNS server then sends an ACCEPT message at 924.
  • a virtualization response is then sent from the virtualization port at 926 to the virtual enclosure server indicating that the virtualization port is now configured to handle messages addressed to the virtual enclosure port of the virtual enclosure.
  • the virtualization response indicates that the virtualization port is now bound to the virtual enclosure port.
  • FCID assigned to the virtual enclosure port VEPl may also be provided to the virtual enclosure server so that the virtual enclosure server may assign the FCID to the virtual enclosure port. More specifically, the FCID may be provided in the virtualization response. In this manner, the virtual enclosure server obtains the FCIDs of SCSI target ports available to it.
  • the virtual enclosure server 902 may also establish trapping by additional virtualization ports within the SAN as described above. For instance, as shown at 928 a trap message indicating that the virtualization port2 906 is to handle messages addressed to an address or identifier assigned to a particular virtual enclosure port is sent. Since the virtual enclosure server 902 has obtained the FCID assigned to the virtual enclosure port, VEPl, it may be provided in the trap message. Alternatively, the FCID may be obtained from a DNS server, as will be described in this example. In this example, the virtualization port 906 is directed to handle messages addressed to the specified virtual enclosure port (e.g., VEPl) that are received by it.
  • the specified virtual enclosure port e.g., VEPl
  • the virtualization port 906 is instructed to handle messages addressed to the address or identifier assigned to the virtual enclosure port.
  • a second trap message is sent at 930 to the virtualization port3 908 indicating that the virtualization port is to handle messages addressed to the first virtual enclosure port.
  • the virtualization port2 906 When the virtualization port2 906 receives the trap message at 928, the virtualization port sends a GET_FCID message at 932 to the DNS server 912 to obtain the address or identifier assigned to the first virtual enclosure port, VEPl. Once it receives the address or identifier at 934, it stores the address or identifier at 936. More specifically, it may update a table such as that described below with reference to FIG. 10. The virtualization port then sends a virtualization response at 938 to the host in response to the virtualization message. For instance, the virtualization response may indicate that the virtualization port has been successfully (or unsuccessfully) configured to handle messages addressed to the virtual enclosure port. The virtualization port may now trap messages addressed to the virtual enclosure port that are received by it as shown at 940.
  • the virtualization port3 908 receives the trap message at 930, the virtualization port sends a GET_FCID message at 942 to the DNS server 912 to obtain the address or identifier assigned to the first virtual enclosure port, VEPl . Once it receives the address or identifier at 944, it stores the address or identifier at 946. The virtualization port sends a virtualization response at 948 indicating that the virtualization port has been successfully (or unsuccessfully) configured as a trapping port, and begins trapping messages directed to the virtualization port at 950.
  • FIG. 10 is a diagram illustrating an exemplary table that may be maintained by a virtualization port indicating FCIDs to be handled by the virtualization port in accordance with various embodiments of the invention.
  • a virtualization port table 1000 may be maintained by a virtualization port to keep track of those virtual enclosure ports 1002 with which it is bound or for which it is configured to trap messages. For each virtual enclosure port 1002, it maintains the associated FCID.
  • FIG. 11 is a transaction flow diagram illustrating one method of establishing communication between a host and one or more virtualization ports (e.g., bound to one or more virtual enclosure ports, trapping ports) such that the host can access one or more LUNs in accordance with various embodiments of the invention.
  • FCID address identifier
  • the switch 1104 provides an FCID to the host 1102 at 1112.
  • the host may perform a DNS query of a DNS server 1106 via a SCSI REPORT command, which requests the SCSI targets that are available, and therefore visible, to the host at block 1114. More specifically, the DNS query determines those SCSI targets that are visible to the host FCID.
  • the FCID of one or more virtual enclosure ports as potential target ports are then provided by the DNS server 1106 to the host 1102 at 1116.
  • the host Once the host has the FCIDs of those SCSI target ports available to it, it sends a fibre channel process login command to one or more of the available virtual enclosure SCSI target ports (identified by its FCID) at 1118. As described above, although the packet is addressed to an FCID assigned to a virtual enclosure port, a virtualization port that is bound to the virtual enclosure port or is trapping on behalf of the virtual enclosure port may actually handle these packets. This process login implements a mapping layer, which "maps" fibre channel to SCSI. Once completed, the SCSI virtual enclosure target port(s) 1108 send a fibre channel accept message at 1120. Since Fibre channel is mapped to SCSI, the host can send SCSI commands to the SCSI target virtual enclosure port(s).
  • the host 1102 can send SCSI commands to the virtual enclosure port(s) 1 108, it performs a SCSI process login at 1122 by sending a SCSI process login command to the SCSI target virtual enclosure port 1108.
  • the SCSI virtual enclosure port 1108 then sends a SCSI accept command 1124. Communication between the host and the SCSI target virtual enclosure port via the SCSI protocol is therefore established. For instance, the host may determine which LUNs are available to it, as well as read and write to those LUNs available to it.
  • the host 1102 Since the host 1102 knows which SCSI targets (and virtual enclosure ports) are available to it and can communicate with each of these ports, it can send a SCSI REPORT LUN command to the SCSI target virtual enclosure port(s) 1108 to determine those LUNs that are visible to the host FCID at 1126. For instance, these LUNs may simply be those VLUNs within the virtual enclosure, or may be a subset of those VLUNs. More specifically, this REPORT message may be sent to the FCIDs assigned to the virtual enclosure ports.
  • the receiving SCSI target virtual enclosure port 1108 determines which LUNs are visible to the host FCID at 1128 and sends a reply message indicating one or more available LUNs (e.g., PLUNs or VLUNs) at 1130.
  • the receiving virtual enclosure port may actually be a virtualization port that is either bound to or trapping on behalf of a virtual enclosure port.
  • these virtualization ports are also responsible for performing LUN mapping at 1128.
  • One method of LUN mapping will be described in further detail below with reference to FIG. 12.
  • the host 1102 may then send SCSI READ and WRITE commands to a particular VLUN at 1132 via the virtual enclosure port 1108.
  • Virtualization messages such as trap and bind messages may be implemented in any communication protocol.
  • the protocol that is implemented is preferably a reliable communication protocol, such as TCP sockets or Dynamic Instantiation Protocol (DIP), which runs on top of TCP/IP.
  • DIP Dynamic Instantiation Protocol
  • FIG. 12 is a diagram illustrating an exemplary LUN mapping table that may be used at step 1128 of FIG. 11 to perform LUN mapping.
  • a LUN mapping table 1200 may be used to identify for each initiator (e.g., host) 1202 those VLUNs 1204 of a virtual enclosure that are visible via one or more virtual enclosure ports 1206 (e.g., FCIDs) of the virtual enclosure.
  • a specified logical unit 1208 such as LUN0 may be used to represent a boot disk. Since it may be desirable to provide a different boot disk for each host, the specified logical unit 1208 is associated with one or more VLUNs 1204 which are made available to the host 1202 via the specified virtual enclosure ports 1206.
  • the virtual enclosure ports 1206 that are available to a particular host 1202 may be specified, or alternatively, a wildcard may be used to indicate that all virtual enclosure ports are available to the host.
  • a request for a specified logical unit 1208 is received from the host via one of the virtual enclosure ports, the VLUNs associated with the specified logical unit 1208 are employed to service the request.
  • FIG. 13 represents a virtual enclosure corresponding to the LUN mapping table of FIG. 12.
  • virtual enclosure 1302 represents four VLUNS, A 1304, B 1306, C 1308, and D 1310.
  • the virtual enclosure 1302 has two virtual enclosure ports, VEPl 1312 and VEP2 1314.
  • host HI sees LUN0 as VLUN A on any virtual enclosure port
  • host H2 sees LUN0 as VLUN B on any virtual enclosure port
  • host H3 sees LUN0 as VLUN C on any virtual enclosure port
  • host H4 sees LUN0 as VLUN D on virtual enclosure port 0 only
  • host H5 sees LUN0 as VLUN C on virtual enclosure port 1 only.
  • the present invention provides an additional level of storage virtualization.
  • the LUN mapping table is preferably maintained at a central location such as the virtual enclosure server. However, in accordance with one embodiment, the LUN mapping table is also be provided as well as periodically distributed to the appropriate network devices (e.g., virtualization ports) within the storage area network. For instance, the LUN mapping table may be distributed at host login to a virtual enclosure port. In other words, the virtualization port is provided a LUN map corresponding to the host that has logged in via the virtual enclosure server.
  • the present invention is described as being applied to frames. However, it should be understood that the invention is not limited to such implementations, but instead would equally apply to packets as well. Moreover, the present invention would apply regardless of the context and system in which it is implemented. Thus, broadly speaking, the coordination of binding and trapping by multiple ports need not be performed using a virtual enclosure server as described above, but may be performed in an alternate manner.
  • the above- described embodiments may be implemented in a variety of network devices (e.g., servers) as well as in a variety of mediums.
  • instructions and data for implementing the above-described invention may be stored on a disk drive, a hard drive, a floppy disk, a server computer, or a remotely networked computer. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Memory System Of A Hierarchy Structure (AREA)

Abstract

Methods and apparatus for implementing storage virtualization on a network device of a storage area network are disclosed. A virtual enclosure is created that has one or more virtual enclosure ports and is adapted for representing one or more virtual storage units. Each of the virtual storage units represents one or more physical storage locations on one or more physical storage units of the storage area network. Each of the virtual enclosure ports of the virtual enclosure is associated with a port of a network device within the storage area network. An address or identifier is then assigned to each of the virtual enclosure ports.

Description

METHODS AND APPARATUS FOR IMPLEMENTING VIRTUALIZATION OF STORAGE WITHIN A STORAGE AREA NETWORK THROUGH A VIRTUAL ENCLOSURE
BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to network technology. More particularly, the present invention relates to methods and apparatus for supporting virtualization of storage within a storage area network.
2. Description of the Related Art
In recent years, the capacity of storage devices has not increased as fast as the demand for storage. Therefore a given server or other host must access multiple, physically distinct storage nodes (typically disks). In order to solve these storage limitations, the storage area network (SAN) was developed. Generally, a storage area network is a high-speed special-purpose network that interconnects different data storage devices and associated data hosts on behalf of a larger network of users. However, although a SAN enables a storage device to be configured for use by various network devices and/or entities within a network, data storage needs are often dynamic rather than static.
FIG. 1 A illustrates an exemplary conventional storage area network. More specifically, within a storage area network 102, it is possible to couple a set of hosts (e.g., servers or workstations) 104, 106, 108 to a pool of storage devices (e.g., disks). In SCSI parlance, the hosts may be viewed as "initiators" and the storage devices may be viewed as "targets." A storage pool may be implemented, for example, through a set of storage arrays or disk arrays 110, 112, 114. Each disk array 110, 112, 114 further corresponds to a set of disks. In this example, first disk array 110 corresponds to disks 116, 118, second disk array 112 corresponds to disk 120, and third disk array 114 corresponds to disks 122, 124. Rather than enabling all hosts 104-108 to access all disks 116-124, it is desirable to enable the dynamic and invisible allocation of storage (e.g., disks) to each of the hosts 104-108 via the disk arrays 110, 112, 114. In other words, physical memory (e.g., physical disks) may be allocated through the concept of virtual memory (e.g., virtual disks). This allows one to connect heterogeneous initiators to a distributed, heterogeneous set of targets (storage pool) in a manner enabling the dynamic and transparent allocation of storage.
The concept of virtual memory has traditionally been used to enable physical memory to be virtualized through the translation between physical addresses in physical memory and virtual addresses in virtual memory. Recently, the concept of "virtualization" has been implemented in storage area networks through various mechanisms. Virtualization interconverts physical storage and virtual storage on a storage network. The hosts (initiators) see virtual disks as targets. The virtual disks represent available physical storage in a defined but somewhat flexible manner. Virtualization provides hosts with a representation of available physical storage that is not constrained by certain physical arrangements/allocation of the storage.
One early technique, Redundant Array of Independent Disks (RAID), provides some limited features of virtualization. Various RAID subtypes have been implemented. In RAEDl, a virtual disk may correspond to two physical disks 116, 118 which both store the same data (or otherwise support recovery of the same data), thereby enabling redundancy to be supported within a storage area network. In RAIDO, a single virtual disk is striped across multiple physical disks. Some other types of virtualization include concatenation, sparing, etc. Some aspects of virtualization have recently been achieved through implementing the virtualization function in various locations within the storage area network. Three such locations have gained some level of acceptance: virtualization in the hosts (e.g., 104-108), virtualization in the disk arrays or storage arrays (e.g., 110-114), and virtualization in a storage appliance 126 separate from the hosts and storage pool. Unfortunately, each of these implementation schemes has undesirable performance limitations.
Virtualization in the storage array is one of the most common storage virtualization solutions in use today. Through this approach, virtual volumes are created over the storage space of a specific storage subsystem (e.g., disk array). Creating virtual volumes at the storage subsystem level provides host independence, since virtualization of the storage pool is invisible to the hosts. In addition, virtualization at the storage system level enables optimization of memory access and therefore high performance. However, such a virtualization scheme typically will allow a uniform management structure only for a homogenous storage environment and even then only with limited flexibility. Further, since virtualization is performed at the storage subsystem level, the physical-virtual limitations set at the storage subsystem level are imposed on all hosts in the storage area network. Moreover, each storage subsystem (or disk array) is managed independently. Virtualization at the storage level therefore rarely allows a virtual volume to span over multiple storage subsystems (e.g., disk arrays), thus limiting the scalability of the storage-based approach.
When virtualization is implemented on each host, it is possible to span multiple storage subsystems (e.g., disk arrays). A host-based approach has an additional advantage, in that a limitation on one host does not impact the operation of other hosts in a storage area network. However, virtualization at the host-level requires the existence of a software layer running on each host (e.g., server) that implements the virtualization function. Running this software therefore impacts the performance of the hosts running this software. Another key difficulty with this method is that it assumes a prior partitioning of the available storage to the various hosts. Since such partitioning is supported at the host-level and the virtualization function of each host is performed independently of the other hosts in the storage area network, it is difficult to coordinate storage access across the hosts. The host-based approach therefore fails to provide an adequate level of security. Due to this security limitation, it is difficult to implement a variety of redundancy schemes such as RAID which require the "locking" of memory during read and write operations. In addition, when mirroring is performed, the host must replicate the data multiple times, increasing its input-output and CPU load, and increasing the traffic over the SAN. Virtualization in a storage area network appliance placed between the hosts and the storage solves some of the difficulties of the host-based and storage-based approaches. The storage appliance globally manages the mapping and allocation of physical storage to virtual volumes. Typically, the storage appliance manages a central table that provides the current mapping of physical to virtual. Thus, the storage appliance-based approach enables the virtual volumes to be implemented independently from both the hosts and the storage subsystems on the storage area network, thereby providing a higher level of security. Moreover, this approach supports virtualization across multiple storage subsystems. The key drawback of many implementations of this architecture is that every input/output (I/O) of every host must be sent through the storage area network appliance, causing significant performance degradation and a storage area network bottleneck. This is particularly disadvantageous in systems supporting a redundancy scheme such as RAID, since data must be mirrored across multiple disks. In another storage appliance-based approach, the appliance makes sure that all hosts receive the current version of the table. Thus, in order to enable the hosts to receive the table from the appliance, a software shim from the appliance to the hosts is required, adding to the complexity of the system. Moreover, since the software layer is implemented on the host, many of the disadvantages of the host-based approach are also present. In view of the above, it would be desirable if various storage devices or portions thereof could be logically and dynamically assigned to various devices and/or entities within a network. Moreover, it would be beneficial if such a mechanism could be implemented to support the virtualization of storage within a SAN without the disadvantages of traditional virtualization approaches.
SUMMARY OF THE INVENTION
Methods and apparatus for implementing virtualization of storage in a storage area network are disclosed. This is accomplished through the use of one or more network devices capable of being placed in a data path between the hosts and the storage devices. As a result, neither the storage devices nor the hosts require additional software or hardware to support storage virtualization. Thus, the present invention is superior to the host based approach, which requires that each host be burdened by additional software to implement virtualization functionality. Moreover, the present invention enables multiple network devices to simultaneously manage the virtualization of various storage devices. Importantly, switch-based virtualization may be implemented on a per port basis. Any number of ports on a switch can manage virtualization of its own traffic. This allows a network's virtualization capacity to scale with the number of ports. Since there are large numbers of ports in any network system, there will nearly always be sufficient bandwidth for virtualization. Accordingly, virtualization of storage may be achieved without many of the drawbacks present in conventional virtualization schemes.
In accordance with one aspect of the invention, a virtual enclosure is created that has one or more virtual enclosure ports and is adapted for representing one or more virtual storage units. In other words, the virtual enclosure serves to "enclose" selected virtual storage units, which may be accessed via the virtual enclosure ports. Each of the virtual storage units represents one or more physical storage locations on one or more physical storage units of the storage area network. In addition, each of the virtual enclosure ports of the virtual enclosure is associated with a port of a network device within the storage area network. An address or identifier is then assigned to each of the virtual enclosure ports. For instance, the address or identifier may be a Fibre Channel identifier (FCID). Thus, a message (e.g., packet or frame) directed to a virtual enclosure port (or its assigned address/identifier) may be handled by the port associated with the virtual enclosure port.
In accordance with various embodiments of the invention, a virtual enclosure is implemented within a Fibre channel network. Thus, a Node World Wide Name (NWWN) is associated with the virtual enclosure. In addition, a Port World Wide Name (PWWN) is associated with each virtual enclosure port.
In accordance with another aspect of the invention, a port of a network device within the storage area network is instructed to handle messages on behalf of a virtual enclosure port. This may be accomplished in two ways. First, the port may be instructed to "bind" itself to the virtual enclosure port. In other words, the port acts as the virtual enclosure port, and all messages directed to the virtual enclosure port and received by the port are handled by that port. Second, the port may be instructed to serve as a "trapping port." More particularly, in addition to the port that is bound to the virtual enclosure port, one or more additional ports may also handle messages they receive that are directed to the virtual enclosure port. A trapping port is preferably a port that is directly connected to a host, and therefore can track those requests received by it as well as the responses associated with those requests. Binding and trapping among multiple ports on behalf of a single virtual enclosure port is preferably coordinated at a central location such as a virtual enclosure server.
Various network devices may be configured or adapted for performing the disclosed virtualization processes. These network devices include, but are not limited to, servers (e.g., hosts), routers, and switches. Moreover, the functionality for the above-mentioned virtualization processes may be implemented in software as well as hardware.
Yet another aspect of the invention pertains to computer program products including machine-readable media on which are provided program instructions for implementing the methods and techniques described above, in whole or in part. Any of the methods of this invention may be represented, in whole or in part, as program instructions that can be provided on such machine-readable media. In addition, the invention pertains to various combinations and arrangements of data generated and/or used as described herein. For example, packets and frames having the format described herein and provided on appropriate media are part of this invention.
These and other features of the present invention will be described in more detail below in the detailed description of the invention and in conjunction with the following figures. BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 A is a block diagram illustrating an exemplary conventional storage area network capable of implementing various embodiments of prior art virtualization functions.
FIG. IB is a block diagram illustrating an exemplary storage area network in which various embodiments of the invention may be implemented.
FIG. 2 is a block diagram illustrating a virtualization model that may be implemented in accordance with various embodiments of the invention.
FIG. 3A is a block diagram illustrating an exemplary virtualization switch in which various embodiments of the present invention may be implemented.
FIG. 3B is a block diagram illustrating an exemplary standard switch in which various embodiments of the present invention may be implemented.
FIG. 4 is a transaction flow diagram illustrating a conventional method of implementing a node world wide name (NWWN) and port world wide name (PWWN) for each SCSI target port.
FIG. 5 is a block diagram illustrating an exemplary virtual enclosure in accordance with one embodiment of the invention.
FIG. 6 is a diagram illustrating an exemplary system in which a virtual enclosure is implemented through the binding of the virtual enclosure ports to various virtualization ports in accordance with various embodiments of the invention.
FIG. 7 is a process flow diagram illustrating a method of creating a virtual enclosure in accordance with various embodiments of the invention. FIG. 8 is a diagram illustrating a conventional fibre channel identifier (FCID) that may be associated with a virtual enclosure port in accordance with various embodiments of the invention.
FIG. 9 is a transaction flow diagram illustrating one method of coordinating virtual enclosure binding and trapping functionality of virtualization ports in accordance with various embodiments of the invention.
FIG. 10 is a diagram illustrating an exemplary table that may be maintained by a virtualization port indicating FCIDs to be handled by the virtualization port in accordance with various embodiments of the invention.
FIG. 11 is a transaction flow diagram illustrating one method of establishing communication between a host and one or more virtualization ports (e.g., virtual enclosure ports, trapping ports) such that the host can access one or more LUNs in accordance with various embodiments of the invention.
FIG. 12 is a diagram illustrating an exemplary LUN mapping table that may be used at step 1128 of FIG. 11 to perform LUN mapping.
FIG. 13 is a diagram representing a virtual enclosure corresponding to the LUN mapping table of FIG. 12.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be obvious, however, to one skilled in the art, that the present invention may be practiced without some or all of these specific details. In other instances, well known process steps have not been described in detail in order not to unnecessarily obscure the present invention.
In accordance with various embodiments of the present invention, virtualization of storage within a storage area network may be implemented { through the creation of a virtual enclosure having one or more virtual enclosure ports. The virtual enclosure is implemented, in part, by one or more network devices, which will be referred to herein as virtualization switches. More specifically, a virtualization switch, or more specifically, a virtualization port within the virtualization switch, may handle messages such as packets or frames on behalf of one of the virtual enclosure ports. Thus, embodiments of the invention may be applied to a packet or frame directed to a virtual enclosure port, as will be described in further detail below. For convenience, the subsequent discussion will describe embodiments of the invention with respect to frames. Switches act on frames and use information about SANs to make switching decisions.
Note that the frames being received and transmitted by a virtualization switch possess the frame format specified for a standard protocol such as Ethernet or fibre channel. Hence, software and hardware conventionally used to generate such frames may be employed with this invention. Additional hardware and/or software is employed to modify and/or generate frames compatible with the standard protocol in accordance with this invention. Those of skill in the art will understand how to develop the necessary hardware and software to allow virtualization as described below.
Obviously, the appropriate network devices should be configured with the appropriate software and/or hardware for performing virtualization functionality. Of course, all network devices within the storage area network need not be configured with the virtualization functionality. Rather, selected switches and or ports may be configured with or adapted for virtualization functionality. Similarly, in various embodiments, such virtualization functionality may be enabled or disabled through the selection of various modes. Moreover, it may be desirable to configure selected ports of network devices as virtualization-capable ports capable of performing virtualization, either continuously, or only when in a virtualization enabled state. The standard protocol employed in the storage area network (i.e., the protocol used to frame the data) will typically, although not necessarily, be synonymous with the "type of traffic" carried by the network. As explained below, the type of traffic is defined in some encapsulation formats. Examples of the type of traffic are typically layer 2 or corresponding layer formats such as Ethernet, Fibre channel, and InfiniBand.
As described above, a storage area network (SAN) is a high-speed special- purpose network that interconnects different data storage devices with associated network hosts (e.g., data servers or end user machines) on behalf of a larger network of users. A SAN is defined by the physical configuration of the system. In other words, those devices in a SAN must be physically interconnected.
Within a storage area network 131 such as that illustrated in FIG. IB, various storage devices 132, 134, 136, 138, 140, and 142 may be implemented, which may be homogeneous (e.g., identical device types, sizes, or configurations) as well as heterogeneous (e.g., different device types, sizes or configurations). Data may be read from, as well as written to, various portions of the storage devices 132-142 in response to commands sent by hosts 144 and 146. Communication among the storage devices and hosts is accomplished by coupling the storage devices and hosts together via one or more switches, routers, or other network nodes configured to perform a switching function. In this example, switches 148, 150, and 152 communicate with one another via interswitch links 154 and 156.
As indicated above, this invention pertains to "virtualization" in storage networks. Unlike prior methods, virtualization in this invention is implemented through the creation and implementation of a virtual enclosure. This is accomplished, in part, through the use of switches or other "interior" network nodes of a storage area network to implement the virtual enclosure. Further, the virtualization of this invention typically is implemented on a per port basis. In other words, a multi-port virtualization switch will have virtualization separately implemented on one or more of its ports. Individual ports have dedicated logic for handing the virtualization functions for packets or frames handled by the individual ports. This allows virtualization processing to scale with the number of ports, and provides far greater bandwidth for virtualization than can be provided with host based or storage based virtualization schemes. In such prior art approaches the number of connections between hosts and the network fabric or between storage nodes and the network fabric are limited - at least in comparison to the number of ports in the network fabric.
In a specific and preferred embodiment of the invention, the virtualization logic is separately implemented at individual ports of a given switch - rather than having centralized processing for all ports of a switch. This allows the virtualization processing capacity to be closely matched with the exact needs of the switch (and the virtual enclosure) on a per port basis. If a central processor is employed for the entire switch (serving numerous ports), the processor must be designed/selected to handle maximum traffic at all ports. For many applications, this represents extremely high processing requirements and a very large/expensive processor. If the central processor is too small, the switch will at times be unable to keep up with the switching/virtualization demands of the network.
Virtualization may take many forms. In general, it may be defined as logic or procedures that inter-relate physical storage and virtual storage on a storage network. Hosts see a representation of available physical storage that is not constrained by the physical arrangements or allocations inherent in that storage. One example of a physical constraint that is transcended by virtualization includes the size and location of constituent physical storage blocks. For example, logical units as defined by the Small Computer System Interface (SCSI) standards come in precise physical sizes (e.g., 36GB and 72GB). Virtualization can represent storage in virtual logical units that are smaller or larger than the defined size of a physical logical unit. Further, virtualization can present a virtual logical unit comprised of regions from two or more different physical logical units, sometimes provided on devices from different vendors. Preferably, the virtualization operations are transparent to at least some network entities (e.g., hosts). In some general ways, virtualization on a storage area network is similar to virtual memory on a typical computer system. Virtualization on a network, however, brings far greater complexity and far greater flexibility. The complexity arises directly from the fact that there are a number of separately interconnected network nodes. Virtualization must span these nodes. The nodes include hosts, storage subsystems, and switches (or comparable network traffic control devices such as routers). Often the hosts and/or storage subsystems are heterogeneous, being provided by different vendors. The vendors may employ distinctly different protocols (standard protocols or proprietary protocols). Thus, in many cases, virtualization provides the ability to connect heterogeneous initiators (e.g., hosts or servers) to a distributed, heterogeneous set of targets (storage subsystems), enabling the dynamic and transparent allocation of storage.
Examples of network specific virtualization operations include the following: RAID 0 through RAID 5, concatenation of memory from two or more distinct logical units of physical memory, sparing (auto-replacement of failed physical media), remote mirroring of physical memory, logging information (e.g., errors and/or statistics), load balancing among multiple physical memory systems, striping (e.g., RAID 0), security measures such as access control algorithms for accessing physical memory, resizing of virtual memory blocks, Logical Unit (LUN) mapping to allow arbitrary LUNs to serve as boot devices, backup of physical memory (point in time copying), and the like. These are merely examples of virtualization functions. This invention is not limited to this full set or any particular subset thereof.
In some of the discussion herein, the functions of virtualization switches of this invention are described in terms of the SCSI protocol. This is because many storage area networks in commerce run a SCSI protocol to access storage sites. Frequently, the storage area network employs fibre channel (FC-PH (ANSI X3.230- 1994, Fibre channel - Physical and Signaling Interface) as a lower level protocol and runs IP and SCSI on top of fibre channel. Note that the invention is not limited to any of these protocols. For example, fibre channel may be replaced with Ethernet, Infiniband, and the like. Further the higher level protocols need not include SCSI. For example, this may include SCSI over FC, iSCSI (SCSI over IP), parallel SCSI (SCSI over a parallel cable), serial SCSI (SCSI over serial cable, and all the other incarnations of SCSI.
Because SCSI is so widely used in storage area networks, much of the terminology used herein will be SCSI terminology. The use of SCSI terminology (e.g., "initiator" and "target") does not imply that the describe procedure or apparatus must employ SCSI. Before going further, it is worth explaining a few of the SCSI terms that will be used in this discussion. First an "initiator" is a device (usually a host system) that requests an operation to be performed by another device. Typically, in the context of this document, a host initiator will request a read or write operation be performed on a region of virtual or physical memory. Next, a "target" is a device that performs an operation requested by an initiator. For example, a target physical memory disk will obtain or write data as initially requested by a host initiator. Note that while the host initiator may provide instructions to read from or write to a "virtual" target having a virtual address, a virtualization switch of this invention must first convert those instructions to a physical target address before instructing the target.
Targets may be divided into physical or virtual "logical units." These are specific devices addressable through the target. For example, a physical storage subsystem may be organized in a number of distinct logical units. In this document, hosts view virtual memory as distinct virtual logical units. Sometimes herein, logical units will be referred to as "LUNs." In the SCSI standard, LUN refers to a logical unit number. But in common parlance, LUN also refers to the logical unit itself.
Central to virtualization is the concept of a "virtualization model." This is the way in which physical storage provided on storage subsystems (such as disk arrays) is related to a virtual storage seen by hosts or other initiators on a network. While the relationship may take many forms and be characterized by various terms, a SCSI-based terminology will be used, as indicated above. Thus, the physical side of the storage area network will be described as a physical LUN. The host side, in turn, sees one or more virtual LUNs, which are virtual representations of the physical LUNs. The mapping of physical LUNs to virtual LUNs may logically take place over one, two, or more levels. In the end, there is a mapping function that can be used by switches of this invention to interconvert between physical LUN addresses and virtual LUN addresses.
FIG. 2 is a block diagram illustrating an example of a virtualization model that may be implemented within a storage area network in accordance with various embodiments of the invention. As shown, the physical storage of the storage area network is made up of one or more physical LUNs, shown here as physical disks 202. Each physical LUN is a device that is capable of containing data stored in one or more contiguous blocks which are individually and directly accessible. For instance, each block of memory within a physical LUN may be represented as a block 204, which may be referred to as a disk unit (DUnit).
Through a mapping function 206, it is possible to convert physical LUN addresses associated with physical LUNs 202 to virtual LUN addresses, and vice versa. More specifically, as described above, the virtualization and therefore the mapping function may take place over one or more levels. For instance, as shown, at a first virtualization level, one or more virtual LUNs 208 each represents one or more physical LUNs 202, or portions thereof. The physical LUNs 202 that together make up a single virtual LUN 208 need not be contiguous. Similarly, the physical LUNs 202 that are mapped to a virtual LUN 208 need not be located within a single target. Thus, through virtualization, virtual LUNs 208 may be created that represent physical memory located in physically distinct targets, which may be from different vendors, and therefore may support different protocols and types of traffic.
Although the virtualization model may be implemented with a single level, a hierarchical arrangement of any number of levels may be supported by various embodiments of the present invention. For instance, as shown, a second virtualization level within the virtualization model of FIG. 2 is referred to as a high-level VLUN or volume 210. Typically, the initiator device "sees" only
VLUN 210 when accessing data. In accordance with various embodiments of the invention, multiple VLUNs are "enclosed" within a virtual enclosure such that only the virtual enclosure may be "seen" by the initiator. In other words, the VLUNs enclosed by the virtual enclosure are not visible to the initiator.
In this example, VLUN 210 is implemented as a "logical" RAID array of virtual LUNs 208. Moreover, such a virtualization level may be further implemented, such as through the use of striping and/or mirroring. In addition, it is important to note that it is unnecessary to specify the number of virtualization levels to support the mapping function 206. Rather, an arbitrary number of levels of virtualization may be supported, for example, through a recursive mapping function. For instance, various levels of nodes may be built and maintained in a tree data structure, linked list, or other suitable data structure that can be traversed.
Each initiator may therefore access physical LUNs via nodes located at any of the levels of the hierarchical virtualization model. Nodes within a given virtualization level of the hierarchical model implemented within a given storage area network may be both visible to and accessible to an allowed set of initiators (not shown). However, in accordance with various embodiments of the invention, these nodes are enclosed in a virtual enclosure, and are therefore no longer visible to the allowed set of initiators. Nodes within a particular virtualization level (e.g., VLUNs) need to be created before functions (e.g., read, write) may be operated upon them. This may be accomplished, for example, through a master boot record of a particular initiator. In addition, various initiators may be assigned read and/or write privileges with respect to particular nodes (e.g., VLUNs) within a particular virtualization level. In this manner, a node within a particular virtualization level may be accessible by selected initiators.
As described above, various switches within a storage area network may be virtualization switches supporting virtualization functionality. FIG. 3A is a block diagram illustrating an exemplary virtualization switch in which various embodiments of the present invention may be implemented. As shown, data or messages are received by an intelligent, virtualization port via a bi-directional connector 302. In addition, the virtualization port is adapted for handling messages on behalf of a virtual enclosure port, as will be described in further detail below. In association with the incoming port, Media Access Control (MAC) block 304 is provided, which enables frames of various protocols such as Ethernet or fibre channel to be received. In addition, a virtualization intercept switch 306 determines whether an address specified in an incoming frame pertains to access of a virtual storage location of a virtual storage unit representing one or more physical storage locations on one or more physical storage units of the storage area network. For instance, the virtual storage unit may be a virtual storage unit (e.g., VLUN) that is enclosed within a virtual enclosure.
When the virtualization intercept switch 306 determines that the address specified in an incoming frame pertains to access of a virtual storage location rather than a physical storage location, the frame is processed by a virtualization processor 308 capable of performing a mapping function such as that described above. More particularly, the virtualization processor 308 obtains a virtual- physical mapping between the one or more physical storage locations and the virtual storage location. In this manner, the virtualization processor 308 may look up either a physical or virtual address, as appropriate. For instance, it may be necessary to perform a mapping from a physical address to a virtual address or, alternatively, from a virtual address to one or more physical addresses.
Once the virtual-physical mapping is obtained, the virtualization processor 308 may then employ the obtained mapping to either generate a new frame or modify the existing frame, thereby enabling the frame to be sent to an initiator or a target specified by the virtual-physical mapping. The mapping function may also specify that the frame needs to be replicated multiple times, such as in the case of a mirrored write. More particularly, the source address and/or destination addresses are modified as appropriate. For instance, for data from the target, the virtualization processor replaces the source address, which was originally the physical LUN address with the corresponding virtual LUN and address. In the destination address, the port replaces its own address with that of the initiator.
For data from the initiator, the port changes the source address from the initiator's address to the port's own address. It also changes the destination address from the virtual LUN/address to the corresponding physical LUN/address. The new or modified frame may then be provided to the virtualization intercept switch 306 to enable the frame to be sent to its intended destination.
While the virtualization processor 308 obtains and applies the virtual- physical mapping, the frame or associated data may be stored in a temporary memory location (e.g., buffer) 310. In addition, it may be necessary or desirable to store data that is being transmitted or received until it has been confirmed that the desired read or write operation has been successfully completed. As one example, it may be desirable to write a large amount of data to a virtual LUN, which must be transmitted separately in multiple frames. It may therefore be desirable to temporarily buffer the data until confirmation of receipt of the data is received. As another example, it may be desirable to read a large amount of data from a virtual LUN, which may be received separately in multiple frames. Furthermore, this data may be received in an order that is inconsistent with the order in which the data should be transmitted to the initiator of the read command. In this instance, it may be beneficial to buffer the data prior to transmitting the data to the initiator to enable the data to be re-ordered prior to transmission. Similarly, it may be desirable to buffer the data in the event that it is becomes necessary to verify the integrity of the data that has been sent to an initiator (or target).
The new or modified frame is then received by a forwarding engine 312, which obtains information from various fields of the frame, such as source address and destination address. The forwarding engine 312 then accesses a forwarding table 314 to determine whether the source address has access to the specified destination address. More specifically, the forwarding table 314 may include physical LUN addresses as well as virtual LUN addresses. The forwarding engine 312 also determines the appropriate port of the switch via which to send the frame, and generates an appropriate routing tag for the frame. Once the frame is appropriately formatted for transmission, the frame will be received by a buffer queuing block 316 prior to transmission. Rather than transmitting frames as they are received, it may be desirable to temporarily store the frame in a buffer or queue 318. For instance, it may be desirable to temporarily store a packet based upon Quality of Service in one of a set of queues that each correspond to different priority levels. The frame is then transmitted via switch fabric 320 to the appropriate port. As shown, the outgoing port has its own MAC block 322 and bi-directional connector 324 via which the frame may be transmitted.
As described above, all switches in a storage area network need not be virtualization switches. In other words, a switch may be a standard switch in which none of the ports implement "intelligent," virtualization functionality. FIG. 3B is a block diagram illustrating an exemplary standard switch in which various embodiments of the present invention may be implemented. As shown, a standard port 326 has a MAC block 304. However, a virtualization intercept switch and virtualization processor such as those illustrated in FIG. 3A are not implemented. A frame that is received at the incoming port is merely processed by the forwarding engine 312 and its associated forwarding table 314. Prior to transmission, a frame may be queued 316 in a buffer or queue 318. Frames are then forwarded via switch fabric 320 to an outgoing port. As shown, the outgoing port also has an associated MAC block 322 and bi-directional connector 324. Of course, each port may support a variey of protocols. For instance, the outgoing port may be an iSCSI port (i.e. a port that supports SCSI over IP over Ethernet), which also supports virtualization, as well as parallel SCSI and serial SCSI.
Although the network devices described above with reference to FIG. 3 A and 3B are described as switches, these network devices are merely illustrative. Thus, other network devices such as routers may be implemented to receive, process, modify and/or generate packets or frames with functionality such as that described above for transmission in a storage area network. Moreover, the above- described network devices are merely illustrative, and therefore other types of network devices may be implemented to perform the disclosed virtualization functionality.
Typically, SCSI targets are directly accessible by SCSI initiators (e.g., hosts). In other words, SCSI targets such as PLUNs are visible to the hosts that are accessing those SCSI targets. Similarly, even when VLUNs are implemented, the VLUNs are visible and accessible to the SCSI initiators. Thus, each host must typically identify those VLUNs that are available to it. More specifically, the host typically determines which SCSI target ports are available to it. The host may then ask each of those SCSI target ports which VLUNs are available via those SCSI target ports.
Within a Fibre channel network, all Fibre channel devices have a World Wide Name (WWN). More specifically, a Node WWN (NWWN) is the WWN of the node that is connected to a particular port. In other words, the NWWN is the WWN of the system, storage device, or subsystem that is connected to the switch port. In addition to a Node WWN, a Port WWN (PWWN) serves as a given name for a particular port. A Fibre channel network ID (FCID) for the particular switch port is used to identify the physical location of a port. Each Fibre channel device may have multiple ports, each of which is uniquely identified by a NWWN and a PWWN.
Unfortunately, there are several disadvantages associated with conventional applications of Fibre channel WWN nomenclature. For instance, a NWWN and a PWWN must be allocated for each port of a Fibre channel device. However, it may be undesirable to allow the NWWN and PWWN of each port to be visible to an initiator such as a host. Moreover, the number of available ports is limited to the number of ports in a particular storage device, subsystem or switch. Similarly, the storage device, subsystem or switch may have a greater number of ports than are needed in a particular storage virtualization scheme. However, virtualization ports are expensive to implement, and it is therefore undesirable to "waste" these intelligent ports. Thus, in accordance with the present invention, virtualization ports may be used on an as-needed basis. In other words, ports may be selected for implementing one or more virtual enclosures, either by binding those ports to virtual enclosure ports or by implementing these ports as trapping ports. As a result, those ports that are most expensive to produce may be used to their maximum capacity.
In one embodiment of the invention, virtualization in a network such as a fibre channel network is implemented without limiting the number of accessible ports in a storage subsystem. Moreover, ports within a SAN may selected and allocated on an as-needed basis for a particular storage virtualization scheme. Various embodiments of the invention are described in further detail below.
In order to understand how the present invention may be applied in a fibre channel network, it is helpful to illustrate a conventional application of WWN nomenclature in a Fibre channel network. FIG. 4 is a transaction flow diagram illustrating a conventional method of implementing a node world wide name (NWWN) and port world wide name (PWWN) for each SCSI target port. As shown, interactions between a host 402, switch 404, Domain Name System (DNS) server 406 and SCSI target port 408 are represented by corresponding labeled vertical lines. As shown, in order for a fibre channel node such as a host 402 to establish a logical connection to a fabric switch 404, it performs a fabric login (FLOGI) 410. Unlike many LAN technologies that use a fixed Media Access Control (MAC) address, Fibre channel uses an address identifier, referred to as an FCID, which is dynamically assigned during login. Thus, the switch 404 provides an FCID to the host 402 at 412. Once the host been assigned a host FCID, the host may perform a DNS query of a DNS server 406 via a SCSI REPORT command, which requests the SCSI targets that are available, and therefore visible, to the host at block 414. More specifically, the DNS query determines those SCSI targets that are visible to the host FCID. The FCID of a SCSI target port is then provided by the DNS server 406 to the host 402 at 416.
Once the host has the FCIDs of those SCSI target ports available to it, it sends a Fibre channel process login command to one of the available SCSI target ports (identified by its FCID) at 418. This process login implements a mapping layer, which "maps" Fibre channel to SCSI. Once completed, the SCSI target port 408 sends a Fibre channel accept message at 420. Since Fibre channel is mapped to SCSI, the host can send SCSI commands to the SCSI target.
Now that the host 402 can send SCSI commands to the SCSI target port 408, it performs a SCSI process login at 422 by sending a SCSI process login command to the SCSI target port 408. The SCSI target port 408 then sends a SCSI accept command 424. Communication between the host and the SCSI target via the SCSI protocol is therefore established. For instance, the host may determine which LUNs are available to it, as well as read and write to those LUNs available to it.
Since the host 402 knows which SCSI target ports are available to it and can communicate with each of these ports, it can send a SCSI REPORT LUN command to the SCSI target port 408 to determine those LUNS that are visible to the host FCID at 426. The SCSI target port 408 determines which LUNs are visible to the host FCID and sends a list of LUNs (e.g., PLUNs or VLUNs) at 428. The host 402 may then send SCSI READ and WRITE commands to a particular PLUN or VLUN at 430 via the SCSI target port 408.
In accordance with various embodiments of the invention, storage virtualization is implemented in a storage area network through the creation of a "virtual enclosure." FIG. 5 is a block diagram of an exemplary virtual enclosure in accordance with one embodiment of the invention. A virtual enclosure 502 is a virtual entity that is that is adapted for "enclosing" or representing one or more virtual storage units. For instance, one or more of the virtual storage units may each comprise a VLUN or other virtual representation of storage on a SAN. As described above, each virtual storage unit represents one or more physical storage locations on one or more physical storage units 504 of the storage area network. Through the creation of a virtual enclosure, various virtualization ports such as those described above may be implemented on an as-needed basis to support virtualization within a particular storage area network. More specifically, the virtual enclosure has one or more "virtual enclosure ports," each of which is associated with a port (e.g., virtualization port) of a network device within the storage area network. For instance, the port of the network device may be instructed, either statically or dynamically, to handle messages addressed to the virtual enclosure port. An address or identifier is also assigned to each of the virtual enclosure ports. Thus, a message (e.g., packet or frame) addressed to a virtual enclosure port may be identified by this address or identifier, thereby enabling a port to intercept and handle messages on behalf of a particular virtual enclosure port. For instance, a FCID may be assigned to each of the virtual enclosure ports for use in a fibre channel network, or with various fibre channel devices within a network such as a SAN. In this manner, a virtual enclosure is created for representing one or more virtual storage units.
Within a SAN, it is possible to create different virtual SANs (VSANs). One method of implementing virtual storage area networks (VSANs) within a single storage area network is described in further detail with reference to U.S. Patent Application No. , Attorney Docket No. ANDIP001 , entitled
"Methods and Apparatus for Encapsulating a Frame for Transmission in a Storage Area Network," Edsall, et al., filed on December 26, 2001, which is incorporated herein by reference for all purposes. In other words, it may be desirable or necessary to distribute physical storage units among different VSANs. Accordingly, when the virtual enclosure ports are bound to virtualization ports, these selected virtualization ports may be distributed among multiple VSANs.
Each virtual enclosure may have any number of virtual enclosure ports. In other words, it may be desirable to enable the number of virtual enclosure ports to be selectable. For instance, a system administrator responsible for creating and maintaining the virtual enclosure definitions may select the appropriate number of virtual enclosure ports. It is important to note that the number of virtual enclosure ports within a virtual enclosure is not limited by the number of ports within a particular network device within the storage area network. More specifically, each virtual enclosure port is associated with a port of a network device within the storage area network. Li other words, the virtual enclosure ports of a single virtual enclosure may be simultaneously associated with ports (e.g., virtualization ports) of multiple, different network devices within the storage area network rather than a single network device. In addition, a single port of a network device such as a virtualization port may be simultaneously associated with or bound to multiple virtual enclosure ports. Accordingly, the number of ports that are available for virtualization of virtual storage units is virtually unlimited.
As described above, an address or identifier is assigned to each of the virtual enclosure ports. Theoretically, this address or identifier may be an address or identifier that has been previously assigned to the port of the storage area network that is later associated with the virtual enclosure port. However, since a single virtualization port may be bound to multiple virtual enclosure ports, an address or identifier previously assigned to the port (e.g., virtualization port) will not uniquely identify a virtual enclosure port. Thus, the address or identifier is assigned to the virtual enclosure port to uniquely identify the virtual enclosure port. At that time, the address or identifier is provided to the virtualization port that is associated with the virtual enclosure port. In accordance with one embodiment, the virtual enclosure is used in a fibre channel network, and therefore each fibre channel device has a NWWN. This nomenclature is leveraged in accordance with various embodiments of the invention to enable a virtual enclosure to be implemented. Thus, both a NWWN and a PWWN together may be used as an address or identifier to uniquely identify a virtual enclosure port. More particularly, a NWWN is associated with the virtual enclosure. In addition, a PWWN is assigned to each virtual enclosure port. The NWWN and PWWN together identify the virtual enclosure port until it is assigned an address or identifier such as an FCID.
Once the virtual enclosure is created, one or more virtual storage units may be assigned to the virtual enclosure. As described above, each of the virtual storage units may be a VLUN or other virtual representation of storage on the storage area network. In this manner, the virtual enclosure "encloses" those virtual storage units, thereby requiring a host to access those virtual storage units via the virtual enclosure. In this manner, the VLUNS may be "hidden" from the host. In other words, by accessing a PWWN, the host merely has access to the virtual enclosure rather than a specific device or VLUN. More specifically, the VLUNs appear as logical units behind a set of virtual enclosure ports. Thus, the PLUNs that are used to create the VLUNs are hidden from the host.
As described above, a storage area network may be implemented with virtualization switches adapted for implementing virtualization functionality as well as standard switches. Each virtualization switch may include one or more "intelligent" virtualization ports as well as one or more standard ports. FIG. 6 is a diagram illustrating an exemplary system in which a virtual enclosure is implemented through the binding of the virtual enclosure ports to various virtualization ports in accordance with various embodiments of the invention. As shown, a virtual enclosure 602 "encloses" two virtual storage units, VLUN1 604 and VLUN2 606. In this example, the virtual enclosure 602 has three virtual enclosure ports, VE1 608, VE2 610, and VE3 612. In addition, four network devices within the storage area network are illustrated. More specifically, four virtualization switches 614, 616, 618, and 620 are illustrated. Each of the virtualization switches has one or more virtualization ports as described above with reference to FIG. 3A. However, in order to simplify the illustration, each virtualization switch is shown to have a single virtualization port, labeled vl, v2, v3, and v4, respectively. Each virtualization port has an associated PWWN (not shown). Communication between the switches may be accomplished by an inter- switch link (not shown) between two switches (or two ports).
Each virtual enclosure port is uniquely identified. More specifically, within a fibre channel network, the virtual enclosure 602 is identified by a NWWN 621. In addition, once each virtual enclosure port is associated with a virtualization port, each of the virtual enclosure ports VE1 608, VE2 610, and
VE3 612 are further identified by a PWWN as well as its FCID, labeled PWWN1 622, PWWN2 624, and PWWN3 626, respectively. More particularly, the virtual enclosure ports may be identified by the NWWN and PWWN until an FCID is assigned to the virtual enclosure ports. An FCID will be further described below with reference to FIG. 8. As shown, the FCIDs that are assigned also identify the virtualization switch associated with the corresponding virtual enclosure port. This is because in Fibre Channel, a switch is seen as a "domain." Thus, the first byte of the FCID of any port in that switch is the same as the domain ID. For instance, the first byte of the FCID assigned to the first virtual enclosure port VE1 608 identifies the virtualization switch 616, shown as switch 50. Similarly, the first byte of the FCID assigned to the second virtual enclosure port VE2 610 identifies the virtualization switch 614, shown as switch 55, while the first byte of the FCID assigned to the third virtual enclosure port VE3 612 identifies the virtualization switch 618, shown as switch 45.
The creation of a virtual enclosure such as virtual enclosure 602 may be performed by a network device such as a virtual enclosure server 627. The virtual enclosure 602 may be created, for example, by a system administrator. Once created, various VLUNs such as VLUNs 604 and 606 may be assigned to the virtual enclosure 602. Such an assignment may be subsequently modified as necessary to include additional VLUNs, remove VLUNs from the virtual enclosure, or otherwise modify the assignment of VLUNs to the virtual enclosure 602. In addition to creating , modifying and maintaining a virtual enclosure, the virtual enclosure server 627 may also inform the virtualization ports of the virtualization switches of their association with the corresponding virtual enclosure ports. In this manner, the virtualization ports may be notified of their responsibility to intercept and handle packets (or frames) directed to those corresponding virtual enclosure ports.
Once a virtual enclosure has been created, one or more hosts 628, 630 may access data in the storage area network via the virtual enclosure 602. More specifically, the hosts 628, 630 may access various physical storage devices, referred to as physical storage units, corresponding to the virtual storage units, shown here as VLUN1 604 and VLUN2 606. For instance, a host may read data from or write data to various PLUNs 632, 634, 636, 638 within the storage area network by sending packets, frames or messages to a virtual address within a VLUN enclosed by a virtual enclosure (and available to the host via the virtual enclosure port). As described above, a virtual enclosure may be created for use in a variety of network environments. FIG. 7 is a process flow diagram illustrating a method of creating a virtual enclosure in accordance with various embodiments of the invention. In this example, the virtual enclosure is for use in a Fibre channel network. Thus, as shown at block 702, a NWWN is associated with the virtual enclosure. In addition, the number of virtual enclosure ports of the virtual enclosure is selected at block 704. Each of these virtual enclosure ports may be identified by a PWWN until they are bound to a virtualization port at block 706 and assigned an FCID at block 708. Once the virtual enclosure is generated, one or more VLUNs are assigned to the virtual enclosure at block 710. Thus, when a host logs into the SAN at block 712, the host may access physical storage units via the virtual enclosure by sending SCSI read and write commands.
Within a Fibre channel network, each SCSI target device is assigned a NWWN. Moreover, each SCSI target port is assigned a PWWN. In addition, a Fibre channel identifier (FCID) identifies a physical location of each SCSI target port, and is therefore associated with the PWWN. FIG. 8 is a diagram illustrating a conventional FCID that may be associated with a virtual enclosure port in accordance with various embodiments of the invention. A FCID is a three byte address identifier, which identifies a switch 802 such as a virtualization switch, an area 804 within the network, and a port 806 of the switch 802 such as a virtualization port.
As described above, in order to create a virtual enclosure, each virtual enclosure port is "bound" to a virtualization port. In accordance with various embodiments of the invention, coordination of binding virtual enclosure ports to virtualization ports is performed by a separate network device. For instance, this network device may be the virtual enclosure server responsible for generating a virtual enclosure. In other words, the virtual enclosure server instructs a virtualization port that it is "bound" to a virtual enclosure port and therefore should handle all messages addressed to that virtual enclosure port (that are received by the virtualization port). In addition to binding a single virtualization port to a virtual enclosure port, it may be desirable to enable additional ports to "trap" messages addressed to the virtual enclosure port. For instance, a virtualization port receiving a message addressed to a virtual enclosure port may be closer to the host than a virtualization port that is "bound" to the virtual enclosure port, and therefore may be better able to provide timely, efficient service. In addition, it may be undesirable to route all messages to a single virtualization port that is bound to a virtual enclosure port, since this may overload a limited number of ports, resulting in ineffective service. Therefore, in accordance with various embodiments of the invention, the virtual enclosure server instructs various virtualization ports to "trap" messages addressed to one or more virtual enclosure ports. In this manner, the power of additional virtualization ports within the SAN to handle messages is leveraged. Accordingly, a virtualization port may service a virtual enclosure even though the virtualization port is not bound to a virtual enclosure port.
A "trapping" virtualization port is preferably an "external" node within the
SAN rather than an internal node. In other words, the virtualization port should be directly connected to a host rather than indirectly connected via one or more other ports (e.g., standard or virtualization ports). This direct connection is particularly important since an internal node may receive a message or request from a host, but the internal node may not receive messages from the target that are directed to the host, since the same return path is not guaranteed. As a result, the virtualization port would not be able to complete this "communication loop," and therefore will not have knowledge of whether a host request has been serviced.
One method of coordinating virtual enclosure binding and trapping functionality of virtualization ports in accordance with various embodiments of the invention is illustrated in FIG. 9. Steps performed at the virtual enclosure server are represented by vertical line 902, and steps performed at virtualization ports 1-3 are represented by vertical lines 904, 906, and 908, respectively. A DNS server and associated switch are represented by vertical lines 912 and 910, respectively. The DNS server 912 is often included in the switch 910. The virtual enclosure server 902 may send various virtualization messages to various ports within the SAN to enable virtualization of storage within a SAN. For instance, as described above, a virtualization message may be a bind message or a trap message, which identifies the PWWN of a virtual enclosure port. The primary difference between a bind message and a trap message is that a bind message indicates that a FCID is to be dynamically assigned to a virtual enclosure port, while a trap message indicates that the FCID that has previously been assigned to another virtual enclosure port is to be obtained so that messages directed to that FCID may be trapped by the receiving port. As shown at 914, the virtual enclosure server sends a bind message 914 instructing the first virtualization port, virtualization portl 904, to handle all messages directed to the first virtual enclosure port, VEP1 that are received by the virtualization portl 904. Each virtual enclosure port such as VEP1 may be uniquely identified by its PWWN. A virtualization message such as a bind or trap message may further indicate that the receiving virtualization port is to obtain an address or identifier assigned to the virtual enclosure port. For instance, the virtualization message may indicate that the address or identifier is to be obtained from a DNS server, which may be specified or unspecified.
Once the bind message is received, the virtualization portl 904 sends a FLOGI message to the switch 910 at 916 to establish a connection with the switch 910. As described above, a FCID is dynamically assigned during login. Thus, the switch 910 assigns a FCID, FCIDl, to the first virtualization port VEP1, which is provided to the DNS server 912 at 918, which then associates the virtual enclosure portl, VEP1 with the FCID FCIDl. The switch 910 then sends an ACCEPT message at 920 indicating that the FCIDl is now associated by the DNS server 912 with the first virtual enclosure port, VEP1. In addition, as shown, the ACCEPT message provides the FCID that has been assigned to the virtualization port 904. A virtualization port keeps track of those virtual enclosure ports with which the virtualization port is bound or for which the virtualization port is responsible for trapping messages. In order to maintain this virtual enclosure port information, the virtualization port may then store the FCID, FCIDl, in a virtualization port table such as that described below with reference to FIG. 10. A DNS REGISTRY is also performed at 922 to inform the DNS server 912 that VEPl understands the SCSI protocol. The DNS server then sends an ACCEPT message at 924. A virtualization response is then sent from the virtualization port at 926 to the virtual enclosure server indicating that the virtualization port is now configured to handle messages addressed to the virtual enclosure port of the virtual enclosure. In this example, the virtualization response indicates that the virtualization port is now bound to the virtual enclosure port. In addition, the FCID assigned to the virtual enclosure port VEPl may also be provided to the virtual enclosure server so that the virtual enclosure server may assign the FCID to the virtual enclosure port. More specifically, the FCID may be provided in the virtualization response. In this manner, the virtual enclosure server obtains the FCIDs of SCSI target ports available to it.
In addition to binding of virtual enclosure ports to virtualization ports within the SAN, the virtual enclosure server 902 may also establish trapping by additional virtualization ports within the SAN as described above. For instance, as shown at 928 a trap message indicating that the virtualization port2 906 is to handle messages addressed to an address or identifier assigned to a particular virtual enclosure port is sent. Since the virtual enclosure server 902 has obtained the FCID assigned to the virtual enclosure port, VEPl, it may be provided in the trap message. Alternatively, the FCID may be obtained from a DNS server, as will be described in this example. In this example, the virtualization port 906 is directed to handle messages addressed to the specified virtual enclosure port (e.g., VEPl) that are received by it. In other words, the virtualization port 906 is instructed to handle messages addressed to the address or identifier assigned to the virtual enclosure port. Similarly, a second trap message is sent at 930 to the virtualization port3 908 indicating that the virtualization port is to handle messages addressed to the first virtual enclosure port.
When the virtualization port2 906 receives the trap message at 928, the virtualization port sends a GET_FCID message at 932 to the DNS server 912 to obtain the address or identifier assigned to the first virtual enclosure port, VEPl. Once it receives the address or identifier at 934, it stores the address or identifier at 936. More specifically, it may update a table such as that described below with reference to FIG. 10. The virtualization port then sends a virtualization response at 938 to the host in response to the virtualization message. For instance, the virtualization response may indicate that the virtualization port has been successfully (or unsuccessfully) configured to handle messages addressed to the virtual enclosure port. The virtualization port may now trap messages addressed to the virtual enclosure port that are received by it as shown at 940.
Similarly, when the virtualization port3 908 receives the trap message at 930, the virtualization port sends a GET_FCID message at 942 to the DNS server 912 to obtain the address or identifier assigned to the first virtual enclosure port, VEPl . Once it receives the address or identifier at 944, it stores the address or identifier at 946. The virtualization port sends a virtualization response at 948 indicating that the virtualization port has been successfully (or unsuccessfully) configured as a trapping port, and begins trapping messages directed to the virtualization port at 950.
As described above, a port within a SAN such as a virtualization port may be configured to handle messages on behalf of multiple virtual enclosure ports. More specifically, a single virtualization port may be bound to multiple virtual enclosure ports as well as be configured as a trapping port for multiple virtual enclosure ports. FIG. 10 is a diagram illustrating an exemplary table that may be maintained by a virtualization port indicating FCIDs to be handled by the virtualization port in accordance with various embodiments of the invention. As shown, a virtualization port table 1000 may be maintained by a virtualization port to keep track of those virtual enclosure ports 1002 with which it is bound or for which it is configured to trap messages. For each virtual enclosure port 1002, it maintains the associated FCID. In this manner, the virtualization port may identify those FCIDs for which it is responsible. When a packet is directed to an FCID for which the virtualization port is not responsible, it merely forwards the packet. Once one or more virtualization ports are configured to handle packets on behalf of a virtual enclosure port (e.g., trapping or bound ports), they may serve as SCSI targets for an initiator such as a host. FIG. 11 is a transaction flow diagram illustrating one method of establishing communication between a host and one or more virtualization ports (e.g., bound to one or more virtual enclosure ports, trapping ports) such that the host can access one or more LUNs in accordance with various embodiments of the invention. As shown, interactions between a host 1102, switch 1104, Domain Name System (DNS) server 1106 and virtualization port 1108 are represented by corresponding labeled vertical lines. As shown, in order for a fibre channel node such as a host 1102 to establish a logical connection to a fabric switch 1104, it performs a fabric login (FLOGI) 1110. As described above, fibre channel uses an address identifier, referred to as an FCID, which is dynamically assigned during login. Thus, the switch 1104 provides an FCID to the host 1102 at 1112. Once the host been assigned a host FCID, the host may perform a DNS query of a DNS server 1106 via a SCSI REPORT command, which requests the SCSI targets that are available, and therefore visible, to the host at block 1114. More specifically, the DNS query determines those SCSI targets that are visible to the host FCID. The FCID of one or more virtual enclosure ports as potential target ports are then provided by the DNS server 1106 to the host 1102 at 1116.
Once the host has the FCIDs of those SCSI target ports available to it, it sends a fibre channel process login command to one or more of the available virtual enclosure SCSI target ports (identified by its FCID) at 1118. As described above, although the packet is addressed to an FCID assigned to a virtual enclosure port, a virtualization port that is bound to the virtual enclosure port or is trapping on behalf of the virtual enclosure port may actually handle these packets. This process login implements a mapping layer, which "maps" fibre channel to SCSI. Once completed, the SCSI virtual enclosure target port(s) 1108 send a fibre channel accept message at 1120. Since Fibre channel is mapped to SCSI, the host can send SCSI commands to the SCSI target virtual enclosure port(s). Now that the host 1102 can send SCSI commands to the virtual enclosure port(s) 1 108, it performs a SCSI process login at 1122 by sending a SCSI process login command to the SCSI target virtual enclosure port 1108. The SCSI virtual enclosure port 1108 then sends a SCSI accept command 1124. Communication between the host and the SCSI target virtual enclosure port via the SCSI protocol is therefore established. For instance, the host may determine which LUNs are available to it, as well as read and write to those LUNs available to it.
Since the host 1102 knows which SCSI targets (and virtual enclosure ports) are available to it and can communicate with each of these ports, it can send a SCSI REPORT LUN command to the SCSI target virtual enclosure port(s) 1108 to determine those LUNs that are visible to the host FCID at 1126. For instance, these LUNs may simply be those VLUNs within the virtual enclosure, or may be a subset of those VLUNs. More specifically, this REPORT message may be sent to the FCIDs assigned to the virtual enclosure ports. The receiving SCSI target virtual enclosure port 1108 determines which LUNs are visible to the host FCID at 1128 and sends a reply message indicating one or more available LUNs (e.g., PLUNs or VLUNs) at 1130. It is important to note that the receiving virtual enclosure port may actually be a virtualization port that is either bound to or trapping on behalf of a virtual enclosure port. Thus, these virtualization ports are also responsible for performing LUN mapping at 1128. One method of LUN mapping will be described in further detail below with reference to FIG. 12. The host 1102 may then send SCSI READ and WRITE commands to a particular VLUN at 1132 via the virtual enclosure port 1108.
Virtualization messages such as trap and bind messages may be implemented in any communication protocol. The protocol that is implemented is preferably a reliable communication protocol, such as TCP sockets or Dynamic Instantiation Protocol (DIP), which runs on top of TCP/IP.
FIG. 12 is a diagram illustrating an exemplary LUN mapping table that may be used at step 1128 of FIG. 11 to perform LUN mapping. As shown, a LUN mapping table 1200 may be used to identify for each initiator (e.g., host) 1202 those VLUNs 1204 of a virtual enclosure that are visible via one or more virtual enclosure ports 1206 (e.g., FCIDs) of the virtual enclosure. For instance, a specified logical unit 1208 such as LUN0 may be used to represent a boot disk. Since it may be desirable to provide a different boot disk for each host, the specified logical unit 1208 is associated with one or more VLUNs 1204 which are made available to the host 1202 via the specified virtual enclosure ports 1206. The virtual enclosure ports 1206 that are available to a particular host 1202 may be specified, or alternatively, a wildcard may be used to indicate that all virtual enclosure ports are available to the host. Thus, when a request for a specified logical unit 1208 is received from the host via one of the virtual enclosure ports, the VLUNs associated with the specified logical unit 1208 are employed to service the request.
FIG. 13 represents a virtual enclosure corresponding to the LUN mapping table of FIG. 12. As shown, virtual enclosure 1302 represents four VLUNS, A 1304, B 1306, C 1308, and D 1310. The virtual enclosure 1302 has two virtual enclosure ports, VEPl 1312 and VEP2 1314. As shown in FIG. 12, host HI sees LUN0 as VLUN A on any virtual enclosure port, host H2 sees LUN0 as VLUN B on any virtual enclosure port, host H3 sees LUN0 as VLUN C on any virtual enclosure port, host H4 sees LUN0 as VLUN D on virtual enclosure port 0 only, and host H5 sees LUN0 as VLUN C on virtual enclosure port 1 only. In this manner, the identity of the VLUNs is invisible to the host. Accordingly, the present invention provides an additional level of storage virtualization.
The LUN mapping table is preferably maintained at a central location such as the virtual enclosure server. However, in accordance with one embodiment, the LUN mapping table is also be provided as well as periodically distributed to the appropriate network devices (e.g., virtualization ports) within the storage area network. For instance, the LUN mapping table may be distributed at host login to a virtual enclosure port. In other words, the virtualization port is provided a LUN map corresponding to the host that has logged in via the virtual enclosure server. Although illustrative embodiments and applications of this invention are shown and described herein, many variations and modifications are possible which remain within the concept, scope, and spirit of the invention, and these variations would become clear to those of ordinary skill in the art after perusal of this application. For instance, the present invention is described as being applied to frames. However, it should be understood that the invention is not limited to such implementations, but instead would equally apply to packets as well. Moreover, the present invention would apply regardless of the context and system in which it is implemented. Thus, broadly speaking, the coordination of binding and trapping by multiple ports need not be performed using a virtual enclosure server as described above, but may be performed in an alternate manner.
In addition, although an exemplary switch is described, the above- described embodiments may be implemented in a variety of network devices (e.g., servers) as well as in a variety of mediums. For instance, instructions and data for implementing the above-described invention may be stored on a disk drive, a hard drive, a floppy disk, a server computer, or a remotely networked computer. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.

Claims

CLAIMSWhat is claimed is:
1. A method of implementing storage virtualization in a storage area network, the method comprising: creating a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network; and assigning an address or identifier to each of the virtual enclosure ports.
2. The method as recited in claim 1, wherein associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network comprises: instructing a port of a network device within the storage area network to handle messages addressed to the address or identifier assigned to the associated virtual enclosure port.
3. The method as recited in claim 1, wherein associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network comprises: instructing a port of a network device within a virtual SAN to handle messages addressed to the address or identifier assigned to the associated virtual enclosure port.
4. The method as recited in claim 1, further comprising: associating a Node World Wide Name with the virtual enclosure.
5. The method as recited in claim 4, further comprising: assigning a Port World Wide Name to each of the virtual enclosure ports such that the Port World Wide Name is associated with an associated port of a network device within the storage area network.
6. The method as recited in claim 1, wherein the port of the network device within the storage area network is a port of a fibre channel device.
7. The method as recited in claim 1, wherein assigning an address or identifier to each of the virtual enclosure ports comprises: assigning a FCID to each of the virtual enclosure ports.
8. The method as recited in claim 1, further comprising: selecting a number of virtual enclosure ports of the virtual enclosure.
9. The method as recited in claim 8, wherein the number of virtual enclosure ports of the virtual enclosure is greater than a number of ports of each network device within the storage area network.
10. The method as recited in claim 1, wherein associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network comprises: associating the virtual enclosure ports with ports of one or more network devices within the storage area network.
11. The method as recited in claim 1 , wherein associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network comprises: binding a port of a network device within the storage area network to one or more of the virtual enclosure ports.
12. The method as recited in claim 11, further comprising: instructing one or more additional ports of one or more network devices within the storage area network to trap messages directed to one of the virtual enclosure ports.
13. The method as recited in claim 1 , wherein one or more of the virtual storage units each comprises a VLUN or other virtual representation of storage on the storage area network.
14. The method as recited in claim 1, further comprising: assigning one or more virtual storage units to the virtual enclosure.
15. The method as recited in claim 14, wherein the one or more virtual storage units each comprise a VLUN or other virtual representation of storage on the storage area network.
16. A computer-readable medium storing thereon computer-readable instructions for implementing storage virtualization in a storage area network, comprising: instructions for creating a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; instructions for associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network; and instructions for assigning an address or identifier to each of the virtual enclosure ports.
17. An apparatus for implementing storage virtualization in a storage area network, comprising: means for creating a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; means for associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network; and means for assigning an address or identifier to each of the virtual enclosure ports.
18. A network device adapted for implementing storage virtualization in a storage area network, comprising: a processor; and a memory, at least one of the processor and the memory being adapted for: creating a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; associating each of the virtual enclosure ports of the virtual enclosure with a port of a network device within the storage area network; and assigning an address or identifier to each of the virtual enclosure ports.
19. A method of performing LUN mapping in a storage area network, the method comprising: accessing a LUN mapping table having one or more entries, each of the entries identifying an initiator in the storage area network, one or more of a set of one or more virtual enclosure ports of a virtual enclosure, and associating a specified logical unit with one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network, wherein the virtual enclosure is adapted for representing the set of one or more virtual storage units and each of the virtual enclosure ports is associated with a port of a network device within the storage area network; and when a request for the specified logical unit is received from the initiator via one of the associated virtual enclosure ports, identifying one of the entries in the LUN mapping table and employing the one or more virtual storage units specified in the entry to service the request.
20. A computer-readable medium storing thereon instructions for performing LUN mapping in a storage area network, comprising: instructions for accessing a LUN mapping table having one or more entries, each of the entries identifying an initiator in the storage area network, one or more of a set of one or more virtual enclosure ports of a virtual enclosure, and associating a specified logical unit with one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network, wherein the virtual enclosure is adapted for representing the set of one or more virtual storage units and each of the virtual enclosure ports is associated with a port of a network device within the storage area network; and instructions for identifying one of the entries in the LUN mapping table and employing the one or more virtual storage units specified in the entry to service the request when a request for the specified logical unit is received from the initiator via one of the associated virtual enclosure ports.
21. A method of implementing storage virtualization in a storage area network, the method comprising: sending a virtualization message to a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and receiving a virtualization response from the port of the network device in response to the virtualization message.
22. The method as recited in claim 21, wherein the virtual enclosure port is identified by a NWWN and a PWWN.
23. The method as recited in claim 21, wherein the virtualization response indicates that the port is configured to handle messages addressed to the virtual enclosure port of the virtual enclosure.
24. The method as recited in claim 21, wherein the virtualization message indicates that the port is to obtain an address or identifier assigned to the virtual enclosure port.
25. The method as recited in claim 24, wherein the virtualization message is a bind message or a trap message.
26. The method as recited in claim 24, wherein the virtualization response comprises the address or identifier assigned to the virtual enclosure port.
27. The method as recited in claim 21, wherein the virtualization message indicates that the port is to obtain an address or identifier assigned to the virtual enclosure port from a DNS server.
28. The method as recited in claim 24, further comprising: receiving the address or identifier assigned to the virtual enclosure port.
29. The method as recited in claim 24, wherein the address or identifier is an FCID.
30. The method as recited in claim 21, wherein the virtualization message indicates that the port is to handle messages addressed to an address or identifier assigned to the virtual enclosure port.
31. The method as recited in claim 30, wherein the address or identifier is an FCID.
32. A computer-readable medium storing thereon computer-readable instructions for implementing storage virtualization in a storage area network, comprising: instructions for sending a virtualization message to a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and instructions for receiving a virtualization response from the port of the network device in response to the virtualization message.
33. An apparatus adapted for implementing storage virtualization in a storage area network, comprising: means for sending a virtualization message to a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and means for receiving a virtualization response from the port of the network device in response to the virtualization message.
34. An apparatus adapted for implementing storage virtualization in a storage area network, comprising: a processor; and a memory, at least one of the processor and the memory being adapted for: sending a virtualization message to a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and receiving a virtualization response from the port of the network device in response to the virtualization message.
35. A method of implementing storage virtualization in a storage area network, the method comprising: receiving a virtualization message at a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and sending a virtualization response from the port of the network device in response to the virtualization message.
36. The method as recited in claim 35, wherein the virtualization message indicates that the port is to obtain an address or identifier assigned to the virtual enclosure port, the method further comprising: obtaining the address or identifier assigned to the virtual enclosure port.
37. The method as recited in claim 36, further comprising: storing the address or identifier.
38. The method as recited in claim 35, wherein the virtualization message indicates that the port is to obtain an address or identifier assigned to the virtual enclosure port from a DNS server, the method further comprising: obtaining the address or identifier assigned to the virtual enclosure port.
39. The method as recited in claim 36, further comprising: sending the address or identifier assigned to the virtual enclosure port.
40. The method as recited in claim 36, wherein the address or identifier is an FCID.
41. The method as recited in claim 35, wherein the virtualization message indicates that the port is to handle messages addressed to an address or identifier assigned to the virtual enclosure port.
42. The method as recited in claim 41, wherein the address or identifier is an FCID.
43. The method as recited in claim 41 , further comprising: handling messages addressed to the address or identifier assigned to the virtual enclosure port.
44. The method as recited in claim 35, further comprising: handling messages addressed to the virtual enclosure port of the virtual enclosure.
45. The method as recited in claim 36, further comprising: handling messages addressed to the address or identifier assigned to the virtual enclosure port.
46. The method as recited in claim 35, further comprising: receiving a report message requesting an identification of one or more of the virtual storage units supported by an address or identifier assigned to one of the virtual enclosure ports; sending a reply message identifying one or more of the virtual storage units.
47. The method as recited in claim 46, wherein the address or identifier is an FCID.
48. The method as recited in claim 46, wherein one or more of the virtual storage units comprises a VLUN or other virtual representation of storage on the storage area network.
49. The method as recited in claim 46, wherein the one or more of the virtual storage units identified in the reply message are those virtual storage units that are visible to an initiator sending the report message.
50. A computer-readable medium storing thereon computer readable instructions for implementing storage virtualization in a storage area network, comprising: instructions for receiving a virtualization message at a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and instructions sending a virtualization response from the port of the network device in response to the virtualization message.
51. A network device adapted for implementing storage virtualization in a storage area network, comprising: means for receiving a virtualization message at a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and means for sending a virtualization response from the port of the network device in response to the virtualization message.
52. A network device adapted for implementing storage virtualization in a storage area network, comprising: a processor; and a memory, at least one of the processor and the memory being adapted for: receiving a virtualization message at a port of a network device within the storage area network, the virtualization message instructing the port to handle messages addressed to a virtual enclosure port of a virtual enclosure, the virtual enclosure having one or more virtual enclosure ports and being adapted for representing one or more virtual storage units, each of the virtual storage units representing one or more physical storage locations on one or more physical storage units of the storage area network; and sending a virtualization response from the port of the network device in response to the virtualization message.
EP02797469A 2002-01-09 2002-12-23 Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure Ceased EP1502179A2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US10/045,883 US7548975B2 (en) 2002-01-09 2002-01-09 Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure
US45883 2002-01-09
PCT/US2002/041079 WO2003060688A2 (en) 2002-01-09 2002-12-23 Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure

Publications (1)

Publication Number Publication Date
EP1502179A2 true EP1502179A2 (en) 2005-02-02

Family

ID=21940363

Family Applications (1)

Application Number Title Priority Date Filing Date
EP02797469A Ceased EP1502179A2 (en) 2002-01-09 2002-12-23 Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure

Country Status (8)

Country Link
US (1) US7548975B2 (en)
EP (1) EP1502179A2 (en)
JP (1) JP4372553B2 (en)
KR (1) KR100927748B1 (en)
CN (1) CN1311326C (en)
AU (1) AU2002361837B2 (en)
CA (1) CA2472992C (en)
WO (1) WO2003060688A2 (en)

Families Citing this family (177)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7171453B2 (en) * 2001-04-19 2007-01-30 Hitachi, Ltd. Virtual private volume method and system
JP4484396B2 (en) * 2001-05-18 2010-06-16 株式会社日立製作所 Turbine blade
US7171434B2 (en) * 2001-09-07 2007-01-30 Network Appliance, Inc. Detecting unavailability of primary central processing element, each backup central processing element associated with a group of virtual logic units and quiescing I/O operations of the primary central processing element in a storage virtualization system
US7472231B1 (en) 2001-09-07 2008-12-30 Netapp, Inc. Storage area network data cache
US7433948B2 (en) * 2002-01-23 2008-10-07 Cisco Technology, Inc. Methods and apparatus for implementing virtualization of storage within a storage area network
US9009427B2 (en) 2001-12-26 2015-04-14 Cisco Technology, Inc. Mirroring mechanisms for storage area networks and network based virtualization
US20070094466A1 (en) * 2001-12-26 2007-04-26 Cisco Technology, Inc., A Corporation Of California Techniques for improving mirroring operations implemented in storage area networks and network based virtualization
US20090259817A1 (en) * 2001-12-26 2009-10-15 Cisco Technology, Inc. Mirror Consistency Checking Techniques For Storage Area Networks And Network Based Virtualization
US20090259816A1 (en) * 2001-12-26 2009-10-15 Cisco Technology, Inc. Techniques for Improving Mirroring Operations Implemented In Storage Area Networks and Network Based Virtualization
US6934799B2 (en) * 2002-01-18 2005-08-23 International Business Machines Corporation Virtualization of iSCSI storage
EP1345113A3 (en) * 2002-03-13 2008-02-06 Hitachi, Ltd. Management server
US6880052B2 (en) * 2002-03-26 2005-04-12 Hewlett-Packard Development Company, Lp Storage area network, data replication and storage controller, and method for replicating data using virtualized volumes
JP2003296037A (en) * 2002-04-05 2003-10-17 Hitachi Ltd Computer system
US9344235B1 (en) * 2002-06-07 2016-05-17 Datacore Software Corporation Network managed volumes
US8200871B2 (en) * 2002-06-28 2012-06-12 Brocade Communications Systems, Inc. Systems and methods for scalable distributed storage processing
US7752361B2 (en) * 2002-06-28 2010-07-06 Brocade Communications Systems, Inc. Apparatus and method for data migration in a storage processing device
US7353305B2 (en) * 2002-06-28 2008-04-01 Brocade Communications Systems, Inc. Apparatus and method for data virtualization in a storage processing device
US7237045B2 (en) * 2002-06-28 2007-06-26 Brocade Communications Systems, Inc. Apparatus and method for storage processing through scalable port processors
US9787524B1 (en) * 2002-07-23 2017-10-10 Brocade Communications Systems, Inc. Fibre channel virtual host bus adapter
US7269168B2 (en) * 2002-07-31 2007-09-11 Brocade Communications Systems, Inc. Host bus adaptor-based virtualization switch
US7120728B2 (en) * 2002-07-31 2006-10-10 Brocade Communications Systems, Inc. Hardware-based translating virtualization switch
US7616631B2 (en) * 2002-08-14 2009-11-10 Lsi Corporation Method and apparatus for debugging protocol traffic between devices in integrated subsystems
US7725568B2 (en) * 2002-09-09 2010-05-25 Netapp, Inc. Method and apparatus for network storage flow control
US8805918B1 (en) 2002-09-11 2014-08-12 Cisco Technology, Inc. Methods and apparatus for implementing exchange management for virtualization of storage within a storage area network
US7774444B1 (en) * 2002-10-16 2010-08-10 Symantec Operating Corporation SAN simulator
JP4139675B2 (en) * 2002-11-14 2008-08-27 株式会社日立製作所 Virtual volume storage area allocation method, apparatus and program thereof
US7263593B2 (en) 2002-11-25 2007-08-28 Hitachi, Ltd. Virtualization controller and data transfer control method
JP2004192105A (en) * 2002-12-09 2004-07-08 Hitachi Ltd Connection device of storage device and computer system including it
US7782784B2 (en) 2003-01-10 2010-08-24 Cisco Technology, Inc. Port analyzer adapter
US7899048B1 (en) 2003-01-15 2011-03-01 Cisco Technology, Inc. Method and apparatus for remotely monitoring network traffic through a generic network
US7606239B2 (en) * 2003-01-31 2009-10-20 Brocade Communications Systems, Inc. Method and apparatus for providing virtual ports with attached virtual devices in a storage area network
JP4237515B2 (en) * 2003-02-07 2009-03-11 株式会社日立グローバルストレージテクノロジーズ Network storage virtualization method and network storage system
US7382776B1 (en) 2003-04-15 2008-06-03 Brocade Communication Systems, Inc. Performing block storage virtualization at a switch
US7460528B1 (en) 2003-04-15 2008-12-02 Brocade Communications Systems, Inc. Processing data packets at a storage service module of a switch
US7739543B1 (en) * 2003-04-23 2010-06-15 Netapp, Inc. System and method for transport-level failover for loosely coupled iSCSI target devices
JP4278444B2 (en) * 2003-06-17 2009-06-17 株式会社日立製作所 Virtual port name management device
US7523201B2 (en) * 2003-07-14 2009-04-21 Network Appliance, Inc. System and method for optimized lun masking
US8239552B2 (en) * 2003-08-21 2012-08-07 Microsoft Corporation Providing client access to devices over a network
US7474666B2 (en) 2003-09-03 2009-01-06 Cisco Technology, Inc. Switch port analyzers
US8165136B1 (en) 2003-09-03 2012-04-24 Cisco Technology, Inc. Virtual port based SPAN
US20050071560A1 (en) * 2003-09-30 2005-03-31 International Business Machines Corp. Autonomic block-level hierarchical storage management for storage networks
US7734748B1 (en) * 2003-10-03 2010-06-08 Nortel Networks Limited Method and apparatus for intelligent management of a network element
US7359993B1 (en) * 2003-10-03 2008-04-15 Nortel Networks Ltd. Method and apparatus for interfacing external resources with a network element
JP4257783B2 (en) * 2003-10-23 2009-04-22 株式会社日立製作所 Logically partitionable storage device and storage device system
US20050091426A1 (en) * 2003-10-23 2005-04-28 Horn Robert L. Optimized port selection for command completion in a multi-ported storage controller system
US20050114464A1 (en) * 2003-10-27 2005-05-26 Shai Amir Virtualization switch and method for performing virtualization in the data-path
US20050114595A1 (en) * 2003-11-26 2005-05-26 Veritas Operating Corporation System and method for emulating operating system metadata to provide cross-platform access to storage volumes
US7669032B2 (en) * 2003-11-26 2010-02-23 Symantec Operating Corporation Host-based virtualization optimizations in storage environments employing off-host storage virtualization
US7934023B2 (en) * 2003-12-01 2011-04-26 Cisco Technology, Inc. Apparatus and method for performing fast fibre channel write operations over relatively high latency networks
US20050138184A1 (en) * 2003-12-19 2005-06-23 Sanrad Ltd. Efficient method for sharing data between independent clusters of virtualization switches
JP2005190036A (en) * 2003-12-25 2005-07-14 Hitachi Ltd Storage controller and control method for storage controller
JP4463042B2 (en) * 2003-12-26 2010-05-12 株式会社日立製作所 Storage system having volume dynamic allocation function
JP4227035B2 (en) 2004-02-03 2009-02-18 株式会社日立製作所 Computer system, management device, storage device, and computer device
JP2005267008A (en) * 2004-03-17 2005-09-29 Hitachi Ltd Method and system for storage management
US20050223180A1 (en) * 2004-03-31 2005-10-06 Philip Derbeko Accelerating the execution of I/O operations in a storage system
US7437506B1 (en) * 2004-04-26 2008-10-14 Symantec Operating Corporation Method and system for virtual storage element placement within a storage area network
JP2005316548A (en) * 2004-04-27 2005-11-10 Hitachi Ltd Computing system having a plurality of type of storage network and information setting method
JP4230410B2 (en) * 2004-05-11 2009-02-25 株式会社日立製作所 Communication quality control device for virtual storage
JP2006024024A (en) * 2004-07-08 2006-01-26 Toshiba Corp Logical disk management method and device
US9264384B1 (en) 2004-07-22 2016-02-16 Oracle International Corporation Resource virtualization mechanism including virtual host bus adapters
US20060101204A1 (en) * 2004-08-25 2006-05-11 Bao Bill Q Storage virtualization
US20060161810A1 (en) * 2004-08-25 2006-07-20 Bao Bill Q Remote replication
JP4646574B2 (en) 2004-08-30 2011-03-09 株式会社日立製作所 Data processing system
US20060047850A1 (en) * 2004-08-31 2006-03-02 Singh Bhasin Harinder P Multi-chassis, multi-path storage solutions in storage area networks
US7062624B2 (en) * 2004-09-29 2006-06-13 Hitachi, Ltd. Method for managing volume groups considering storage tiers
JP2006134208A (en) * 2004-11-09 2006-05-25 Fujitsu Ltd Storage virtualization device
US7571274B2 (en) * 2004-12-30 2009-08-04 Intel Corporation Method and system for virtual enclosure management
US20060271706A1 (en) * 2005-05-24 2006-11-30 Dugan Robert J Method, apparatus, and computer program product for requesting deterministic virtual address identifiers for nodes in a fibre channel network
US8037026B1 (en) * 2005-07-01 2011-10-11 Hewlett-Packard Development Company, L.P. Protected user-controllable volume snapshots
WO2007011164A1 (en) * 2005-07-21 2007-01-25 Kyungpook National University Industry-Academic Cooperation Foundation Virtual storage system and method for managementing virtual storage based on ad-hoc network
US9813283B2 (en) 2005-08-09 2017-11-07 Oracle International Corporation Efficient data transfer between servers and remote peripherals
US8127059B1 (en) * 2005-09-02 2012-02-28 Pmc-Sierra Us, Inc. Apparatus for interconnecting hosts with storage devices
US8072987B1 (en) 2005-09-30 2011-12-06 Emc Corporation Full array non-disruptive data migration
US8107467B1 (en) * 2005-09-30 2012-01-31 Emc Corporation Full array non-disruptive failover
US20070088917A1 (en) * 2005-10-14 2007-04-19 Ranaweera Samantha L System and method for creating and maintaining a logical serial attached SCSI communication channel among a plurality of storage systems
US7644179B1 (en) 2005-12-01 2010-01-05 Cisco Technology, Inc. Inter-VSAN routing with NAT
US8001267B2 (en) * 2005-12-15 2011-08-16 International Business Machines Corporation Apparatus, system, and method for automatically verifying access to a multipathed target at boot time
US7882562B2 (en) * 2005-12-15 2011-02-01 International Business Machines Corporation Apparatus, system, and method for deploying iSCSI parameters to a diskless computing device
US8166166B2 (en) * 2005-12-15 2012-04-24 International Business Machines Corporation Apparatus system and method for distributing configuration parameter
US7769023B2 (en) * 2005-12-21 2010-08-03 Cisco Technology, Inc. Fibre channel traffic redirect scheme using access control lists
US9348530B2 (en) * 2005-12-27 2016-05-24 Emc Corporation Presentation of virtual arrays using n-port ID virtualization
US7697515B2 (en) * 2005-12-27 2010-04-13 Emc Corporation On-line data migration of a logical/virtual storage array
US7697554B1 (en) * 2005-12-27 2010-04-13 Emc Corporation On-line data migration of a logical/virtual storage array by replacing virtual names
US7685395B1 (en) 2005-12-27 2010-03-23 Emc Corporation Spanning virtual arrays across multiple physical storage arrays
US7500134B2 (en) * 2005-12-27 2009-03-03 Emc Corporation Virtual array failover
US7631143B1 (en) * 2006-01-03 2009-12-08 Emc Corporation Data storage system employing virtual disk enclosure
US7549017B2 (en) * 2006-03-07 2009-06-16 Cisco Technology, Inc. Methods and apparatus for selecting a virtualization engine
CN100423491C (en) * 2006-03-08 2008-10-01 杭州华三通信技术有限公司 Virtual network storing system and network storing equipment thereof
US20070271560A1 (en) * 2006-05-18 2007-11-22 Microsoft Corporation Deploying virtual machine to host based on workload characterizations
US8274993B2 (en) * 2006-06-16 2012-09-25 Cisco Technology, Inc. Fibre channel dynamic zoning
US8539177B1 (en) * 2006-06-29 2013-09-17 Emc Corporation Partitioning of a storage array into N-storage arrays using virtual array non-disruptive data migration
US8589504B1 (en) 2006-06-29 2013-11-19 Emc Corporation Full array non-disruptive management data migration
US7757059B1 (en) 2006-06-29 2010-07-13 Emc Corporation Virtual array non-disruptive management data migration
US8452928B1 (en) * 2006-06-29 2013-05-28 Emc Corporation Virtual array non-disruptive migration of extended storage functionality
US8533408B1 (en) * 2006-06-29 2013-09-10 Emc Corporation Consolidating N-storage arrays into one storage array using virtual array non-disruptive data migration
US8583861B1 (en) * 2006-06-29 2013-11-12 Emc Corporation Presentation of management functionality of virtual arrays
JP5087249B2 (en) * 2006-09-06 2012-12-05 株式会社日立製作所 Storage system and storage system control method
US7761738B2 (en) * 2006-09-07 2010-07-20 International Business Machines Corporation Establishing communications across virtual enclosure boundaries
US9104599B2 (en) 2007-12-06 2015-08-11 Intelligent Intellectual Property Holdings 2 Llc Apparatus, system, and method for destaging cached data
US20080140724A1 (en) 2006-12-06 2008-06-12 David Flynn Apparatus, system, and method for servicing object requests within a storage controller
US8443134B2 (en) * 2006-12-06 2013-05-14 Fusion-Io, Inc. Apparatus, system, and method for graceful cache device degradation
US8706968B2 (en) * 2007-12-06 2014-04-22 Fusion-Io, Inc. Apparatus, system, and method for redundant write caching
US8489817B2 (en) 2007-12-06 2013-07-16 Fusion-Io, Inc. Apparatus, system, and method for caching data
JP5081441B2 (en) * 2006-12-13 2012-11-28 株式会社日立製作所 Storage control device and control method of storage control device
US8489811B1 (en) * 2006-12-29 2013-07-16 Netapp, Inc. System and method for addressing data containers using data set identifiers
US7711979B2 (en) * 2007-02-16 2010-05-04 Symantec Corporation Method and apparatus for flexible access to storage facilities
US7958263B2 (en) * 2007-02-20 2011-06-07 International Business Machines Corporation Address reduction for data storage enclosures
US8312046B1 (en) 2007-02-28 2012-11-13 Netapp, Inc. System and method for enabling a data container to appear in a plurality of locations in a super-namespace
US8661518B2 (en) * 2007-06-13 2014-02-25 Cisco Technology, Inc. N-port ID virtualization (NPIV) proxy module, NPIV proxy switching system and methods
US7856022B1 (en) * 2007-06-28 2010-12-21 Emc Corporation Non-disruptive data migration with external virtualization engine
US9063896B1 (en) 2007-06-29 2015-06-23 Emc Corporation System and method of non-disruptive data migration between virtual arrays of heterogeneous storage arrays
US9098211B1 (en) * 2007-06-29 2015-08-04 Emc Corporation System and method of non-disruptive data migration between a full storage array and one or more virtual arrays
US9063895B1 (en) 2007-06-29 2015-06-23 Emc Corporation System and method of non-disruptive data migration between heterogeneous storage arrays
US8131857B2 (en) * 2007-07-11 2012-03-06 Cisco Technology, Inc. Fibre channel intelligent target management system
US8209365B2 (en) * 2007-07-23 2012-06-26 Hewlett-Packard Development Company, L.P. Technique for virtualizing storage using stateless servers
US7836226B2 (en) 2007-12-06 2010-11-16 Fusion-Io, Inc. Apparatus, system, and method for coordinating storage requests in a multi-processor/multi-thread environment
US9519540B2 (en) 2007-12-06 2016-12-13 Sandisk Technologies Llc Apparatus, system, and method for destaging cached data
US8028062B1 (en) * 2007-12-26 2011-09-27 Emc Corporation Non-disruptive data mobility using virtual storage area networks with split-path virtualization
US8380674B1 (en) * 2008-01-09 2013-02-19 Netapp, Inc. System and method for migrating lun data between data containers
US10108460B2 (en) * 2008-02-28 2018-10-23 International Business Machines Corporation Method and system for integrated deployment planning for virtual appliances
US8914540B1 (en) * 2008-07-01 2014-12-16 Cisco Technology, Inc. Multi-fabric SAN based data migration
US9973446B2 (en) 2009-08-20 2018-05-15 Oracle International Corporation Remote shared server peripherals over an Ethernet network for resource virtualization
CN102045603B (en) * 2009-10-19 2013-09-11 杭州华三通信技术有限公司 Negotiation method of fiber channel (FC) port and FC exchanger system
US20110153905A1 (en) * 2009-12-23 2011-06-23 Hitachi, Ltd. Method and apparatus for i/o path switching
US8468302B2 (en) 2010-01-14 2013-06-18 Hitachi, Ltd. Storage system
US8473947B2 (en) * 2010-01-18 2013-06-25 Vmware, Inc. Method for configuring a physical adapter with virtual function (VF) and physical function (PF) for controlling address translation between virtual disks and physical storage regions
US8402196B2 (en) * 2010-03-04 2013-03-19 Xyratex Technology Limited Storage assembly, a physical expander and a method
US8341457B2 (en) * 2010-03-11 2012-12-25 Lsi Corporation System and method for optimizing redundancy restoration in distributed data layout environments
CN102412978B (en) * 2010-09-21 2014-04-16 杭州华三通信技术有限公司 Method for carrying out network configuration for VM and system thereof
US9331963B2 (en) 2010-09-24 2016-05-03 Oracle International Corporation Wireless host I/O using virtualized I/O controllers
US8719253B2 (en) * 2010-12-01 2014-05-06 Cisco Technology, Inc. Method and apparatus for efficiently organizing hierarchical QoS policies
US20120155461A1 (en) * 2010-12-16 2012-06-21 Electronics And Telecommunications Research Institute Apparatus for managing virtual network
US8725691B1 (en) * 2010-12-16 2014-05-13 Emc Corporation Dynamic LUN resizing in a replication environment
WO2012106362A2 (en) 2011-01-31 2012-08-09 Fusion-Io, Inc. Apparatus, system, and method for managing eviction of data
US9141527B2 (en) 2011-02-25 2015-09-22 Intelligent Intellectual Property Holdings 2 Llc Managing cache pools
US8712962B1 (en) * 2011-12-01 2014-04-29 Emc Corporation Snapshots in de-duplication
US9767032B2 (en) 2012-01-12 2017-09-19 Sandisk Technologies Llc Systems and methods for cache endurance
US9251086B2 (en) 2012-01-24 2016-02-02 SanDisk Technologies, Inc. Apparatus, system, and method for managing a cache
CN102662610B (en) * 2012-04-20 2016-05-11 中国科学院遥感应用研究所 A kind of Remote Sensing Image Management System and method
US9710762B2 (en) * 2012-06-06 2017-07-18 Juniper Networks, Inc. Dynamic logging
US9819737B2 (en) 2012-08-23 2017-11-14 Cisco Technology, Inc. System and method for policy based fibre channel zoning for virtualized and stateless computing in a network environment
US9083550B2 (en) 2012-10-29 2015-07-14 Oracle International Corporation Network virtualization over infiniband
US9304875B2 (en) * 2012-11-30 2016-04-05 International Business Machines Corporation Dynamically tracking logical units moving between input/output ports of a storage area network target
US10013185B2 (en) * 2013-03-15 2018-07-03 Srinivasan Viswanathan Mapping systems and methods of an accelerated application-oriented middleware layer
US9503523B2 (en) * 2013-09-20 2016-11-22 Cisco Technology, Inc. Hybrid fibre channel storage with end-to-end storage provisioning and external connectivity in a storage area network environment
CN104636076B (en) * 2013-11-15 2017-12-05 中国电信股份有限公司 A kind of distributed block device drives method and system for cloud storage
CN104750427B (en) * 2013-12-27 2019-05-17 伊姆西公司 Store device and method thereof
US9933946B2 (en) * 2014-09-15 2018-04-03 Hewlett Packard Enterprise Development Lp Fibre channel storage array methods for port management
US9853873B2 (en) 2015-01-10 2017-12-26 Cisco Technology, Inc. Diagnosis and throughput measurement of fibre channel ports in a storage area network environment
US20160277498A1 (en) * 2015-03-20 2016-09-22 Intel Corporation Location and boundary controls for storage volumes
US9900250B2 (en) 2015-03-26 2018-02-20 Cisco Technology, Inc. Scalable handling of BGP route information in VXLAN with EVPN control plane
US10222986B2 (en) 2015-05-15 2019-03-05 Cisco Technology, Inc. Tenant-level sharding of disks with tenant-specific storage modules to enable policies per tenant in a distributed storage system
US11588783B2 (en) 2015-06-10 2023-02-21 Cisco Technology, Inc. Techniques for implementing IPV6-based distributed storage space
US10778765B2 (en) 2015-07-15 2020-09-15 Cisco Technology, Inc. Bid/ask protocol in scale-out NVMe storage
US9892075B2 (en) 2015-12-10 2018-02-13 Cisco Technology, Inc. Policy driven storage in a microserver computing environment
JP6734058B2 (en) 2016-01-27 2020-08-05 株式会社バイオス Control device
US10140172B2 (en) 2016-05-18 2018-11-27 Cisco Technology, Inc. Network-aware storage repairs
US20170351639A1 (en) 2016-06-06 2017-12-07 Cisco Technology, Inc. Remote memory access using memory mapped addressing among multiple compute nodes
US10664169B2 (en) 2016-06-24 2020-05-26 Cisco Technology, Inc. Performance of object storage system by reconfiguring storage devices based on latency that includes identifying a number of fragments that has a particular storage device as its primary storage device and another number of fragments that has said particular storage device as its replica storage device
US11563695B2 (en) 2016-08-29 2023-01-24 Cisco Technology, Inc. Queue protection using a shared global memory reserve
US10545914B2 (en) 2017-01-17 2020-01-28 Cisco Technology, Inc. Distributed object storage
US11070512B2 (en) * 2017-02-10 2021-07-20 International Business Machines Corporation Server port virtualization for guest logical unit number (LUN) masking in a host direct attach configuration
US10243823B1 (en) 2017-02-24 2019-03-26 Cisco Technology, Inc. Techniques for using frame deep loopback capabilities for extended link diagnostics in fibre channel storage area networks
US10713203B2 (en) 2017-02-28 2020-07-14 Cisco Technology, Inc. Dynamic partition of PCIe disk arrays based on software configuration / policy distribution
US10254991B2 (en) 2017-03-06 2019-04-09 Cisco Technology, Inc. Storage area network based extended I/O metrics computation for deep insight into application performance
US10560525B2 (en) * 2017-06-21 2020-02-11 Juniper Networks, Inc. Synchronization between virtual network functions and host systems
US10303534B2 (en) 2017-07-20 2019-05-28 Cisco Technology, Inc. System and method for self-healing of application centric infrastructure fabric memory
US10404596B2 (en) 2017-10-03 2019-09-03 Cisco Technology, Inc. Dynamic route profile storage in a hardware trie routing table
US10942666B2 (en) 2017-10-13 2021-03-09 Cisco Technology, Inc. Using network device replication in distributed storage clusters
CN110502184B (en) * 2018-05-17 2021-01-05 杭州海康威视系统技术有限公司 Data storage method, data reading method, device and system
CN108920101B (en) * 2018-06-27 2021-03-30 河北工业大学 Soft and hard cooperative dynamic load storage method
US11137913B2 (en) 2019-10-04 2021-10-05 Hewlett Packard Enterprise Development Lp Generation of a packaged version of an IO request
CN113297116A (en) * 2020-04-08 2021-08-24 阿里巴巴集团控股有限公司 Information configuration method, port access method, device and storage medium
US11119965B1 (en) 2020-04-20 2021-09-14 International Business Machines Corporation Virtualized fabric controller for a storage area network
US11683372B2 (en) 2020-11-24 2023-06-20 International Business Machines Corporation Virtualized fabric management server for storage area network
US11709699B2 (en) 2020-11-24 2023-07-25 International Business Machines Corporation Virtualized fabric name server for storage area network
US11522814B2 (en) 2020-11-24 2022-12-06 International Business Machines Corporation Virtualized fabric login server for storage area network

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6260120B1 (en) * 1998-06-29 2001-07-10 Emc Corporation Storage mapping and partitioning among multiple host processors in the presence of login state changes and host controller replacement

Family Cites Families (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5617421A (en) * 1994-06-17 1997-04-01 Cisco Systems, Inc. Extended domain computer network using standard links
US5809285A (en) * 1995-12-21 1998-09-15 Compaq Computer Corporation Computer system having a virtual drive array controller
US6035105A (en) * 1996-01-02 2000-03-07 Cisco Technology, Inc. Multiple VLAN architecture system
US5742604A (en) * 1996-03-28 1998-04-21 Cisco Systems, Inc. Interswitch link mechanism for connecting high-performance network switches
US5764636A (en) * 1996-03-28 1998-06-09 Cisco Technology, Inc. Color blocking logic mechanism for a high-performance network switch
US5740171A (en) * 1996-03-28 1998-04-14 Cisco Systems, Inc. Address translation mechanism for a high-performance network switch
US6101497A (en) * 1996-05-31 2000-08-08 Emc Corporation Method and apparatus for independent and simultaneous access to a common data set
EP0972247B1 (en) * 1996-08-02 2004-03-17 Hewlett-Packard Company Method and apparatus for allowing distributed control of shared resources
US6202135B1 (en) * 1996-12-23 2001-03-13 Emc Corporation System and method for reconstructing data associated with protected storage volume stored in multiple modules of back-up mass data storage facility
US6209059B1 (en) * 1997-09-25 2001-03-27 Emc Corporation Method and apparatus for the on-line reconfiguration of the logical volumes of a data storage system
US6188694B1 (en) * 1997-12-23 2001-02-13 Cisco Technology, Inc. Shared spanning tree protocol
US6208649B1 (en) * 1998-03-11 2001-03-27 Cisco Technology, Inc. Derived VLAN mapping technique
US6295575B1 (en) * 1998-06-29 2001-09-25 Emc Corporation Configuring vectors of logical storage units for data storage partitioning and sharing
US6269381B1 (en) * 1998-06-30 2001-07-31 Emc Corporation Method and apparatus for backing up data before updating the data and for restoring from the backups
US6269431B1 (en) * 1998-08-13 2001-07-31 Emc Corporation Virtual storage and block level direct access of secondary storage for recovery of backup data
US6148414A (en) 1998-09-24 2000-11-14 Seek Systems, Inc. Methods and systems for implementing shared disk array management functions
US6266705B1 (en) * 1998-09-29 2001-07-24 Cisco Systems, Inc. Look up mechanism and associated hash table for a network switch
US6226771B1 (en) * 1998-12-14 2001-05-01 Cisco Technology, Inc. Method and apparatus for generating error detection data for encapsulated frames
JP2000242434A (en) 1998-12-22 2000-09-08 Hitachi Ltd Storage device system
US6542961B1 (en) * 1998-12-22 2003-04-01 Hitachi, Ltd. Disk storage system including a switch
US6400730B1 (en) * 1999-03-10 2002-06-04 Nishan Systems, Inc. Method and apparatus for transferring data between IP network devices and SCSI and fibre channel devices over an IP network
US6877044B2 (en) * 2000-02-10 2005-04-05 Vicom Systems, Inc. Distributed storage management platform architecture
US20020103889A1 (en) * 2000-02-11 2002-08-01 Thomas Markson Virtual storage layer approach for dynamically associating computer storage with processing hosts
US20020120741A1 (en) * 2000-03-03 2002-08-29 Webb Theodore S. Systems and methods for using distributed interconnects in information management enviroments
US6898670B2 (en) * 2000-04-18 2005-05-24 Storeage Networking Technologies Storage virtualization in a storage area network
US6772231B2 (en) * 2000-06-02 2004-08-03 Hewlett-Packard Development Company, L.P. Structure and process for distributing SCSI LUN semantics across parallel distributed components
US6779094B2 (en) * 2000-06-19 2004-08-17 Storage Technology Corporation Apparatus and method for instant copy of data by writing new data to an additional physical storage area
US6952734B1 (en) * 2000-08-21 2005-10-04 Hewlett-Packard Development Company, L.P. Method for recovery of paths between storage area network nodes with probationary period and desperation repair
US6847647B1 (en) * 2000-09-26 2005-01-25 Hewlett-Packard Development Company, L.P. Method and apparatus for distributing traffic over multiple switched fiber channel routes
US6978300B1 (en) * 2000-10-19 2005-12-20 International Business Machines Corporation Method and apparatus to perform fabric management
US7165096B2 (en) * 2000-12-22 2007-01-16 Data Plow, Inc. Storage area network file system
WO2002065275A1 (en) * 2001-01-11 2002-08-22 Yottayotta, Inc. Storage virtualization system and methods
US6748502B2 (en) * 2001-01-12 2004-06-08 Hitachi, Ltd. Virtual volume storage
US6880062B1 (en) * 2001-02-13 2005-04-12 Candera, Inc. Data mover mechanism to achieve SAN RAID at wire speed
US7200144B2 (en) * 2001-10-18 2007-04-03 Qlogic, Corp. Router and methods using network addresses for virtualization
JP2003162439A (en) * 2001-11-22 2003-06-06 Hitachi Ltd Storage system and control method therefor
US6986015B2 (en) * 2001-12-10 2006-01-10 Incipient, Inc. Fast path caching
US7155494B2 (en) * 2002-01-09 2006-12-26 Sancastle Technologies Ltd. Mapping between virtual local area networks and fibre channel zones
US6983303B2 (en) * 2002-01-31 2006-01-03 Hewlett-Packard Development Company, Lp. Storage aggregator for enhancing virtualization in data storage networks
JP3993773B2 (en) * 2002-02-20 2007-10-17 株式会社日立製作所 Storage subsystem, storage control device, and data copy method
US6907419B1 (en) * 2002-02-27 2005-06-14 Storage Technology Corporation Method, system, and product for maintaining within a virtualization system a historical performance database for physical devices
US8051197B2 (en) * 2002-03-29 2011-11-01 Brocade Communications Systems, Inc. Network congestion management systems and methods
US6683883B1 (en) * 2002-04-09 2004-01-27 Sancastle Technologies Ltd. ISCSI-FCP gateway
US7353305B2 (en) 2002-06-28 2008-04-01 Brocade Communications Systems, Inc. Apparatus and method for data virtualization in a storage processing device
US7237045B2 (en) 2002-06-28 2007-06-26 Brocade Communications Systems, Inc. Apparatus and method for storage processing through scalable port processors
US7120728B2 (en) * 2002-07-31 2006-10-10 Brocade Communications Systems, Inc. Hardware-based translating virtualization switch
US7269168B2 (en) 2002-07-31 2007-09-11 Brocade Communications Systems, Inc. Host bus adaptor-based virtualization switch
US7467406B2 (en) 2002-08-23 2008-12-16 Nxp B.V. Embedded data set processing
US7352706B2 (en) * 2002-09-16 2008-04-01 Finisar Corporation Network analysis scalable analysis tool for multiple protocols
US7277431B2 (en) * 2002-10-31 2007-10-02 Brocade Communications Systems, Inc. Method and apparatus for encryption or compression devices inside a storage area network fabric
US20050050211A1 (en) 2003-08-29 2005-03-03 Kaul Bharat B. Method and apparatus to manage network addresses
US20050076113A1 (en) * 2003-09-12 2005-04-07 Finisar Corporation Network analysis sample management process
US20050091426A1 (en) * 2003-10-23 2005-04-28 Horn Robert L. Optimized port selection for command completion in a multi-ported storage controller system
US7934023B2 (en) * 2003-12-01 2011-04-26 Cisco Technology, Inc. Apparatus and method for performing fast fibre channel write operations over relatively high latency networks

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6260120B1 (en) * 1998-06-29 2001-07-10 Emc Corporation Storage mapping and partitioning among multiple host processors in the presence of login state changes and host controller replacement

Also Published As

Publication number Publication date
JP2006505831A (en) 2006-02-16
KR100927748B1 (en) 2009-11-20
KR20040068368A (en) 2004-07-30
WO2003060688A3 (en) 2004-12-09
CN1311326C (en) 2007-04-18
AU2002361837B2 (en) 2009-06-04
CA2472992A1 (en) 2003-07-24
US20030131182A1 (en) 2003-07-10
CN1623136A (en) 2005-06-01
US7548975B2 (en) 2009-06-16
JP4372553B2 (en) 2009-11-25
WO2003060688A2 (en) 2003-07-24
CA2472992C (en) 2013-02-19
AU2002361837A1 (en) 2003-07-30

Similar Documents

Publication Publication Date Title
CA2472992C (en) Methods and apparatus for implementing virtualization of storage within a storage area network through a virtual enclosure
CA2473832C (en) Methods and apparatus for implementing virtualization of storage within a storage area network
US9733868B2 (en) Methods and apparatus for implementing exchange management for virtualization of storage within a storage area network
US7437507B2 (en) Online restriping technique for distributed network based virtualization
AU2003238219A1 (en) Methods and apparatus for implementing virtualization of storage within a storage area network
US7506124B2 (en) Apparatus and methods for facilitating data tapping with host clustering in a storage area network
JP5026283B2 (en) Collaborative shared storage architecture
US7609649B1 (en) Methods and apparatus for improving network based virtualization performance
US8549240B2 (en) Apparatus and methods for an appliance to recover data from a storage device
US20050033878A1 (en) Apparatus and method for data virtualization in a storage processing device
US20050232285A1 (en) System and method of providing network node services
US20040141498A1 (en) Apparatus and method for data snapshot processing in a storage processing device
US7685223B1 (en) Network-wide service discovery
US20040143639A1 (en) Apparatus and method for data replication in a storage processing device

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20040806

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LI LU MC NL PT SE SI SK TR

AX Request for extension of the european patent

Extension state: AL LT LV MK RO

17Q First examination report despatched

Effective date: 20070608

REG Reference to a national code

Ref country code: DE

Ref legal event code: R003

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN REFUSED

18R Application refused

Effective date: 20121212