WO2002037212A2 - Modele de donnes pouvant etre utilise pour l'approvisionnement automatique de dispositifs de stockage de donnees centralises - Google Patents

Modele de donnes pouvant etre utilise pour l'approvisionnement automatique de dispositifs de stockage de donnees centralises Download PDF

Info

Publication number
WO2002037212A2
WO2002037212A2 PCT/US2001/042859 US0142859W WO0237212A2 WO 2002037212 A2 WO2002037212 A2 WO 2002037212A2 US 0142859 W US0142859 W US 0142859W WO 0237212 A2 WO0237212 A2 WO 0237212A2
Authority
WO
WIPO (PCT)
Prior art keywords
entities
storage device
switch
data model
host
Prior art date
Application number
PCT/US2001/042859
Other languages
English (en)
Other versions
WO2002037212A3 (fr
Inventor
Gerardo Lopez-Fernandez
Christine Hsieh
Glenn Ferguson
Original Assignee
Loudcloud, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/699,347 external-priority patent/US6751702B1/en
Application filed by Loudcloud, Inc. filed Critical Loudcloud, Inc.
Priority to AU2002214679A priority Critical patent/AU2002214679A1/en
Publication of WO2002037212A2 publication Critical patent/WO2002037212A2/fr
Publication of WO2002037212A3 publication Critical patent/WO2002037212A3/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0629Configuration or reconfiguration of storage systems
    • G06F3/0637Permissions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/062Securing storage systems
    • G06F3/0622Securing storage systems in relation to access
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/067Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]

Definitions

  • the present invention relates generally to centralized storage devices, and more particularly to automated provisioning of centralized storage devices to be used in network computing applications.
  • Storage networks are used to tie multiple hosts to a single storage system, and may be either a storage area network (SAN), or a network attached storage system (NAS). These two types of storage networks differ primarily in the manner in which the devices on the network are attached to the storage system.
  • SAN storage area network
  • NAS network attached storage system
  • the devices are attached to the centralized storage device by way of channels, or direct connections from the devices to the centralized storage device.
  • NAS the devices of the computer network are attached to the centralized storage device by way of a network, or virtual, connection.
  • storage devices in SANs are considered to be channel attached devices
  • storage devices in NASs are known as network attached devices.
  • many different types of storage devices may be used.
  • RAID redundant array of independent disks
  • a RAID uses a controller and two or more disk drives to store data.
  • RAID systems have different configuration levels, such as RAIDO, RAID1, RAID2, RAID3, RAID4, RAID5, RAID6, RAID7, RAID 10, and RAID53.
  • RAIDO redundant array of independent disks
  • RAID1, RAID2, RAID3, RAID4, RAID5, RAID6, RAID7, RAID 10, and RAID53 Depending upon the specific configuration of the RAID device used for centralized data storage, various advantages may be obtained. Some of the advantages generally obtained through use of a RAID device include increased input/output (I/O) performance, increased fault tolerance, and data redundancy. The degree to which each of these advantages is obtained depends upon the specific RAID configuration. For more information regarding RAID technology in general, a general description of each of the RAID configurations can be found on the Internet at the following URL: http://www.raid5.com.
  • Data striping is a technique whereby data elements are broken into specific blocks and written to different disks within the disk array of the RAID storage device. This improves access time as the controller, which accesses each of the disks within the disk array, may spread the load of I/O requests across many channels and many disk drives. Additionally, data may be backed up using various data redundancy algorithms, such as parity storage algorithms, or the like.
  • RAID storage devices One of the leading manufactures of RAID storage devices is EMC Corporation of Hopkinton, Massachusetts. EMC Corporation manufactures a variety of RAID storage devices that may be used in storage networks such as
  • EMC devices manufactured by EMC Corporation, are generally highly scalable and provide high availability of information to network clients. While EMC devices are among some of the more commonly used RAID devices for a large storage network applications, other systems using RAID technology are also used in many storage network applications.
  • RAID storage devices such as EMC devices for example
  • EMC devices EMC devices
  • client A have access to only client A's data, and that no other clients, such as clients B and C have access to client A's data, e.g., financial transaction information, such as credit card account information, or proprietary business operations information.
  • financial transaction information such as credit card account information, or proprietary business operations information.
  • a central storage device 8 e.g., an EMC frame
  • the number of ports provided to the central storage device 8 may be very limited, e.g., twelve.
  • using the direct attachment scheme depicted in Figure 1 would only permit the central storage device to be connected to N hosts, where N is defined by the number of ports provided to the central storage device 8.
  • a switching matrix 18 can be introduced between the central storage device 8 and the hosts which store their data on the central storage device as shown in Figure 2.
  • the switching matrix 18 may, for example, be comprised of a number of fiber switches (not shown).
  • a larger number of hosts 1-Y can share the limited number of ports 1-N available at the central storage device 8. This provides a more efficient usage of the capabilities of the central storage device 8, but at the expense of introducing additional complexity into the pathway between any given host and its data. This complexity makes it particularly challenging to accomplish the task of provisioning these types of central storage devices, i.e., to identify and allocate storage space to any of the hosts 1-Y on an ongoing basis.
  • these objectives are achieved by the provision of a data model for characterizing a storage device, a system and method for managing data storage device on a network, and a system and method for modeling data of a network storage device.
  • An exemplary data model for characterizing a data storage device connected to a computer network may include a plurality of related entities. Each entity is associated with one or more entities in a one-to-many relationship or a many-to-one relationship. Each entity is also characterized by a variable set, the values of which identify physical instances of each entity in the data model, e.g. , a storage device or a switch. The variable sets also include primary keys which interrelate the variable sets of interconnected entities within the model.
  • a method for allocating storage to a host within a central data storage device having at least one switch disposed therebetween using a data model may, for example, include the steps of: (a) informing the central storage device that the host is authorized to access a predetermined storage area, the predetermined storage area being a subset of said unallocated storage space; (b) creating a path through the switch between the central data storage device and the host; and (c) informing the host that the predetermined storage area has been allocated thereto, wherein at least one of steps (a)-(c) is performed using information extracted from a data model associated with said the storage device.
  • Figure 1 depicts a conventional direct attachment topology between a central storage device and a plurality of hosts.
  • Figure 2 depicts a conventional attachment topology wherein a switching matrix has been inserted between the central storage device and a plurality of hosts.
  • Figure 3 depicts an attachment topology according to the present invention.
  • Figure 4 is a diagram illustrating details of storage media used in connection with the present invention.
  • Figure 5 is a diagram depicting ports which are available at the central storage device.
  • Figure 6 illustrates an exemplary zoning of the ports of a switch.
  • Figure 7 is a flow diagram depicting an overall method for provisioning data storage devices.
  • Figure 8 is an entity relationship diagram of the data model of a network storage device used in connection with one embodiment of the present invention.
  • Figures 9 and 10 are flow diagram illustrating steps involving defining commands used to perform some of the provisioning steps of Figure 7.
  • Figure 11 is a flow diagram illustrating steps for automated provisioning of a central storage device using a data model according to an exemplary embodiment of the present invention.
  • FIG. 3 An exemplary storage network of the type in which the present invention can be employed is illustrated in block diagram form in Figure 3.
  • a central storage device 30, e.g., an EMC frame is connected via a plurality of optical links to a switching matrix 32.
  • four fiber switches 34, 36, 38 and 40 are employed in the switching matrix 32, each of which are connected to the central storage device 30 using three ports and three optical links.
  • the switches 34, 36, 38 and 40 are each connected to two of the other switches 34, 36, 38 and 40 as shown to enable a single switch configuration to promulgate throughout the switching matrix 32 rather than requiring that each switch in the matrix be separately configured.
  • FIG. 3 is only one exemplary manner in which a switching matrix 32 can be configured for interfacing between central storage device 30 and a plurality of hosts and that the present invention is applicable to any such configurations.
  • a host 42 is illustrated in Figure 3, to permit the exemplary configuration of the switching matrix 32 to stand out more clearly, those skilled in the art will further appreciate that a large number of hosts may be connected to the switching matrix 32 so as to access their data stored in the central data storage device 30.
  • An access terminal 43 is also shown which permits an operator to have access to each of the switches 34, 36, 38 and 40, as well as the central storage device 30 and the host 42 for configuration purposes associated with the provisioning of the central storage device 30 which will be described in more detail below.
  • FIG. 4 is a diagram of the storage media contained within an exemplary central storage device 30. It will be recognized by those skilled in the art that similar diagrams may be devised for specific network storage devices, such as EMC devices and other devices, by way of modifications to the elements of this exemplary diagram.
  • multiple segments of the storage medium are defined as logical volumes (sometimes referred to as "hypers") 202, 204, 206, 208, 210, 212, 214, 216, and 218.
  • Each of these volumes is defined by a beginning and ending address, and has a field indicating the size of the volume.
  • each volume contains a volume identification field that uniquely identifies that particular volume within the central storage device 30.
  • Volumes may be of different sizes, as illustrated by the varying widths of the volumes in Figure 4.
  • the first volume 202 is approximately one half of the size of the third volume 206, and approximately one third of the size of the fifth volume 210. That is, if the first volume 202 has a size of one gigabyte, the third volume 206 would have a size of approximately two gigabytes, and the fifth volume 210 would have a size of approximately three gigabytes.
  • the storage medium illustrated in the volume diagram 200 of Figure 4 is also characterized by larger segments of memory called groups (sometimes referred to as"metas").
  • a group is made up of multiple, contiguous volumes within the storage medium.
  • group A 220 is made up of the first five volumes 202, 204, 206, 208, 210.
  • Group B 222 is made up of the next four contiguous volumes 212, 214, 216, 218.
  • the storage medium 200 of Figure 4 may include a larger number of groups, however only two are illustrated therein for the sake of convenience.
  • the groups of the storage medium 200 are defined by the beginning address of the first volume contained within that group, and the ending address of the last volume of the contiguous of volumes contained within the group.
  • group A 220 is defined by the beginning address of volume 202, which is the first volume contained within group A 220, and the ending address of volume 210, which is the last contiguous volume contained within group 220.
  • group B 222 is defined by the beginning address of volume 212, which is the first volume contained within group B 222, and the ending address of group 218, which is the last contiguous volume contained within group B 222.
  • the number of groups and volumes defined within a central storage device 30 may vary according to the various data characteristics intended to be stored thereon.
  • the definition groups and volumes need not be limited to those illustrated in Figure 4, which are provided for illustration purposes only.
  • an access panel 300 is provided for allowing direct connections to the central storage device for accessing data contained on the storage mediums thereof. This access is provided by way of multiple adaptors 302, 304, 306, 308, 310, 312. Each of these adaptors may be an individual computer interface card, which may be interchangeable, or more permanent.
  • each of these adaptors 302, 304, 306, 308, 310, 312 is one or more adaptor ports 314, 316, 318, 320, 322, 324, 326, 328, 330, 332, 334, 336, which correspond to the ports 1-N described above with respect to Figures 1 and 2.
  • access may be gained to the storage medium of the storage device.
  • group A 220 of the storage medium 200 in Figure 4 may be displayed, or accessed, through one or more ports contained within the access panel 300. Which group is seen through which port or ports may be a dynamic definition, that changes as data contained within the storage device changes.
  • Figure 6 is a block diagram which depicts an exemplary set of ports associated with fiber switch unit 34. Although sixteen ports are illustrated by the dots set forth therein, those skilled in the art will appreciate that any number of ports may be built into fiber switch 34. Three of the ports 400, 402 and 404, in the upper left-hand portion of Figure 6, are connected to the central storage device 30. Two of the ports are connected to switches 36 and 40 in switching matrix 32. Some or all of the remaining ports are connected to host devices (not shown in this Figure). These connections may be accomplished by way of any communication channel, such as direct cable wiring, infrared line-of-sight connections, or the like.
  • the switch ports shown in Figure 6 may be grouped into zones, which, by way of wiring or other connections, provide common access to the ports within each zone.
  • the six ports in zone 1 share common access to the portions of the media within central storage device 30 that are made available by device 30' s internal controller (not shown) to the adaptor port to which port 400 is connected.
  • those ports zoned in zones 2 and zone 3 share common access with respective portions of the media of central control device 30 to which ports 402 and 404 are connected via adaptor ports, respectively.
  • ports can belong to more than one zone, e.g., port 408, such that those ports can transact data through multiple adaptor ports on the central storage device 30.
  • Switch port zoning provides the mechanism whereby a plurality of hosts can be aggregated into the three data pipes available to switch 34 for connection with the central storage device 30.
  • this feature also adds to the complexity associated with allocating storage on the central storage device and defining the pathway thereto. Accordingly, exemplary embodiments of the present invention provide techniques for handling these issues.
  • provisioning of central storage devices like those described above comprises three general steps which are depicted by the flow diagram of Figure 7.
  • the controller (not shown) in the central storage device 30 needs to be informed that a particular host is authorized to access a particular range of storage within the central storage device.
  • a path is created from the host through the switching matrix to the allocated storage at step 702. Note that steps 700 and 702 can be performed in the reverse of the order in which they are described herein.
  • the host is informed of how to find the allocated storage via the path that has been created at step 704.
  • One solution to this problem is to automate the provisioning of the storage medium within the centralized storage device, and automate the assignment of ports to specific devices belonging to various client or customer groups.
  • devices such as the ports of the switch in Figure 6, and adaptor ports of the storage device, such as the adaptor ports of Figure 5, it is extremely difficult to assign ports, and provision the storage device with any degree of reliability.
  • This data model 800 is illustrated in the form of an entity relationship diagram, which identifies the relationships between various entities associated with the storage device, which entities may be physical or virtual.
  • Each entity is represented by a rectangular block. These rectangular blocks are labeled to indicate their significance, and are connected by way of connection lines. Each line has a single connection on one end and multiple connections on the opposite end, indicating a one-to-many relationship, i.e., or indicating that a single entity is related to multiple, entities on the end of the line with multiple connections.
  • each device 802 e.g., a computer
  • Each instance of the entities illustrated in Figure 8 can be characterized by the values of a set of variables that interrelate the entities multiple variables to aid in relaying information to one another.
  • Each variable set also includes a primary key, denoted in the following tables by an asterisk following the variable name.
  • Each entities' primary key provides a link between the variable sets of related entities, i.e., each entity which has many-to-one relationship with another entity will include, in its variable set, the primary key of the other entity.
  • the Switch Zone Configuration entity 832 has a many-to-one relationship with both the Switch Zones entity 830 and the Storage Device Switch Configuration entity 834.
  • the variable set associated with the Switch Zone Configuration entity 832 includes the primary keys from the variable sets of both entities 830 and 834.
  • the exemplary entity variables described below are not intended to be a comprehensive list, but merely to indicate some of the variables which can be used in relating information contained within the storage device to various ports, and zones contained on the switching unit. It will be recognized by those skilled in the art, however, that additional variables, or in some cases fewer variables, may be used to accomplish similar results. Also shown in the following charts are indications that the variables are "required” or “optional” for this exemplary embodiment. Those skilled in the art will also appreciate that these designations may vary.
  • device entities 802 are each connected to (or associated with) multiple storage device unit entities 804. These device entities 802 may relate to, for example, computers, computer peripherals, and other computer networking equipment. As mentioned above, the device entities 802 illustrated in Figure 8 are characterized by a variable set, which is defined to aid the other entities connected thereto in conveying information to, and extracting information from, this entity. These variables are indicated below in Table 1.
  • Table 1 VARIABLE SET OF THE DEVICE ENTITIES 802
  • the primary key for each of the device entities 802 is the device identification variable, or DVC_ID, which uniquely identifies each central storage device.
  • DVC_ID device identification variable
  • the device type and status indicate the particular type of device, and its operating status to the network.
  • the effective beginning date indicates to the network the initial date for this record in the database, and the discovery date indicates the date at which the device is first registered with the system.
  • other variables can also be provided.
  • the machine ID variable may reflect an identifier which is associated with the device for fast, remote identification. More information regarding these machine ID values may be found in U.S. Patent Application Serial
  • NETWORK SERVERS filed on August 4, 2000, to Joshua T. LeVasseur et al., the disclosure of which is incorporated here by reference.
  • information such as serial numbers, device manufacturers, chassis serial numbers, system information, operating system version information, and mail protocol capability may all be used to characterize devices by way of variables contained in Table 1.
  • the system information and operating system version information can be that which would be reported from the device in response to an SNMP inquiry.
  • a device description variable may be used to indicate any additional information that is needed to be known by the network, and a monitoring flag may indicate any problems the device is having to the network.
  • These storage device unit entities 804 may correspond to, for example, RAID based storage devices such as EMC storage devices, or other suitable storage device units. As seen in Figure 8, the data model depicts a many- to-one relationship between the entities 804 and 802 and a one-to-many relationship between entity 804 and entities 806 and 814.
  • the storage device unit entities 804 are characterized by their own variable set, which values are illustrated below in Table 2.
  • Table 2 are variables which define the device identification number, the storage device site identification, the storage device name, and the maximum number of drivers and adaptors that can be used with the present invention.
  • each storage device unit entity 804 relates to multiple storage device storage block entities 806, which entities may, for example, correspond to the storage blocks illustrated in Figure 2, for example.
  • the storage device storage block entities 806 used in the present invention are characterized a set of variables that are illustrated below in Table 3.
  • Table 3 above illustrates the set of variables used by the storage device storage block entities 806.
  • Each storage block has its own identification number, and retains the group volume identification number of the group containing the block. Additionally, a device identification number is stored, and information such as the hexadecimal memory address of the block, the block status, the block storage size, and the group order of the block. This latter variable identifies the order in which the blocks are stored within the group and can be used to determine the begin and end addresses for each volume.
  • Each of these variables is designated, in this exemplary embodiment, as required for the storage device storage block entities 806.
  • multiple storage block entities 806 make up various storage device storage group entities 808.
  • the storage device storage group entities 808 each utilize multiple variables, which are each required for the operation of these groups. These variables are indicated below in Table 4.
  • the variables associated with the storage device storage group entities 808, shown above in table 4, include a group identification number to identify a particular memory group.
  • beginning and ending addresses for the memory group are provided.
  • the target and LUN variables are part of the characterizing string which identifies a mounted disk in the central storage device 30, e.g., dev/dsk/c#t#d#s#, where c refers to the controller number, t# refers to the target number (stored as a variable in this entity), d# refers to the disk number (referred to herein as the logical unit number or LUN) and s# refers to the slice number.
  • variables define the function type and protection type of the group, along with the status of the group.
  • the storage device group entities 808 may be connected to one or more adaptor group port entities 810, one or more of which may be connected to various storage device adaptor port entities 812.
  • the variable set associated with the adaptor group port entities 810 are shown below in Table 8.
  • Table 5 contains the variables of the adaptor group port entities 810. These variables include the adaptor port identification number and the group identification number associated with each adaptor port identified by the adaptor port identification number. Each of these variables is required, according to this exemplary embodiment of the present invention, for the characterization of the adaptor group port entities 810.
  • the storage device adaptor port entities 812 may be associated with storage device adaptor ports, such as ports 314-316 illustrated in Figure 3.
  • the variable set of the storage device adaptor port entities 812 are outlined in Table 6, shown below.
  • Table 6 indicates the set of variables associated with the storage device adaptor port entities 812. These variables include an adaptor port identification number, a global name identification number, an adaptor slot identification.
  • the adaptor identification may correspond to the physical and adaptor upon which the adaptor port resides. Additionally, an operating system indication may optionally be made which indicates the operating system used by the adaptor port.
  • one or more storage device adaptor port entities 812 may be associated with physical device adaptor port situated on each storage device adaptor associated with the multiple storage device adaptor entities 814 related to each storage device unit entity 804.
  • the set of variables associated with the storage device adaptors are shown below in Table 7.
  • an adaptor identification number is indicated along with a device identification number.
  • the type of adaptor along with the physical slot in which it is inserted are contained within variables in this table as well. Additionally, an indication as to the size of the adaptor cache may be made.
  • Storage device global name entities 816 are assigned to one or more storage device adaptor port entities 812 and one or more interface component entities 818. These storage device global name entities 816 serve as global variables, which globally define identification names for each of the storage device adaptor port entities 812, interface component entities 818, storage device VCM entities 822, and switching unit entities 824.
  • the storage device global name entities 816 is a key aspect to the present invention. As can be discerned from Figure 8, storage device global name entities 816 is a central to the entity relationships described therein and are used to relate devices to the storage device units by way of their adaptor ports, and to switching units. This is accomplished by assigning specific, unique global names to each of the entities with which it relates, and with which it comes in contact.
  • variable set associated with the storage device global name entities 816 include a worldwide, or global, identification number, and a worldwide, or global, name. Additionally, the storage device global name entities 816 provide a variable that indicates the type of device for each device associated with each worldwide, or global, name.
  • the variables set of the storage device global name entities 816 are shown below in Table 8.
  • One or more interface component entities 818 is associated with each interface card entity 820.
  • the variable set associated with the interface component entities 818 is set forth below in Table 9.
  • the variable in Table 9, associated with the interface component entities 818 include an interface component identification number, and an interface card identification number. Additionally, the type of interface is indicated, as is the effective beginning date of the interface component. The effective beginning date corresponds to the first date upon which the device is available to the network. Additional variables of the interface component entities 818 may optionally be provided where necessary to accomplish desired functionality. For example, optional variables corresponding to a hardware address, the speed of the network interface, an effective ending date, an interface alias, an interface descriptor, an interface index, and a worldwide, or global, name identification number may all be provided for as variables of the interface component entities 818.
  • the variables IF_ALIAS and IF INDEX are switch related variables that are read using SNMP.
  • IF_ INDEX stores a value that can be supplied to a switch to access a particular port on that switch.
  • IF_ALIAS stores a MAC address of the device that is plugged in to the other side of the switch.
  • IF_DESCRIPTOR provides a description of the interface which may be set by the user to aid in understanding the relationship of this record to the physical implementation, e.g., this record relates to a particular VLAN.
  • Interface card entities 820 are associated with the device entities 802.
  • the variables associated with the interface card entities 820 are set forth below in Table 10.
  • the interface card variables set forth in Table 10 include an interface identification number for unique identification of each interface card. Additionally, a device identification number, and an effective beginning date are provided by these variables. Additional optional variables may be provided, which are associated with the interface card entities 820, depending upon the functionality desired from the interface card entities 820. For example, variables associated with the interface cards may optionally indicate the interface card index number, serial number, or slot in which it is inserted. Additionally the speed of the central processing unit (CPU), and interface card memory quantity may be indicated. Additionally, the amount of dynamic random access memory (DRAM) may be indicated in megabytes (MBs).
  • DRAM dynamic random access memory
  • the storage device global names are also assigned to one or more storage device VCM entities 822.
  • the VCM (also sometimes referred to as the EMC/ECC) is the controller which provides access to the disk devices within the central storage device, e.g., as mentioned above with respect to step 700.
  • One or more storage device VCM entities 822 may be associated with each storage device storage group entity 808. The variables associated with the storage device VCM entities 822 are indicated below in Table 11.
  • Storage device global name entities 816 are assigned to one or more switching unit entities 824, which may be related to physical switching units, such as the switch of Figure 6, for example.
  • Table 12 below indicates variable associated with the switching unit entities 824.
  • Each of the variables indicated in Table 12 are important, and associated with the switching unit 824. These variables include a switch identification number, which identifies a unique switching unit. Also, a worldwide, or global, name identification number is given to provide a unique name of the device on the network. A domain number, host name, internet protocol (IP) address, and operating system (OS) conversion are all provided as variables of the switching unit entities 824.
  • IP internet protocol
  • OS operating system
  • Each switching unit entity 824 is associated with one or more switch port entities 826.
  • These switch port entities 826 may be configured in a manner similar to that illustrated in Figure 6, or in another suitable manner.
  • the variable set associated with the switch port entities 826 are indicated below in Table 13.
  • Each of the switch port entities 826 is identified by a switch port identification number. Additionally, the switch port entities 826 utilize variables that indicate the switch identification number with which the switch port entities 826 are associated, a worldwide name identification number associated with the switch port, a slot identification, and a switch port status indication.
  • each of the switch port entities 826 may be associated with parts of one or more zones of a physical switching unit. As previously discussed, these zones indicate which ports have access to the data of other ports contained within the same zones.
  • each of the switch port entities 826 is defined by one or more switch zone mapping entities 828. These switch zone mapping entities 828 relate to the various switch zone entities 830 in a many-to-one relationship. The variables associated with the switch zone mapping entities 828 are indicated below in Table 14.
  • the switch zone mapping entities 828 make use of a zone identification number, and a switch port identification number. These variables are used to map the zones of the switching unit, such as the switch illustrated in Figure 6.
  • the switch zone entities 830 relate to the switch zone mapping entity in a one-to-many relationship.
  • the variables associated with the switch zone entities 830 include a zone identification number, and a zone name, which help to further define the zones of the switching unit, such as those zones illustrated on the switch of Figure 6. These variables are shown below in Table 18.
  • the switch zone entities 830 are related to one or more switch zone configuration entities 832, which define the configuration of each zone of the switching unit which may change.
  • the variables associated with the switch zone configuration entities 832 are illustrated below in Table 16.
  • switch zone configuration entities 832 variables associated with the switch zone configuration entities 832 are shown.
  • a switch zone configuration identification number is given, which uniquely identifies each switch configuration.
  • a zone identification number which uniquely identifies each zone related to the switch zone configuration entities 832.
  • the storage device switch configuration entities 834 may be defined by one or more switch zone configuration entities 832.
  • the variable associated with the storage device switch configuration entities 834 are illustrated below in Table 17.
  • variables associated with the storage device switch configuration entities 832 are illustrated. Particularly, a switch configuration identification number is given along with a switch identification number and the name of the configuration being referred to. Additionally, a status variable indicates the configuration status.
  • the present invention involves a complex network which relates the various entities to each other. The purpose for which these variables are given is to provide a convenient environment for provisioning each of the storage device units utilized in the network with various devices by way of switching units.
  • the storage device global names are especially useful for uniquely naming and identifying each of the entities shown in the entity relationship diagram of Figure 8.
  • the network, and storage device units may be readily reconfigured for use with different zones, depending upon customer or client demand.
  • new devices having new interface cards and new interface components may be added to the network, and may gain access only to those groups contained within the storage device media to which access is desired to be granted.
  • a system administrator may accomplish this by way of using zones to which the various ports of switching units are assigned. These zones provide a convenient mechanism whereby system administrators may readily determine which ports are able to see specific storage information contained within specific blocks or groups on the storage media of the storage device unit.
  • Step 700 Informing the Central Data Storage Device of the Host's Authorization
  • each central storage device 30 will typically include a volume on which is stored a database that masks the data based on the authorization granted to each host device. This database can be edited for the purpose of granting new rights, e.g. , those described in table above, as part of the provisioning process.
  • this editing process can be performed using a utility provided by the EMC Corporation known as fpath.
  • the fpath utility includes functions for backing up the database, adding a new host device, changing the name of a host, listing the contents of the database and refreshing the configuration of the database which are used in this process. These functions can be used as described in the flow diagram of to inform the central storage device 30 of the new host's access authorization.
  • the database is backed up to preserve the original records.
  • an entry is added to the data base using the appropriate fpath command for adding a device.
  • This command will include, as arguments, a unique identifier of the host (sometimes referred to as a worldwide name (wwn)), the port on the central storage device through which this host will obtain access and the range of storage (e.g., as in the table) to which this host will be permitted access.
  • a unique identifier of the host sometimes referred to as a worldwide name (wwn)
  • the range of storage e.g., as in the table
  • an alias of the entry created in step 902 may be added to the database at step 904 to make it easier to visually locate the entry of step 902 in the database.
  • the entries are verified using the fpath listing function (step 906) and the database can be saved (step 908) to complete the task of informing the central storage device 30 of the new host's access authorization.
  • Step 702 Creating a Path Through the Switching Matrix by Zoning
  • the next step in the provisioning process of Figure 7 is to establish the zones on the switch which aggregate multiple hosts ports on the switch to a single central storage device port on the switch, as depicted conceptually in Figure 6.
  • step 1000 it is first determined the switch port to which the host of interest is connected. This can be accomplished by establishing a Telnet session with the switch using access terminal 43 which can establish a communication link with a processor (not shown) embedded in the switch. Those skilled in the art will be aware of various Telnet emulation programs which can be used to establish these types of communication sessions. Once the Telnet session is established, the port connections of the switch can be revealed using a "Switchshow" command, which command might, for example, reveal:
  • zoning requires the addition of port 11 to the zone which includes port 15. This is performed in step 1002 by executing a "zoneCreate" command via the Telnet session with the switch as:
  • zoneCreate "AAACORP_m0042_fcaO_sym2621_falA", "1,15;1,11"
  • the zone is added to a selected configuration file to be stored on the switch, again using a suitable Telnet command.
  • Each fiber switch 34, 36, 38, and 40 may have multiple configuration files which accumulate the matchings between host ports and central storage device ports on the switch.
  • the new configuration is enabled via the terminal 43 which performs compilation and verification of the new configuration. If errors are detected by the switch processor, then the switch will indicate the error to the terminal 43 and the configuration will not be modified. Otherwise, the new configuration is saved and the zoning step is complete.
  • Step 704 Informing the Host How to Find the Allocated Storage
  • the next step is to configure the host so that it can access the specified storage space. This is accomplished by modifying a configure file (sd.conf) stored in the host and rebooting the host.
  • the syntax for this command is:
  • ⁇ fa-wwn> is the symm's fa wwn
  • a process referred to as persistent binding can then be applied to tie devices to a specific wwn jn the central storage device.
  • exemplary embodiments of the present invention automate much of the process using the data model of Figure 8. More specifically, exemplary embodiments request certain information from a user who wishes to provision the central storage, extract information from the data model of Figure 8 (stored as a number of database tables) and process that information to output the configuration commands which can be used to provision the central storage device.
  • a general method for automating the provisioning of a central storage device using a data model according to an exemplary embodiment of the present invention will now be described with reference to the flow diagram of Figure 11 and the exemplary software code attached as Appendix A.
  • the user inputs one or more identifiers associated with the host(s) to which he or she wants to allocate storage from the central storage device 32.
  • the software code set forth in Appendix A uses this information to first determine, and then present a list of free storage groups (also sometimes referred to as "metavolumes" that are in the same location as the hosts, e.g., using the information stored in the storage device group entities 808 and the information stored in the device entities 802 of the data model.
  • the user selects one (or more) groups for allocation at step 1104.
  • the software then continues to output the commands for configuring the central storage device, the host(s), and the switch(es) at step 1106, again by using information stored in the data model.
  • the switch zone creation and configuration is performed using information in the data model regarding the switches to which the selected host(s) is connected to using information stored in the interface card entities 820 and interface component entities 818 of the data model and the ports of the central storage device on which the selected group(s) is available (using information stored in entities 808, 810 and 818.
  • the present invention is directed to a system and method for automated provisioning of central data storage devices, which allows for multiple network clients to share a centralized storage device without sharing access to their information.
  • clients may desire to share their information with specified individuals or companies, and therefore may increase the number of parties having access to this information, according to the individual client's goals and desires.
  • #sw_nm spin.Device.getLocation( ⁇ "id” : switch ["dvcjd”] ⁇ )
  • cmds in symm_cmds.items() cmds.append( "fpath Isdb -s on” ) cmds.append( "[verify]” ) cmds.append( "fpath backupdb -o /root/vcm/$MYSYM. ' date +%d%m%YJ J /oH%M%S '

Abstract

L'invention concerne des techniques et des systèmes permettant de modéliser et d'approvisionner des dispositifs de stockage complexes. Un modèle de données permet d'établir des rapports entre entités entre des éléments physiques et des éléments logiques des dispositifs de stockage. Le modèle de données peut être utilisé pour approvisionner les dispositifs de manière uniforme, automatique et sécurisée.
PCT/US2001/042859 2000-10-31 2001-10-31 Modele de donnes pouvant etre utilise pour l'approvisionnement automatique de dispositifs de stockage de donnees centralises WO2002037212A2 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2002214679A AU2002214679A1 (en) 2000-10-31 2001-10-31 A data model for use in the automatic provisioning of central data storage devices

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US69935100A 2000-10-31 2000-10-31
US09/699,347 US6751702B1 (en) 2000-10-31 2000-10-31 Method for automated provisioning of central data storage devices using a data model
US09/699,347 2000-10-31
US09/699,351 2000-10-31

Publications (2)

Publication Number Publication Date
WO2002037212A2 true WO2002037212A2 (fr) 2002-05-10
WO2002037212A3 WO2002037212A3 (fr) 2003-07-24

Family

ID=27106399

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2001/042859 WO2002037212A2 (fr) 2000-10-31 2001-10-31 Modele de donnes pouvant etre utilise pour l'approvisionnement automatique de dispositifs de stockage de donnees centralises

Country Status (2)

Country Link
AU (1) AU2002214679A1 (fr)
WO (1) WO2002037212A2 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9100283B2 (en) 2002-06-12 2015-08-04 Bladelogic, Inc. Method and system for simplifying distributed server management

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5325505A (en) * 1991-09-04 1994-06-28 Storage Technology Corporation Intelligent storage manager for data storage apparatus having simulation capability
US5835758A (en) * 1995-02-28 1998-11-10 Vidya Technologies, Inc. Method and system for respresenting and processing physical and conceptual entities
WO1998053399A2 (fr) * 1997-05-21 1998-11-26 British Telecommunications Public Limited Company Analyse operationnelle pour systemes commandes par logiciels
WO2000029954A1 (fr) * 1998-11-14 2000-05-25 Mti Technology Corporation Mise en correspondance d'une unite logique dans un environnement de reseau a zone de memoire (san)

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5325505A (en) * 1991-09-04 1994-06-28 Storage Technology Corporation Intelligent storage manager for data storage apparatus having simulation capability
US5835758A (en) * 1995-02-28 1998-11-10 Vidya Technologies, Inc. Method and system for respresenting and processing physical and conceptual entities
WO1998053399A2 (fr) * 1997-05-21 1998-11-26 British Telecommunications Public Limited Company Analyse operationnelle pour systemes commandes par logiciels
WO2000029954A1 (fr) * 1998-11-14 2000-05-25 Mti Technology Corporation Mise en correspondance d'une unite logique dans un environnement de reseau a zone de memoire (san)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9100283B2 (en) 2002-06-12 2015-08-04 Bladelogic, Inc. Method and system for simplifying distributed server management
US9794110B2 (en) 2002-06-12 2017-10-17 Bladlogic, Inc. Method and system for simplifying distributed server management
US10659286B2 (en) 2002-06-12 2020-05-19 Bladelogic, Inc. Method and system for simplifying distributed server management

Also Published As

Publication number Publication date
WO2002037212A3 (fr) 2003-07-24
AU2002214679A1 (en) 2002-05-15

Similar Documents

Publication Publication Date Title
US6751702B1 (en) Method for automated provisioning of central data storage devices using a data model
US7734712B1 (en) Method and system for identifying storage devices
JP6208207B2 (ja) オブジェクト・ストレージ・システムにアクセスするコンピュータ・システム
US10140045B2 (en) Control device for storage system capable of acting as a constituent element of virtualization storage system
US6839824B2 (en) System and method for partitioning a storage area network associated data library employing element addresses
JP5963864B2 (ja) 入力/出力オペレーションのためにオブジェクト・ストレージ・システムを構成すること
US8516191B2 (en) Storage system and method of managing a storage system using a management apparatus
US6845431B2 (en) System and method for intermediating communication with a moveable media library utilizing a plurality of partitions
US20030126225A1 (en) System and method for peripheral device virtual functionality overlay
US7237083B2 (en) Storage control system
US6496914B1 (en) Method and system for administering storage devices on a network
US20020029319A1 (en) Logical unit mapping in a storage area network (SAN) environment
US20100250878A1 (en) Storage System of Storing Golden Image of a Server or Any Physical/Virtual Machine Execution Environment
US20070079060A1 (en) Method, apparatus and program storage device for providing virtual disk service (VDS) hints based storage
US8898418B2 (en) Method, apparatus and computer program for provisioning a storage volume to a virtual server
US8271759B2 (en) Storage system
WO2014162497A1 (fr) Système informatique intégré et son procédé de commande
US10852980B1 (en) Data migration techniques
WO2002037212A2 (fr) Modele de donnes pouvant etre utilise pour l'approvisionnement automatique de dispositifs de stockage de donnees centralises
US20120226669A1 (en) Merging a storage cluster into another storage cluster
US11269792B2 (en) Dynamic bandwidth management on a storage system
US11922043B2 (en) Data migration between storage systems
WO2002037282A2 (fr) Procede pour alimenter des dispositifs de stockage complexes en donnees

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase in:

Ref country code: JP