US20120016950A1 - Method and apparatus for dynamically managing bandwidth for clients in a storage area network - Google Patents
Method and apparatus for dynamically managing bandwidth for clients in a storage area network Download PDFInfo
- Publication number
- US20120016950A1 US20120016950A1 US13/246,505 US201113246505A US2012016950A1 US 20120016950 A1 US20120016950 A1 US 20120016950A1 US 201113246505 A US201113246505 A US 201113246505A US 2012016950 A1 US2012016950 A1 US 2012016950A1
- Authority
- US
- United States
- Prior art keywords
- area network
- storage area
- client devices
- network according
- data storage
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
- H04L67/1097—Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/60—Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources
- H04L67/61—Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources taking into account QoS or priority requirements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L9/00—Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
- H04L9/40—Network security protocols
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L69/00—Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
- H04L69/30—Definitions, standards or architectural aspects of layered protocol stacks
- H04L69/32—Architecture of open systems interconnection [OSI] 7-layer type protocol stacks, e.g. the interfaces between the data link level and the physical level
- H04L69/322—Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions
- H04L69/329—Intralayer communication protocols among peer entities or protocol data unit [PDU] definitions in the application layer [OSI layer 7]
Definitions
- the present invention relates to the field of data storage. More particularly, the present invention relates to dynamically restricting the bandwidth of one I/O path in a storage area network to provide additional bandwidth to another I/O path.
- the primary data storage device is usually a hard drive with a storage capacity measured in gigabytes.
- computers may store data using such devices as CD-ROM drives, floppy disk drives, tape drive, etc.
- the computers of the network may also store data on network servers or other data storage devices, such as those mentioned above, that are accessible through the network. For larger systems with even greater data storage needs, arrays of data storage disks may be added to the network.
- SANs Storage Area Networks
- a SAN is essentially a high-speed network between client devices, such as servers and data storage devices, particularly disk arrays.
- client devices such as servers and data storage devices, particularly disk arrays.
- a SAN overcomes the limitations and inflexibility of traditional attached data storage.
- a SAN can overcome limitations of traditional attached data storage but also introduces new considerations.
- SANs experience competition for resources when more than one server is attempting to access the same data storage device.
- a typical storage device has a limited amount of bandwidth in its Input/Output (I/O) paths, and this bandwidth must be shared by the clients accessing the storage device.
- I/O Input/Output
- the present invention provides a method for managing bandwidth allocation in a storage area network that receiving a plurality of Input/Output (I/O) requests from a plurality of client devices, determining a priority of each of the client devices relative to other client devices, and dynamically allocating bandwidth resources to each client device based on the priority assigned to that client device.
- I/O Input/Output
- FIG. 1 is a block diagram illustrating an embodiment of a system according to principles of the present invention.
- FIG. 2 is a block diagram illustrating an additional embodiment of a system according to principles of the present invention.
- FIG. 3 is a block diagram illustrating an additional embodiment of a system according to principles of the present invention.
- FIG. 4 is a block diagram of a system according to one embodiment of the present invention.
- FIG. 5 is a flow diagram for assigning array performance groups according to principles of one embodiment of the present invention.
- Embodiments of the invention include a method for managing bandwidth associated with a Storage Area Network (SAN).
- SAN Storage Area Network
- an innovative method limits the bandwidth associated with one I/O path so that a different I/O path may consume the extra bandwidth.
- Storage area networks vary in size and complexity, and are flexible in their configurations for meeting the storage needs of a network.
- a simplified storage area network configuration is depicted in FIG. 1 to illustrate the transfer of data between a limited number of devices interfaced with a storage area network. More complex storage area networks may interface with any number of devices as needed to meet a given user's storage needs.
- FIG. 1 illustrates a data retrieval system according to one embodiment of the present invention.
- an embodiment of a data retrieval system includes a number of servers or host computers ( 100 , 110 ), referred to collectively as “clients.”
- data retrieval systems may assign a different priority to each client within the data retrieval system, as illustrated by the priority server ( 100 ) and the non-priority server ( 110 ).
- Each server is communicatively coupled to a Host Bus Adapter (HBA) ( 115 ) which is in turn coupled to a communication line ( 120 ).
- HBA Host Bus Adapter
- the communication line ( 120 ) that couples the servers ( 100 , 110 ) to the storage disk array ( 150 ) is preferably a fibre channel loop compliant with the “Fibre Channel Physical and Signaling Interface” ((FC-PH) Rev. 4.3, X3T11, Jun. 1, 1994 standard, American National Standards for Information Systems), which standard is hereby incorporated by reference.
- Each device on the loop ( 120 ) by virtue of the fiber channel host bus adapter, has a unique identifier referred to as its world wide name (WWN).
- WWN world wide name
- the present invention may also use any unique identifier associated with the servers ( 100 , 110 ) so long as that identifying means is unique for each device among the interconnected devices.
- the line ( 120 ) is fed into a fibre channel switch ( 130 ).
- the switch ( 130 ) continues on to a port ( 140 ) of the storage disk array ( 150 ).
- storage disk arrays ( 150 ) divide the storage into a number of logical volumes. These volumes are accessed through a logical unit number (LUN) ( 155 ) addressing scheme as is common in SCSI protocol based storage systems, including SCSI protocol based, fibre channel loop, physical layer configurations.
- LUN refers to a logical unit or logical volume, or, in the context of a SCSI protocol based device or system, to an SCSI logical unit or SCSI logical volume.
- the number of physical disk drives may be the same as, or different from, the number of logical drives or logical volumes. However, for the sake of simplicity and clarity, we use these terms interchangeably here, focusing primarily on logical volumes as compared to the physical disk drives that make up those logical volumes.
- the storage disk array ( 150 ) also contains a resource manager ( 160 ).
- the resource manager ( 160 ) contains firmware that enables the resource manager ( 160 ) to identify each server ( 100 , 110 ) accessing the storage array ( 150 ) and to allot I/O bandwidth at the port ( 140 ) to each such server ( 100 , 110 ) as specified by the firmware.
- FIG. 2 illustrates an additional configuration of one embodiment of the present invention.
- a number of servers ( 200 , 210 ) may be connected via a fibre channel loop ( 220 ) to a plurality of FC switches ( 230 ) leading to a plurality of storage disk arrays ( 250 ) that are communicatively coupled to the network through the switches ( 230 ).
- FC switches 230
- storage disk arrays 250
- FIG. 2 illustrates an additional configuration of one embodiment of the present invention.
- An I/O path is the path from the client's Host Bus Adapter (HBA), over a Storage Network, to a block of storage on a storage device (e.g. a disk array) ( 250 ).
- HBA Host Bus Adapter
- a storage device e.g. a disk array
- the system recognizes that the host systems are not all of the same priority, i.e., in order to optimize the operation of the overall system, some clients or servers need more I/O performance and bandwidth from the storage devices ( 250 ) than do other clients.
- the maximum storage device performance available for lower priority client systems shouldn't impact the storage device performance available to higher priority clients.
- the resource manager ( 260 ) is a product that monitors the I/O performance and bandwidth usage of the storage system, and sets performance caps based on user-established policies.
- One aspect of the present invention concerns the ability to set an upper limit or cap and a minimum threshold on bandwidth usage.
- the cap limits the amount of bandwidth a client may use at any one time.
- the minimum threshold establishes a minimum level of performance below which the user-defined policies, i.e., the caps, are relaxed.
- There are a number of ways to administer caps and thresholds including, but not limited to, assigning a cap and/or threshold to each port, to each client/array port pair, or to each client/array LUN pair.
- FIG. 3 illustrates an embodiment of the present invention that bases the threshold and cap on a port.
- contention may occur.
- the path between each client and the storage disk array ( 350 ) is considered a separate I/O path here.
- the other I/O path can be allowed to use the extra I/O on the Port.
- utilization of the I/O path is optimized. This concept can be expanded out to a very large Storage Network, but can get difficult to manage.
- Groups By grouping servers into priority categories (Groups), a single setting can be made to all servers in a category automatically.
- the servers ( 300 , 310 ) have been grouped into priority groups: priority servers ( 300 ) and ordinary servers ( 310 ). Each group has a number of HBAs ( 315 ) connecting the respective groups to the fibre channel loop ( 320 ) leading to a switch ( 330 ). Each switch ( 330 ) leads to a port ( 340 , 345 ) for each server group. Port 1 ( 340 ) is dedicated to the group of ordinary servers ( 310 ) and port 2 ( 345 ) is dedicated to the priority servers ( 300 ).
- the resource manager ( 360 ) can allocate bandwidth resources to each port in proportion to the importance of the corresponding client group. By assigning a cap and a threshold quantity to each port, the bandwidth can be efficiently distributed.
- logic unit ( 355 ) of the storage disk array ( 350 ) can only handle 7,000 input and output operations per second (IOPS)
- port 1 ( 340 ) may be capped at 2,000 IOPS.
- all the servers attached to port 2 ( 345 ) can access the remaining 5,000 IOPS associated with the logic unit ( 355 ). Accordingly, port 2 may be capped at 5,000 IOPS.
- the bandwidth resources can be dynamically managed. If, by way of example only, port 1 ( 340 ) had a threshold of 2,000 IOPS and activity at port 1 ( 340 ) drops below that threshold, the cap assigned to port 2 ( 345 ) is subsequently released allowing the servers ( 300 ) associated with port 2 ( 345 ) to make use of the unused bandwidth.
- the embodiment demonstrated in FIG. 3 may also be implemented using a single high priority server and a single regular priority server.
- a single port ( 340 , 345 ) is associated with each server.
- a cap may be implemented on one or both servers through the corresponding port ( 340 , 345 ) with the residual bandwidth available to the server associated with the other port.
- a threshold may be implemented on either or both of the ports ( 340 , 345 ). If the activity at one of the ports ( 340 , 345 ) dropsdrops below the threshold assigned it, the assigned cap at the other port is subsequently released to allow the second server make use of the unused bandwidth.
- FIG. 4 illustrates how a cap and threshold may be based upon a client/array port pair.
- clients with different priorities 400 , 410
- the resource manager ( 460 ) can identify which clients ( 400 , 410 ) are high priority clients ( 400 ) and which ones are not ( 410 ).
- the resource manager ( 460 ) can place caps and/or thresholds at each port for specific clients.
- FIG. 4 illustrates the client/array port pair embodiment of the present invention. If port 1 ( 440 ) is “capped” at 500 IOPS for HBAs 1 , 3 , and 5 only ( 415 ), and Port 1 ( 440 ) has threshold of 2,000 IOPS then if total activity on Port 1 ( 440 ) drops below 2,000 IOPs, the caps for HBAs 1 , 3 , and 5 ( 415 ) arc released.
- Port 2 ( 445 ) is “capped” at 500 IOPS for HBAs 1 , 3 , and 5 ( 415 ) only and Port 2 ( 445 ) has threshold of 3,000 IOPS, then if total activity on Port 2 ( 445 ) drops below 3,000 IOPs, the caps for HBAs 1 , 3 , and 5 ( 415 ) are released.
- the client/array LUN pair embodiment of the present invention uses the resource manager ( 460 ) to identify the client ( 400 , 410 ) requesting bandwidth performance and applying corresponding caps and thresholds to the individual logic unit ( 455 ) in the storage disk array ( 450 ) rather than the ports ( 440 , 445 ).
- the dynamic management of the bandwidth resources is still triggered by a drop in activity below the threshold. However, in this embodiment, the activity is measured at the individual logic unit ( 455 ).
- This invention allows for a fine level of user control over the threshold settings.
- the array can relax performance caps when they are not needed and, thereby, not unduly restrict the bandwidth available to the capped port, client/array port pairs, or host/LUN pairs.
- the bandwidth resources When all of the bandwidth resources arc being used, they are distributed according to priority designations. If, however, activity drops below threshold values, caps for each group may be released to allow for a dynamic redistribution of the available resources.
- the settings can be made in either I/O per second or MB per second.
- the present invention allows the caps and threshold to be time dependant corresponding to predicted spurts of activities.
- the priority of a client may be determined by its application performance requirements.
- a user application can interrogate various storage arrays and add its connectivity port to an array performance group which meets the application's transaction bandwidth.
- this embodiment allows for the dynamic grouping of host applications by acquiring information on available performance groups within the storage array.
- the user application evaluates an established group ( 610 ). The user application first determines if the performance cap of the first established group is greater than or equal to the application's bandwidth requirement ( 620 ). If it is, there is sufficient room in the established group to include the desired application. In that case, the port associated with the application is added to the first established group ( 630 ). If, however, the performance cap of the first established group is less than the application bandwidth requirement, the user application determines if there are additional groups ( 640 ).
- the application again evaluates the cap ( 620 ) to see if there is sufficient bandwidth to perform the desired application. If all of the groups have been considered and none meet the application bandwidth requirements, the application is added to the group with the highest cap ( 650 ).
- the embodiment of the present invention disclosed above allows for dynamic grouping of server priorities.
- the groups may be arranged so as to utilize the maximum bandwidth available.
- a user application can also dynamically add connectivity bandwidth and increase the performance capability of a host application by controlling multiple connectivity paths between the host computer and the storage array.
- GUI Graphical User Interface
- the resource manager can then increase available bandwidth accordingly by dedicating additional ports to the host application.
- This embodiment of the present invention can then be tied into billing applications for demand-based performance such as pay for performance applications.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Computer Security & Cryptography (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
- Computer And Data Communications (AREA)
Abstract
A method for managing bandwidth allocation in a storage area network includes receiving a plurality of Input/Output (I/O) requests from a plurality of client devices, determining a priority of each of the client devices relative to other client devices, and dynamically allocating bandwidth resources to each client device based on the priority assigned to that client device.
Description
- The present invention relates to the field of data storage. More particularly, the present invention relates to dynamically restricting the bandwidth of one I/O path in a storage area network to provide additional bandwidth to another I/O path.
- The use of computers and computer networks pervade virtually every business and other enterprise in the modem world. With computers, users generate vast quantities of data that can be stored for a variety of purposes. This storehouse of data can grow at a phenomenal pace and become critically valuable to those who have generated it. Consequently, there is an ever-present need for data storage systems that improve on capacity, speed, reliability, etc.
- In a single computer, the primary data storage device is usually a hard drive with a storage capacity measured in gigabytes. Additionally, computers may store data using such devices as CD-ROM drives, floppy disk drives, tape drive, etc. Within a computer network, the computers of the network may also store data on network servers or other data storage devices, such as those mentioned above, that are accessible through the network. For larger systems with even greater data storage needs, arrays of data storage disks may be added to the network.
- Storage Area Networks (SANs) are an emerging technology being implemented to accommodate high-capacity data storage devices, particularly disk arrays, within a network. A SAN is essentially a high-speed network between client devices, such as servers and data storage devices, particularly disk arrays. A SAN overcomes the limitations and inflexibility of traditional attached data storage.
- A SAN can overcome limitations of traditional attached data storage but also introduces new considerations. In particular, SANs experience competition for resources when more than one server is attempting to access the same data storage device. A typical storage device has a limited amount of bandwidth in its Input/Output (I/O) paths, and this bandwidth must be shared by the clients accessing the storage device.
- In one of many possible embodiments, the present invention provides a method for managing bandwidth allocation in a storage area network that receiving a plurality of Input/Output (I/O) requests from a plurality of client devices, determining a priority of each of the client devices relative to other client devices, and dynamically allocating bandwidth resources to each client device based on the priority assigned to that client device.
- The accompanying drawings illustrate various embodiments of the present invention and are a part of the specification. Together with the following description, the drawings demonstrate and explain the principles of the present invention. The illustrated embodiments arc examples of the present invention and do not limit the scope of the invention.
-
FIG. 1 is a block diagram illustrating an embodiment of a system according to principles of the present invention. -
FIG. 2 is a block diagram illustrating an additional embodiment of a system according to principles of the present invention. -
FIG. 3 is a block diagram illustrating an additional embodiment of a system according to principles of the present invention. -
FIG. 4 is a block diagram of a system according to one embodiment of the present invention. -
FIG. 5 is a flow diagram for assigning array performance groups according to principles of one embodiment of the present invention. - Throughout the drawings, identical reference numbers designate similar, but not necessarily identical, elements.
- Embodiments of the invention include a method for managing bandwidth associated with a Storage Area Network (SAN). According to one exemplary embodiment, described more fully below, an innovative method limits the bandwidth associated with one I/O path so that a different I/O path may consume the extra bandwidth. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the invention. It will be apparent, however, to one skilled in the art that the invention can be practiced without these specific details.
- Reference in the specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the invention. The several appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
- Storage area networks vary in size and complexity, and are flexible in their configurations for meeting the storage needs of a network. A simplified storage area network configuration is depicted in
FIG. 1 to illustrate the transfer of data between a limited number of devices interfaced with a storage area network. More complex storage area networks may interface with any number of devices as needed to meet a given user's storage needs. -
FIG. 1 illustrates a data retrieval system according to one embodiment of the present invention. As shown inFIG. 1 , an embodiment of a data retrieval system includes a number of servers or host computers (100, 110), referred to collectively as “clients.” As demonstrated inFIG. 1 , data retrieval systems may assign a different priority to each client within the data retrieval system, as illustrated by the priority server (100) and the non-priority server (110). Each server is communicatively coupled to a Host Bus Adapter (HBA) (115) which is in turn coupled to a communication line (120). - The communication line (120) that couples the servers (100, 110) to the storage disk array (150) is preferably a fibre channel loop compliant with the “Fibre Channel Physical and Signaling Interface” ((FC-PH) Rev. 4.3, X3T11, Jun. 1, 1994 standard, American National Standards for Information Systems), which standard is hereby incorporated by reference. Each device on the loop (120), by virtue of the fiber channel host bus adapter, has a unique identifier referred to as its world wide name (WWN). The present invention may also use any unique identifier associated with the servers (100, 110) so long as that identifying means is unique for each device among the interconnected devices.
- Continuing in the direction of the communication line (120), the line (120) is fed into a fibre channel switch (130). The switch (130) continues on to a port (140) of the storage disk array (150).
- In computing systems, storage disk arrays (150) divide the storage into a number of logical volumes. These volumes are accessed through a logical unit number (LUN) (155) addressing scheme as is common in SCSI protocol based storage systems, including SCSI protocol based, fibre channel loop, physical layer configurations. The term LUN refers to a logical unit or logical volume, or, in the context of a SCSI protocol based device or system, to an SCSI logical unit or SCSI logical volume.
- Those of ordinary skill in the art will appreciate that the number of physical disk drives may be the same as, or different from, the number of logical drives or logical volumes. However, for the sake of simplicity and clarity, we use these terms interchangeably here, focusing primarily on logical volumes as compared to the physical disk drives that make up those logical volumes.
- The storage disk array (150) also contains a resource manager (160). The resource manager (160) contains firmware that enables the resource manager (160) to identify each server (100, 110) accessing the storage array (150) and to allot I/O bandwidth at the port (140) to each such server (100, 110) as specified by the firmware.
-
FIG. 2 illustrates an additional configuration of one embodiment of the present invention. As shown inFIG. 2 , a number of servers (200, 210) may be connected via a fibre channel loop (220) to a plurality of FC switches (230) leading to a plurality of storage disk arrays (250) that are communicatively coupled to the network through the switches (230). It will be appreciated by those of ordinary skill in the art that the present invention may be practiced with a number of configurations without varying from the teachings of the present invention. - As mentioned earlier, SANs experience competition for resources when more than one client is attempting to access the same data storage device. A typical storage device has a limited amount of bandwidth in its I/O paths and this bandwidth should be properly apportioned out to the clients accessing the storage device. An I/O path is the path from the client's Host Bus Adapter (HBA), over a Storage Network, to a block of storage on a storage device (e.g. a disk array) (250). In order to properly allocate the bandwidth, the system recognizes that the host systems are not all of the same priority, i.e., in order to optimize the operation of the overall system, some clients or servers need more I/O performance and bandwidth from the storage devices (250) than do other clients. In order to maximize the performance of the system, the maximum storage device performance available for lower priority client systems shouldn't impact the storage device performance available to higher priority clients.
- The resource manager (260) is a product that monitors the I/O performance and bandwidth usage of the storage system, and sets performance caps based on user-established policies. One aspect of the present invention concerns the ability to set an upper limit or cap and a minimum threshold on bandwidth usage. The cap limits the amount of bandwidth a client may use at any one time. The minimum threshold establishes a minimum level of performance below which the user-defined policies, i.e., the caps, are relaxed. There are a number of ways to administer caps and thresholds, including, but not limited to, assigning a cap and/or threshold to each port, to each client/array port pair, or to each client/array LUN pair.
-
FIG. 3 illustrates an embodiment of the present invention that bases the threshold and cap on a port. When multiple I/O paths cross the same piece of hardware (e.g. an array port), contention may occur. The path between each client and the storage disk array (350) is considered a separate I/O path here. By allowing one I/O path (non-priority) to be throttled or capped above a certain performance level, the other I/O path (priority) can be allowed to use the extra I/O on the Port. Thus, utilization of the I/O path is optimized. This concept can be expanded out to a very large Storage Network, but can get difficult to manage. By grouping servers into priority categories (Groups), a single setting can be made to all servers in a category automatically. - In
FIG. 3 , the servers (300, 310) have been grouped into priority groups: priority servers (300) and ordinary servers (310). Each group has a number of HBAs (315) connecting the respective groups to the fibre channel loop (320) leading to a switch (330). Each switch (330) leads to a port (340, 345) for each server group. Port 1 (340) is dedicated to the group of ordinary servers (310) and port 2 (345) is dedicated to the priority servers (300). - By providing independent ports to each respective group of clients, the resource manager (360) can allocate bandwidth resources to each port in proportion to the importance of the corresponding client group. By assigning a cap and a threshold quantity to each port, the bandwidth can be efficiently distributed. By way of example only, if logic unit (355) of the storage disk array (350) can only handle 7,000 input and output operations per second (IOPS), port 1 (340) may be capped at 2,000 IOPS. By setting the cap at 2,000 IOPS for port 1 (340), all the servers attached to port 2 (345) can access the remaining 5,000 IOPS associated with the logic unit (355). Accordingly,
port 2 may be capped at 5,000 IOPS. By assigning a threshold equal to the aforementioned cap at port 1 (340) and port 2 (345), the bandwidth resources can be dynamically managed. If, by way of example only, port 1 (340) had a threshold of 2,000 IOPS and activity at port 1 (340) drops below that threshold, the cap assigned to port 2 (345) is subsequently released allowing the servers (300) associated with port 2 (345) to make use of the unused bandwidth. - The embodiment demonstrated in
FIG. 3 may also be implemented using a single high priority server and a single regular priority server. In this embodiment, a single port (340, 345) is associated with each server. A cap may be implemented on one or both servers through the corresponding port (340, 345) with the residual bandwidth available to the server associated with the other port. Just as indicated above, a threshold may be implemented on either or both of the ports (340, 345). If the activity at one of the ports (340, 345) dropsdrops below the threshold assigned it, the assigned cap at the other port is subsequently released to allow the second server make use of the unused bandwidth. -
FIG. 4 illustrates how a cap and threshold may be based upon a client/array port pair. As shown inFIG. 4 , clients with different priorities (400, 410) may be commonly linked to the various ports (440, 445) of the storage disk array (450) rather than grouped as inFIG. 3 . By virtue of the unique identifier WWN associated with each client, the resource manager (460) can identify which clients (400, 410) are high priority clients (400) and which ones are not (410). By recognizing which HBA (415) is associated with which client (400, 410), the resource manager (460) can place caps and/or thresholds at each port for specific clients. - By way of example only,
FIG. 4 illustrates the client/array port pair embodiment of the present invention. If port 1 (440) is “capped” at 500 IOPS for HBAs 1, 3, and 5 only (415), and Port 1 (440) has threshold of 2,000 IOPS then if total activity on Port 1 (440) drops below 2,000 IOPs, the caps forHBAs HBAs - Similar to the client/array port pair embodiment of the present invention explained above, the client/array LUN pair embodiment of the present invention uses the resource manager (460) to identify the client (400, 410) requesting bandwidth performance and applying corresponding caps and thresholds to the individual logic unit (455) in the storage disk array (450) rather than the ports (440, 445). The dynamic management of the bandwidth resources is still triggered by a drop in activity below the threshold. However, in this embodiment, the activity is measured at the individual logic unit (455).
- This invention allows for a fine level of user control over the threshold settings. By setting a threshold, the array can relax performance caps when they are not needed and, thereby, not unduly restrict the bandwidth available to the capped port, client/array port pairs, or host/LUN pairs. When all of the bandwidth resources arc being used, they are distributed according to priority designations. If, however, activity drops below threshold values, caps for each group may be released to allow for a dynamic redistribution of the available resources. The settings can be made in either I/O per second or MB per second.
- Often computer systems have periodic spurts of activity or schedules on which they operate. Events such as nightly backups or daytime merchant hours affect the quantity of I/O traffic and the required quality of service. In order to maximize the efficiency of bandwidth resources, the present invention allows the caps and threshold to be time dependant corresponding to predicted spurts of activities.
- In an additional embodiment of the present invention is demonstrated in
FIG. 5 . As shown inFIG. 5 , the priority of a client may be determined by its application performance requirements. Using the command line interface or script API of a storage management software application, a user application can interrogate various storage arrays and add its connectivity port to an array performance group which meets the application's transaction bandwidth. - As demonstrated in
FIG. 5 , this embodiment allows for the dynamic grouping of host applications by acquiring information on available performance groups within the storage array. - Initially, the user application evaluates an established group (610). The user application first determines if the performance cap of the first established group is greater than or equal to the application's bandwidth requirement (620). If it is, there is sufficient room in the established group to include the desired application. In that case, the port associated with the application is added to the first established group (630). If, however, the performance cap of the first established group is less than the application bandwidth requirement, the user application determines if there are additional groups (640).
- If there are additional groups, the application again evaluates the cap (620) to see if there is sufficient bandwidth to perform the desired application. If all of the groups have been considered and none meet the application bandwidth requirements, the application is added to the group with the highest cap (650).
- The embodiment of the present invention disclosed above allows for dynamic grouping of server priorities. In this manner the groups may be arranged so as to utilize the maximum bandwidth available.
- Using the command line interface, script API or Web-based Graphical User Interface (GUI) of a storage management software application, a user application can also dynamically add connectivity bandwidth and increase the performance capability of a host application by controlling multiple connectivity paths between the host computer and the storage array.
- When a host application indicates a desire for increased performance capability, the resource manager can then increase available bandwidth accordingly by dedicating additional ports to the host application. This embodiment of the present invention can then be tied into billing applications for demand-based performance such as pay for performance applications.
- The preceding description has been presented only to illustrate and describe the invention. It is not intended to be exhaustive or to limit the invention to any precise form disclosed. Many modifications and variations are possible in light of the above teaching.
- The foregoing embodiments were chosen and described in order to illustrate principles of the invention and its practical applications. The preceding description is intended to enable others skilled in the art to best utilize the invention in various embodiments and with various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims.
Claims (14)
1-39. (canceled)
40. A storage area network comprising:
a plurality of client devices; and
a data storage system communicating with said plurality of client devices, wherein said data storage system receives Input/Output (I/O) requests from the plurality of client devices to both read from and write data to said data storage system;
wherein said data storage system dynamically allocates bandwidth resources for said (I/O) requests to said plurality of client devices based on a priority assigned to each said client device;
wherein said data storage system automatically assigns a priority to a new client device based on performance requirements reported electronically by that client.
41. A storage area network according to claim 40 , further comprising a resource manager of said data storage system for allocating said bandwidth resources among said client device, wherein said resource manager:
sets an upper limit on an amount of bandwidth allocated to each of said client devices; and
sets a lower threshold on an amount of bandwidth allocated to each of said client devices;
wherein if said bandwidth resources used by one or more client devices are lower than a corresponding lower threshold, upper limits for one or more other client devices are released.
42. A storage area network according to claim 41 , wherein said upper limit is substantially equal to said lower threshold value.
43. A storage area network according to claim 40 , wherein said data storage system comprises a disk array.
44. A storage area network according to claim 40 , further comprising a fibre channel loop connected between said plurality of client devices and said data storage system.
45. A storage area network according to claim 40 , wherein said plurality of client devices comprise a network server.
46. A storage area network according to claim 40 , wherein said plurality of client devices are divided into groups according to said priority, wherein a group of client devices all receive a same bandwidth allocation setting.
47. A storage area network according to claim 46 , further comprising a plurality of ports of said data storage system, wherein all client devices of one of said groups are connected to said data storage system through a single one of said ports.
48. A storage area network according to claim 46 , wherein said priority levels each require a different fee corresponding to said bandwidth allocation.
49. A storage area network according to claim 41 , wherein said upper limit and lower threshold are measured in Megabytes per second.
50. A storage area network according to claim 41 , wherein said upper limit and lower threshold are measured in input and output operations per second.
51. A storage area network according to claim 41 , wherein either or both of said upper limit and lower threshold vary with time.
52. A storage area network according to claim 40 , wherein a client reports its performance requirements using a command line interface or script Application Programming Interface of a storage management software application.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/246,505 US20120016950A1 (en) | 2002-08-30 | 2011-09-27 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/233,234 US8060643B2 (en) | 2002-08-30 | 2002-08-30 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
US13/246,505 US20120016950A1 (en) | 2002-08-30 | 2011-09-27 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/233,234 Division US8060643B2 (en) | 2002-08-30 | 2002-08-30 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
Publications (1)
Publication Number | Publication Date |
---|---|
US20120016950A1 true US20120016950A1 (en) | 2012-01-19 |
Family
ID=31977188
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/233,234 Expired - Fee Related US8060643B2 (en) | 2002-08-30 | 2002-08-30 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
US13/246,505 Abandoned US20120016950A1 (en) | 2002-08-30 | 2011-09-27 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/233,234 Expired - Fee Related US8060643B2 (en) | 2002-08-30 | 2002-08-30 | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
Country Status (1)
Country | Link |
---|---|
US (2) | US8060643B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100312944A1 (en) * | 2009-06-04 | 2010-12-09 | Micron Technology, Inc. | Control of page access in memory |
Families Citing this family (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8060643B2 (en) * | 2002-08-30 | 2011-11-15 | Hewlett-Packard Development Company, L.P. | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
US7889761B2 (en) * | 2002-09-17 | 2011-02-15 | Broadcom Corporation | Method and system for providing bandwidth allocation and sharing in a hybrid wired/wireless network |
US7324523B2 (en) * | 2003-03-26 | 2008-01-29 | Sony Corporation | System and method for dynamically allocating bandwidth to applications in a network based on utility functions |
US7747255B2 (en) * | 2003-03-26 | 2010-06-29 | Sony Corporation | System and method for dynamic bandwidth estimation of network links |
TWI227398B (en) * | 2003-04-15 | 2005-02-01 | Asustek Comp Inc | Automatic adjusting device of computer system performance |
US7523200B2 (en) * | 2003-07-02 | 2009-04-21 | International Business Machines Corporation | Dynamic access decision information module |
US20100011090A1 (en) * | 2008-07-14 | 2010-01-14 | Limelight Networks, Inc. | Network-aware download manager |
US7688733B1 (en) * | 2003-08-04 | 2010-03-30 | Sprint Communications Company L.P. | System and method for bandwidth selection in a communication network |
US7640316B2 (en) * | 2003-09-05 | 2009-12-29 | International Business Machines Corporation | Apparatus and method to write information to two virtual tape servers |
US7484040B2 (en) * | 2005-05-10 | 2009-01-27 | International Business Machines Corporation | Highly available removable media storage network environment |
US20070130344A1 (en) * | 2005-11-14 | 2007-06-07 | Pepper Timothy C | Using load balancing to assign paths to hosts in a network |
US7797468B2 (en) * | 2006-10-31 | 2010-09-14 | Hewlett-Packard Development Company | Method and system for achieving fair command processing in storage systems that implement command-associated priority queuing |
US20080109867A1 (en) * | 2006-11-07 | 2008-05-08 | Microsoft Corporation | Service and policies for coordinating behaviors and connectivity of a mesh of heterogeneous devices |
EP2003556A1 (en) * | 2007-05-25 | 2008-12-17 | Axalto SA | Method of processing by a portable electronical device of applicative commands originating from physical channels, corresponding device and system |
JP5098886B2 (en) * | 2008-08-18 | 2012-12-12 | 富士通株式会社 | Network group management method |
US20110238857A1 (en) * | 2010-03-29 | 2011-09-29 | Amazon Technologies, Inc. | Committed processing rates for shared resources |
US8793334B1 (en) * | 2010-07-19 | 2014-07-29 | Applied Micro Circuits Corporation | Network-attached storage (NAS) bandwidth manager |
US9466036B1 (en) * | 2012-05-10 | 2016-10-11 | Amazon Technologies, Inc. | Automated reconfiguration of shared network resources |
JP6179321B2 (en) * | 2013-09-27 | 2017-08-16 | 富士通株式会社 | Storage management device, control method, and control program |
JP6273966B2 (en) | 2014-03-27 | 2018-02-07 | 富士通株式会社 | Storage management device, performance adjustment method, and performance adjustment program |
US9450879B2 (en) | 2014-05-09 | 2016-09-20 | Nexgen Storage, Inc. | Adaptive bandwidth throttling |
US10474383B1 (en) | 2016-12-29 | 2019-11-12 | EMC IP Holding Company LLC | Using overload correlations between units of managed storage objects to apply performance controls in a data storage system |
US10681572B2 (en) | 2017-03-30 | 2020-06-09 | International Business Machines Corporation | Dynamic bandwidth analysis for mobile devices |
US10990447B1 (en) * | 2018-07-12 | 2021-04-27 | Lightbits Labs Ltd. | System and method for controlling a flow of storage access requests |
CN111046007B (en) * | 2018-10-12 | 2023-08-18 | 伊姆西Ip控股有限责任公司 | Method, apparatus and computer program product for managing a storage system |
US11481156B1 (en) * | 2020-02-26 | 2022-10-25 | Marvell Asia Pte, Ltd. | Throughput management of distributed storage system |
US12363050B2 (en) * | 2021-11-29 | 2025-07-15 | Dish Network Technologies India Private Limited | Internet bandwidth/data management by gateway device |
Family Cites Families (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5778165A (en) * | 1995-10-20 | 1998-07-07 | Digital Equipment Corporation | Variable-level backup scheduling method and apparatus |
CN1094277C (en) * | 1996-03-18 | 2002-11-13 | 通用仪器公司 | Dynamic bandwidth allocation for communication network |
US6886035B2 (en) * | 1996-08-02 | 2005-04-26 | Hewlett-Packard Development Company, L.P. | Dynamic load balancing of a network of client and server computer |
US5953338A (en) * | 1996-12-13 | 1999-09-14 | Northern Telecom Limited | Dynamic control processes and systems for asynchronous transfer mode networks |
EP1008051A4 (en) * | 1997-03-12 | 2007-04-25 | Storage Technology Corp | Network attached virtual tape data storage subsystem |
US6968379B2 (en) * | 1997-05-30 | 2005-11-22 | Sun Microsystems, Inc. | Latency-reducing bandwidth-prioritization for network servers and clients |
US6003062A (en) * | 1997-07-16 | 1999-12-14 | Fore Systems, Inc. | Iterative algorithm for performing max min fair allocation |
US7079534B1 (en) * | 1998-02-20 | 2006-07-18 | Sprint Communications Company L.P. | System and method for allocating bandwidth for a call |
US6631477B1 (en) * | 1998-03-13 | 2003-10-07 | Emc Corporation | Host system for mass storage business continuance volumes |
CA2332413A1 (en) * | 1998-05-15 | 1999-11-25 | Rick W. Landsman | A technique for implementing browser-initiated network-distributed advertising and for interstitially displaying an advertisement |
US20020133589A1 (en) * | 1998-09-11 | 2002-09-19 | Rajugopal R. Gubbi | Dynamic bandwidth negotiation scheme for wireless computer networks |
US6421723B1 (en) | 1999-06-11 | 2002-07-16 | Dell Products L.P. | Method and system for establishing a storage area network configuration |
US6343324B1 (en) | 1999-09-13 | 2002-01-29 | International Business Machines Corporation | Method and system for controlling access share storage devices in a network environment by configuring host-to-volume mapping data structures in the controller memory for granting and denying access to the devices |
US7072295B1 (en) * | 1999-09-15 | 2006-07-04 | Tellabs Operations, Inc. | Allocating network bandwidth |
FR2801455B1 (en) * | 1999-11-23 | 2002-02-22 | France Telecom | METHOD FOR TRANSMITTING DATA STREAMS OVER AN ATM NETWORK, AND DEVICE FOR IMPLEMENTING THE METHOD |
US6571354B1 (en) * | 1999-12-15 | 2003-05-27 | Dell Products, L.P. | Method and apparatus for storage unit replacement according to array priority |
US6847984B1 (en) * | 1999-12-16 | 2005-01-25 | Livevault Corporation | Systems and methods for backing up data files |
US6526418B1 (en) * | 1999-12-16 | 2003-02-25 | Livevault Corporation | Systems and methods for backing up data files |
US6460055B1 (en) * | 1999-12-16 | 2002-10-01 | Livevault Corporation | Systems and methods for backing up data files |
US6564336B1 (en) * | 1999-12-29 | 2003-05-13 | General Electric Company | Fault tolerant database for picture archiving and communication systems |
GB0001804D0 (en) * | 2000-01-26 | 2000-03-22 | King S College London | Pre-emptive bandwidth allocation by dynamic positioning |
US20050259682A1 (en) * | 2000-02-03 | 2005-11-24 | Yuval Yosef | Broadcast system |
AU4195601A (en) * | 2000-03-01 | 2001-09-12 | Computer Ass Think Inc | Method and system for updating an archive of a computer file |
US20020049841A1 (en) * | 2000-03-03 | 2002-04-25 | Johnson Scott C | Systems and methods for providing differentiated service in information management environments |
US20020129048A1 (en) | 2000-03-03 | 2002-09-12 | Surgient Networks, Inc. | Systems and methods for resource monitoring in information storage environments |
US20020108059A1 (en) | 2000-03-03 | 2002-08-08 | Canion Rodney S. | Network security accelerator |
US6671724B1 (en) * | 2000-03-21 | 2003-12-30 | Centrisoft Corporation | Software, systems and methods for managing a distributed network |
US7009992B2 (en) * | 2000-05-19 | 2006-03-07 | Scientific-Atlanta, Inc. | Methods of allocating access across a shared communications medium |
US7499453B2 (en) * | 2000-05-19 | 2009-03-03 | Cisco Technology, Inc. | Apparatus and methods for incorporating bandwidth forecasting and dynamic bandwidth allocation into a broadband communication system |
NZ523328A (en) * | 2000-07-05 | 2005-04-29 | Ernst & Young Llp | Method and apparatus for providing computer services |
US6704885B1 (en) * | 2000-07-28 | 2004-03-09 | Oracle International Corporation | Performing data backups with a stochastic scheduler in a distributed computing environment |
US20020103772A1 (en) * | 2001-01-31 | 2002-08-01 | Bijoy Chattopadhyay | System and method for gathering of real-time current flow information |
US20030067874A1 (en) * | 2001-10-10 | 2003-04-10 | See Michael B. | Central policy based traffic management |
JP3879471B2 (en) * | 2001-10-10 | 2007-02-14 | 株式会社日立製作所 | Computer resource allocation method |
US20030100307A1 (en) * | 2001-11-27 | 2003-05-29 | Intel Corporation | Automatic service handoff method and apparatus |
US7085848B2 (en) * | 2002-03-15 | 2006-08-01 | Microsoft Corporation | Time-window-constrained multicast using connection scheduling |
US20040015602A1 (en) * | 2002-07-19 | 2004-01-22 | Roving Planet, Inc. | Network bandwidth allocation and access method and apparatus |
US8060643B2 (en) * | 2002-08-30 | 2011-11-15 | Hewlett-Packard Development Company, L.P. | Method and apparatus for dynamically managing bandwidth for clients in a storage area network |
US7586944B2 (en) * | 2002-08-30 | 2009-09-08 | Hewlett-Packard Development Company, L.P. | Method and system for grouping clients of a storage area network according to priorities for bandwidth allocation |
US7130890B1 (en) * | 2002-09-04 | 2006-10-31 | Hewlett-Packard Development Company, L.P. | Method and system for adaptively prefetching objects from a network |
-
2002
- 2002-08-30 US US10/233,234 patent/US8060643B2/en not_active Expired - Fee Related
-
2011
- 2011-09-27 US US13/246,505 patent/US20120016950A1/en not_active Abandoned
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100312944A1 (en) * | 2009-06-04 | 2010-12-09 | Micron Technology, Inc. | Control of page access in memory |
US8380916B2 (en) * | 2009-06-04 | 2013-02-19 | Micron Technology, Inc. | Control of page access in memory |
US8738837B2 (en) | 2009-06-04 | 2014-05-27 | Micron Technology, Inc. | Control of page access in memory |
US20140258649A1 (en) * | 2009-06-04 | 2014-09-11 | Micron Technology, Inc. | Control of page access in memory |
US9176904B2 (en) * | 2009-06-04 | 2015-11-03 | Micron Technology, Inc. | Control of page access in memory |
US9524117B2 (en) | 2009-06-04 | 2016-12-20 | Micron Technology, Inc. | Control of page access in memory |
Also Published As
Publication number | Publication date |
---|---|
US20040044770A1 (en) | 2004-03-04 |
US8060643B2 (en) | 2011-11-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20120016950A1 (en) | Method and apparatus for dynamically managing bandwidth for clients in a storage area network | |
US7586944B2 (en) | Method and system for grouping clients of a storage area network according to priorities for bandwidth allocation | |
US5996014A (en) | Distribution multimedia server system using shared disk arrays connected in a chain and having two ports each that are striped with digitalized video data | |
EP1370950B1 (en) | System and method for policy based storage provisioning and management | |
US20220413976A1 (en) | Method and System for Maintaining Storage Device Failure Tolerance in a Composable Infrastructure | |
JP4264001B2 (en) | Quality of service execution in the storage network | |
US20080162735A1 (en) | Methods and systems for prioritizing input/outputs to storage devices | |
EP1071989B1 (en) | Intelligent data storage manager | |
US7568037B2 (en) | Apparatus and method for using storage domains for controlling data in storage area networks | |
US8732339B2 (en) | NPIV at storage devices | |
US6976134B1 (en) | Pooling and provisioning storage resources in a storage network | |
US7032041B2 (en) | Information processing performing prefetch with load balancing | |
US7428614B2 (en) | Management system for a virtualized storage environment | |
US7984251B2 (en) | Autonomic storage provisioning to enhance storage virtualization infrastructure availability | |
CA2498174C (en) | Dynamic storage device pooling in a computer system | |
US11474704B2 (en) | Target path selection for storage controllers | |
US20030079018A1 (en) | Load balancing in a storage network | |
US20020004883A1 (en) | Network attached virtual data storage subsystem | |
JP4975399B2 (en) | Resource management method in logical partitioning storage system | |
JP2005321959A (en) | Information processing apparatus, information processing apparatus control method, and program | |
JP2004199697A (en) | Method and device for dynamically assigning storage array band | |
US20040181594A1 (en) | Methods for assigning performance specifications to a storage virtual channel | |
US20040181589A1 (en) | Storage virtual channels and method for using the same | |
US20100242048A1 (en) | Resource allocation system | |
US9065740B2 (en) | Prioritising data processing operations |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION |