US20200267219A1 - Systems and methods for granular resource management in a storage network - Google Patents

Systems and methods for granular resource management in a storage network Download PDF

Info

Publication number
US20200267219A1
US20200267219A1 US16/800,284 US202016800284A US2020267219A1 US 20200267219 A1 US20200267219 A1 US 20200267219A1 US 202016800284 A US202016800284 A US 202016800284A US 2020267219 A1 US2020267219 A1 US 2020267219A1
Authority
US
United States
Prior art keywords
storage
data
storage operation
policies
media
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/800,284
Inventor
Rajiv Kottomtharayil
Parag Gokhale
Marcus S. Muller
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Commvault Systems Inc
Original Assignee
Commvault Systems Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Commvault Systems Inc filed Critical Commvault Systems Inc
Priority to US16/800,284 priority Critical patent/US20200267219A1/en
Assigned to COMMVAULT SYSTEMS, INC. reassignment COMMVAULT SYSTEMS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MULLER, MARCUS S., GOKHALE, PARAG, KOTTOMTHARAYIL, RAJIV
Publication of US20200267219A1 publication Critical patent/US20200267219A1/en
Assigned to JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT reassignment JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: COMMVAULT SYSTEMS, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/66Arrangements for connecting between networks having differing types of switching systems, e.g. gateways
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/70Admission control; Resource allocation

Definitions

  • the inventions disclosed herein relate generally to performing storage operations on electronic data in a computer network. More particularly, aspects of the present invention relate to data transmission schemes used during a storage operation including data pathways and other components used in the transfer of data.
  • RAM Random Access Memory
  • Serial, parallel, Small Computer System Interface (SCSI), or other cables directly connect such stand-alone storage devices to individual computers that are part of a network of other computers such as a Local Area Network (LAN) or a Wide Area Network (WAN).
  • LAN Local Area Network
  • WAN Wide Area Network
  • each individual computer on the network controlled the storage devices that were physically attached to that computer and could also access the storage devices of the other network computers to perform backups, transaction processing, file sharing, and other storage-related operations.
  • NAS Network Attached Storage
  • a storage controller computer typically controls the storage device to the exclusion of other computers on the network, but the SCSI or other cabling directly connecting that storage device to the individual controller is eliminated. Instead, storage devices are directly attached to the network itself.
  • Yet another network storage scheme is modular storage architecture which is more fully described in U.S. Pat. Nos. 7,035,880 and 6,542,268.
  • An example of such a software application is the GalaxyTM system, by CommVault Systems of Oceanport, N.J.
  • the GalaxyTM system is a multi-tiered storage management solution which includes, among other components, a storage manager, one or more media agents, and one or more storage devices.
  • the storage manager directs storage operations of client data to storage devices such magnetic and optical media libraries.
  • Media agents are storage controller computers that serve as intermediary devices managing the flow of data from client information stores to individual storage devices. Each storage device may be uniquely associated with a particular media agent and this association may be tracked by the storage manager.
  • a common feature shared by all of the above-described network architectures is the substantially static relationship between storage controller computers and storage devices.
  • storage devices are generally connected, virtually or physically, to a single storage controller computer.
  • One computer typically cannot control the drive pool and media group be that is being controlled by another. Requests to store and retrieve data from such a drive pool and media group would have to be coordinated by the controlling computer.
  • storage media reserved or being written to by one media agent cannot be written to be another media agent.
  • storage media being used pursuant to one storage policy cannot be used by another storage policy and vice versa often resulting in the inefficient use of storage resources.
  • storage policies may specify alternate data paths or resources in the case device failure or an otherwise unavailable data path.
  • such systems typically specify a single alternate data path.
  • backup operations are traditional performed on a client by client basis, each client may store information on different media, resulting in inefficient media use.
  • failover conditions often result in the use of additional media further resulting in inefficient use of resources.
  • alternate data paths are defined in a static fashion, and thus conventional data protection schemes are unable to adapt to changing network conditions.
  • systems and methods are provided for dynamically or automatically selecting and/or modifying data path definitions that are used in performing storage operations.
  • Alternate data paths may be specified or selected that use some or all resources that communicate with a particular destination to improve system reliability and performance.
  • the system may also dynamically monitor and choose data path definitions to optimize system performance, conserve storage media, prevent resource exhaustion and promote balanced load distribution.
  • a method for configuring a storage operation system includes defining a first storage operation path to be used in performing a storage operation.
  • the first storage operation path may specify a destination and substantially all of the resources capable of communicating with the destination.
  • the system may define a second storage operation path used in the storage operation when the first storage path is unavailable.
  • a storage operation system may include a management module for controlling or coordinating a storage operation to a destination, a plurality of storage devices, and at least two storage operation paths linking a client to one or more storage devices.
  • the first storage operation path may specify many, most or substantially all of the resources capable of communicating with the destination, while the second storage operation path may be used in the storage operation when the first storage path is unavailable.
  • a method for consolidating storage policies within a storage operation network may include, analyzing storage operation paths, which may defined in storage policies. Determining whether any of the storage operation paths have common element points and consolidating two or more of the storage policies having at least one common element into a single storage operation policy such that the single storage operation policy supports copy operations to or with the common element point such as a common destination.
  • the system may include a management module for directing a storage operation to a destination, a plurality of storage devices and a plurality of storage operation paths.
  • the storage operation paths may be defined within a plurality of storage policies and have a series of element points defining locations or resources along the path, ending with the destination.
  • the management module may consolidate two or more of the storage policies having at least one common element point into a single storage policy such that the single storage policy supports copy operations to the common element point.
  • One embodiment of the present invention includes a method for consolidating storage policies within a storage operation network that includes analyzing a plurality of storage operation paths that are defined in storage policies. This may involve identifying certain inefficiencies in the storage operation paths and reconfiguring the storage operation paths to improve system performance. This may further involve monitoring the storage network for the inefficiencies in the storage network subsequent to redefining the plurality of storage operation paths to determine whether the reconfiguration has achieved the desired effect.
  • the system may include a management component for controlling or coordinating a storage operation to a destination using one of a plurality of storage operation paths defined within a plurality of storage policies.
  • the management component may identify inefficiencies in the storage operation paths and reconfigure or redefine the storage operation paths to correct or improve or the modified inefficiencies.
  • the management component may also monitor the storage network including any reconfigurations subsequent to redefining the storage operation paths to determine whether the reconfigurations provided the desired correct or improvement. If, not, additional analysis and reconfiguration may be performed.
  • FIG. 1 is a block diagram of a network architecture for a system to perform storage operations on electronic data in a computer network according to an embodiment of the invention
  • FIG. 2 is a block diagram of an exemplary media storage device for performing storage operations on electronic data in a computer network according to an embodiment of the invention
  • FIG. 3 is a flow chart illustrating some of the steps of a storage operation in accordance with an embodiment of the invention
  • FIG. 4 is a flow chart illustrating some steps of assigning storage policies to system resources and evaluating existing storage policies of in accordance with an embodiment of the invention.
  • FIG. 5 is a flow chart illustrating some of the steps of a method of dynamically analyzing and managing storage policies and data paths in accordance with an embodiment of the invention.
  • FIG. 1 a block diagram of one network architecture suitable for performing storage operations on electronic data in a computer network according to an embodiment of the invention is shown.
  • the embodiment may include a storage management component such as manager 100 and one or more of the following: a client 85 , an information store 90 , a data agent 95 , a media agent 105 , an index cache 110 , and a storage device 115 .
  • the system and elements thereof are exemplary of a three-tier backup system such as the CommVault GalaxyTM backup system, available from CommVault Systems, Inc. of Oceanport, N.J., and further described in U.S. Pat. No. 7,035,880 which is incorporated herein by reference in its entirety.
  • a data agent 95 is generally a software module that responsible for archiving, migrating, and recovering data of a client computer 85 stored in an information store 90 or other memory location.
  • Each client computer 85 may have one or more data agent(s) 95 and the system can support multiple client computers 85 .
  • the system may include a plurality of data agents 95 each of which is intended to backup, migrate, and recover data associated with a different application.
  • different individual data agents 95 may be designed to handle Microsoft Exchange® data, Lotus Notes® data, Microsoft Windows 2000® file system data, Microsoft Active Directory Objects® data, and other types of data known in the art.
  • a dedicated data agent 95 may be used for each data type to archive, migrate, and restore the client computer 85 data.
  • the client computer 85 would use one Microsoft Exchange 2000® Mailbox data agent 95 to backup the Exchange 2000® mailboxes, one Microsoft Exchange 2000® Database data agent 95 to backup the Exchange 2000® databases, one Microsoft Exchange 2000® Public Folder data agent 95 to backup the Exchange 2000® Public Folders, and one Microsoft Windows 2000® File System data agent 95 to backup the client computer's 85 file system.
  • These data agents 95 would be treated as four separate data agents 95 by the system even though they reside on the same client computer 85 .
  • multipurpose or generic data agents may be used that operate on multiple data without types.
  • one data agent may operate on Microsoft Exchange 2000® Mailbox and Microsoft Windows 2000® File System data, etc.
  • Storage manager 100 may be implemented as a software module or application that coordinates and controls various aspects of the system shown in FIG. 1 .
  • storage manager 100 may communicate with some or all elements of the system including client computers 85 , data agents 95 , media agents 105 , and storage devices 115 , to schedule, initiate, manage and coordinate system backups, migrations, and data recoveries.
  • a media agent 105 is may be implemented as a software module that conducts data, as directed by storage manager 100 , between the client computer 85 and one or more storage devices 115 such as a tape library, a magnetic media storage device, an optical media storage device, or other storage device known in the art.
  • storage manager 100 may direct data agents 95 to copy data from one or more clients 85 to storage device 115 through media agents 105 .
  • media agent 105 communicates with and controls the storage device 115 .
  • media agent 105 may instruct storage device 115 to use a robotic arm or other means to load or eject a media cartridge, to archive, migrate, or restore data to or from certain media present in device 115 .
  • Media agents 105 may also communicate with the storage devices 115 via a local bus such as a SCSI adaptor, or other suitable connection means.
  • storage device 115 may communicate to the data agent 105 via a Storage Area Network (“SAN”).
  • SAN Storage Area Network
  • Each media agent 105 may maintain an index cache 110 which stores the index data the system generates during backup, migration, and restore storage operations as further described herein.
  • index cache 110 stores the index data the system generates during backup, migration, and restore storage operations as further described herein.
  • storage operations for Microsoft Exchange® data generate index data containing the location and other information such as metadata regarding the data on the storage device 145 the Exchange data is stored on.
  • Index data provides the system with an efficient mechanism for locating user files or data for recovery operations.
  • This index data is generally stored with the data backed up to the storage device 115 .
  • the media agent 105 that controls the storage operation may also write an additional copy of the index data to its index cache 110 .
  • the data in media agent 105 and index cache 110 is thus readily available to the system for use (in storage and retrieval operations and other activities) without having to be first retrieved from a storage device 115 .
  • Storage manager 100 also maintains an index cache 110 .
  • index data may include logical associations between components of the system, user preferences, metadata regarding application data or user preferences, management tasks, and other useful data.
  • the storage manager 100 may use its index cache 110 to track the logical associations between media agents 105 and storage devices 115 .
  • Index caches 110 typically reside on their corresponding storage component's hard disk or other fixed storage device. Like any cache, the index cache 110 has finite capacity and the amount of index data that can be maintained directly corresponds to the size of that portion of the disk that is allocated to the index cache 110 . In one embodiment, the system may manage the index cache 110 on a least recently used (“LRU”) basis as known in the art. When the capacity of the index cache 110 is reached, the system may overwrite those files in the index cache 110 that have been least recently used with the new index data. In some embodiments, before data in the index cache 110 is overwritten, the data may be copied to an index cache copy and stored on a storage device 115 . If a recovery operation requires index data that is no longer stored in the index cache 110 , such as in the case of a cache miss, the system may recover the index data from the copy stored in storage device 115 .
  • LRU least recently used
  • FIG. 2 a block diagram of an exemplary media library storage device 120 for performing storage operations on electronic data in a computer network according to an embodiment of the invention is presented.
  • Media library device 120 represents one specific type of storage device 115 ( FIG. 1 ) that may be used with an implementation if the invention.
  • Media library storage device 120 may contain any suitable magnetic, optical or other storage media 145 and associated drives 125 , 130 , 135 , and 140 .
  • Media 145 may store electronic data containing backups of application data, user preferences, metadata, system information, and other useful information known in the art.
  • Drives 125 , 130 , 135 and 140 are used to store and retrieve electronic data from media 145 .
  • drives 125 , 130 , 135 and 140 may function as a drive pool, as further described in application Ser. No. 10/658,095 which is hereby incorporated herein by reference in its entirety.
  • a drive pool is a logical concept that associates drives and storage media with a storage policy and a source device such as a client 85 . Storage policies representing storage patterns and preferences are more fully discussed in U.S. Pat. No. 6,542,972 which is hereby incorporated by reference herein in its entirety.
  • a drive pool may be identified by a set of drives within a library storage device 120 as pointed to by one or more media agents 105 .
  • a drive pool known as DP 1 consisting of drives 125 and 130 in library 120 known as LIB 1 may be associated by a storage policy, with a first media agent 105 MA 1 in an index cache 110 entry as follows: LIB 1 /MA 1 /DP 1 .
  • a second drive pool consisting of drives 130 , 135 , and 140 within the library storage device 120 associated with the same media agent 105 may be expressed in index cache 110 as follows: LIB 1 /MA 1 /DP 2 .
  • an index cache 110 entry for a storage policy may combine the two previous entries instead and thus may be logically represented as: ##STR 1 ##
  • media 145 may be associated by the system with drive pools or storage policies, and not necessarily with individual drives 125 , 130 , 135 and 140 .
  • a media group may be a collection of media 145 or other storage media assigned to a specific storage policy.
  • the media group may dynamically point to different drive pools as further described herein, including those with different recording formats as the system may update the recording format of the media group in a media group table stored in an index cache 110 .
  • aspects of the present invention permit data associated with a particular storage policy copy to be stored on and share certain media 145 .
  • Data from each storage policy copy may be appended to media 145 shared by other storage policy copies.
  • a storage policy copy may be shared between several media agents 105 in a dynamic drive pooling environment with media 145 also being shared by the different media agents 105 and storage policies.
  • Media 145 can be located in virtually any storage device 115 and for a given storage policy copy may be spread across multiple storage devices 115 .
  • an index cache entry may associate multiple media sets 145 with multiple media agents, storage policies, drive pools, and other system components.
  • a single index cache 110 entry may also specify and associate multiple media agents 105 , storage policies, drive pools, network pathways, and other components.
  • FIG. 3 a flow chart 300 illustrating some of the steps involved in performing storage operations on electronic data in a computer network according to an embodiment of the invention is shown. Selection of desired storage components for storage operations may be performed manually or automatically in dynamic fashion. In operation, the system may initiate a storage operation in response to a scheduled procedure or as directed by a user, system administrator, or as otherwise directed by the system (step 310 ).
  • the system may initiate a backup operation or a restore operation at a specific time of day or in response to a certain threshold being exceeded as specified in a storage policy.
  • the system may select a media agent 105 ( FIG. 1 ) according to selection logic or a specified data path as further described herein (step 320 ).
  • the selection logic and data paths may determined by a set of criteria defined in the storage policies or according to system configuration or operational rules or guidelines. Examples of such criteria may include, load balancing within the network, bandwidth use and efficiency, media usage, available media space, etc.
  • the selection logic includes the ability to conduct a LAN-free storage operation, such as using a SAN, when it is desired to optimize storage operations via load balancing.
  • a LAN-free storage operation such as using a SAN
  • an index entry in index cache 110 may associate certain media agents 105 , storage devices 115 , or other components with LAN-free storage operations either via user input, network topology detection algorithms known in the art, or other methods.
  • the system may select a free media agent 105 to optimize storage operations via load balancing when a default media agent 105 or other media agent 105 specified in a storage policy is already performing other storage operations or otherwise occupied.
  • the system may also select an appropriate drive pool in a network storage device according to selection logic further described herein (step 330 ). Once the system has selected an appropriate media agent and drive pool, the storage operation is performed, using the selected storage components (step 340 ).
  • Another embodiment of the present invention allows storage policies to be recognized or be defined in terms of sub-clients (e.g., processes or portions of data of a volume that are mutually exclusive) and have data protection operations performed at the sub-client level.
  • a storage policy may specify a path similar to those described above for each sub-client operating on a client.
  • storage policies associated with each sub-client specify a default data path and one or more alternate data paths.
  • media agent 2 library 1
  • Alternate data paths are desirable as they provide additional means by which a storage operation may be completed and thus improve system reliability and promote robust operation.
  • the system may automatically select certain available alternate data paths to facilitate load balancing and failover recovery.
  • Such alternate data paths may be specified using some or all of the additional routing resources available in the system.
  • the alternate data path above may specify as alternates some or all of the media agents in the system that are capable of communicating with library 1 and may be expressed as follows:
  • media agent 2 media agent 3; . . . media agent n: library 1
  • n is the total number of media agents specified in the alternate data path.
  • a storage policy may specify all of the media agents 105 in the system capable of communicating with a particular destination (e.g., library 1). This provides the greatest likelihood that a storage operation will be completed, assuming that at least some storage resources are functioning or not otherwise congested.
  • some of the available routing resources may be specified as alternates, providing a greater likelihood that the storage operation will be completed, rather than relying on a single alternate.
  • Such alternate resources may be selected based on the degree of utilization, capacity, bandwidth, physical location, the desired confidence factor or other considerations and may be specified manually or assigned automatically based on data protections goals specified for the system.
  • alternate data paths may be specified in many ways to provide robust routing options.
  • alternate data paths may be specified according to user preferences.
  • a system administrator may specify certain alternate data paths and the priority and/or order in which the data paths are to be used.
  • Another method for providing alternate data paths may involve using the “round robin” approach in which alternate data paths are selected from a group of available data paths such that each alternate data path is selected and used before any previously used data path is selected and used again. This approach is typically useful in promoting load balancing within the system as it tends to spread out data transfer operations across available data paths in a substantially uniform fashion.
  • Other approaches may include specifying alternate data paths to emphasize the ability to complete a storage operation in the event of a failover condition.
  • alternate data paths may be specified such that data from one client or sub-client may be routed to a particular destination through substantially every available data path that may potentially link the client or sub-client to the destination.
  • alternate data paths may be defined such that data is restored to particular client, sub-client or computing device from some or all storage devices within the system. This arrangement provides significant flexibility within the system for performing and completing both storage and restore operations.
  • alternate destinations may be specified and used in failover or other emergency data protection operations.
  • a storage policy may specify a data path including library 1, library 2, and others, with the provision that the specified media agents have access to each of the specified libraries.
  • media agents specified in such data paths may share an index cache.
  • storage policies may not have the ability to share storage media due to certain conflicts within programming logic or the need for storage policies to resolve any such conflicts in mutually exclusive manner to ensure computational integrity.
  • storage operations performed pursuant to different storage policies are generally required to write to different media, often resulting in the inefficient use of media.
  • two groups of clients may specify two sets of client or sub-client data paths (e.g., a default and alternate for each), but, may be governed by a single storage policy in accordance with one embodiment of the present invention. This may be accomplished by examining the data paths and combining or rearranging them into a suitable form for use in the storage policy.
  • a first group of sub-clients may specify the following data paths:
  • the second group of sub-clients may specify the following data paths:
  • the system may consult this modified storage policy (default first and alternate second) to obtain data path preferences when moving data from the first set of clients.
  • this storage policy may be consulted in reverse order, thus preserving the original preferences (i.e., Media agent 2 as the default with Media agent 1 as the alternate).
  • data from various storage operations may be written to the same media, rather than using separate media as explained above, promoting media conservation.
  • specifying data paths on a sub-client level allows multiple clients to write data to the same media and avoids the potential logical conflicts described above.
  • Another embodiment in accordance with the present invention includes the case where two client domains separated by a firewall, each domain containing multiple sub-clients. Assume, for example that each domain has set of sub-clients with different data paths as shown below:
  • these data paths may be modified (or combined) into one storage policy set forth below having four specified data paths rather than two storage polices with two data paths each, thus maintaining failover protection and promote minimum media utilization:
  • the first entry When moving data from the first domain, the first entry is consulted and vice versa for the second domain which allows information from both domains to be written to the same media, promoting efficient media utilization.
  • This order of operations may be defined within the storage policy or may be specified by placing the appropriate pointers or other referential elements in an index or other entry that governs data path preferences.
  • Another benefit of the present invention includes the ability to use a single storage policy to govern multiple clients. This provides users with significant flexibility by allowing them to define a storage policy and “point” to that policy through referential elements to multiple clients, thereby simplifying system administration.
  • a single policy may be changed having a global effect rather than requiring a similar change be made to multiple individual polices.
  • clients may easily be assigned or moved from one storage policy to another merely by changing a pointer or other referential element. This eliminates having to copy, significantly modify, change or create a new storage policy from scratch. Further, storage policies are no longer defined and associated on an individual client by client basis.
  • storage policies and associated storage domains may be associated with one another based on system configuration, user needs, or other considerations. This process may be performed either manually, automatically, or may be partially automated, requiring certain user input such as customization information, intended or expected use, etc. For example, at system setup a configuration program may walk an administrator through a configuration program and prompt the user for certain customization information. In alternative embodiments, this process may be predominantly or completely automated based on certain specific goals including, but not limited to, efficient media usage, degree of desired data protection, and substantially even and/or efficient load distribution.
  • Flowchart 400 of FIG. 4 illustrates some of the steps involved in assigning storage policies to system resources or in evaluating existing storage polices for possible consolidation as part of an ongoing effort to analyze and increase system efficiency.
  • any existing storage polices or defined data paths for performing storage operations may be retrieved, examined and analyzed. This may involve, for example, retrieving path information from an index cache associated with a media agent or master storage manager or retrieving similar information from a metabase that may be associated with such components.
  • the analysis may include examining data path information such as origination point (e.g., clients and/or sub-clients), destination point (storage device, library, media pool, etc.), transmission resources scheduled to be involved including media agents, data conduits and other transmission elements. In some embodiments, this may involve the creation of a system wide or more limited process-based netlist to obtain a basic understanding of system routing options and transmission patterns and preferences.
  • the system may determine whether any identified clients or sub-clients have a common destination point.
  • the destination points are typically defined as a storage device for receiving data from copy operations representing the last location of data at the completion of a particular copy operation.
  • a list of origination points (e.g., clients and or media agents) and common destination points may be compiled as a starting point to determine similarities between various identified data paths that may be suitable for combination or rearrangement into one or more storage policy to improve overall system efficiency and/or reduce media consumption.
  • media agents and other data transfer resources may be associated with the list to generate a more complete picture of the routes and resources involved/available in traversing the data paths between origination and destination points.
  • the netlist may be substantially complete taking into account available routing and resource information.
  • the system examines any pre-existing or identified storage polices and compares them with other storage polices and the information generated at step 430 to identify common elements that may be combined or folded into the existing storage policies. This may also involve identifying and comparing clients/sub-clients with common origination points and correlating them with storage polices having common destination points as a basis for potentially creating new storage polices. Other information of interest may include identifying common media agents and associated destination points, etc. for similar reasons.
  • step 450 it may be determined whether any identified common elements are precluded from writing information to the same storage device and/or storage media. If so, in some embodiments, these elements may be noted on a list of items not suitable for combination to into storage polices and may be identified as needing individual treatment. The gathered information may be analyzed to determine if the number of storage polices may be reduced by combining common elements, by combining or modifying existing storage polices, or by recasting storage polices with other identified data paths into more efficient storage pathways (step 460 ).
  • the analysis may reveal four existing storage polices that have many common elements. Depending on the management goals of the storage system, these four storage polices may be combined into one comprehensive storage policy with a common destination if minimal media usage is desired or may be combined into two storage polices to minimize the possibility of alternate data path congestion.
  • factors other than media consumption or possible congestion may be taken into account when determining how to create, change or modify storage polices to accommodate certain system management goals.
  • considerations may include load balancing, optimization, service level performance or other operational goals including adjustments to account for changes that may occur over time.
  • a system administrator may wish to maintain a substantially even workload across the storage network and maintain that distribution on a going forward basis.
  • Other goals may include maintaining operational performance within a certain percentage level to ensure a specified level of data protection or maximizing system efficiency during peak usage periods. Achieving these and other goals may involve the dynamic and periodic redefinition of data paths and associated storage policies
  • FIG. 5 a flow chart 500 illustrating some of the steps involved with the dynamic analysis and potential redefinition of storage policies/selection of alternate data paths in accordance with aspects of the present invention is illustrated.
  • client/sub-client data paths within the system are analyzed similarly to step 410 described above in connection with FIG. 4 .
  • step 520 with the data paths identified, certain system performance and forecasting reports may be run as described in co-pending, commonly assigned cases entitled Systems and Methods for Allocation of Organizational Resources Application, and Hierarchical Systems and Methods for Providing a Unified View of Storage Information, Ser. No. 11/120,619, filed May 2, 2005, which are hereby incorporated by reference in its entirety.
  • Such reports may forecast, based on past performance or other parameters, how resource utilization may grow or otherwise change and predict how capacity, efficiency, failure rates, and traffic load may impact storage operations over time.
  • the system analyzes, at step 530 , data paths to identify which ones are susceptible to or likely to experience an adverse impact due to the changing conditions (e.g., based on predefined thresholds or resource capacity). Such data path definitions or storage policies may then be modified on a dynamic basis to accommodate or otherwise account for predicted conditions to minimize impact (step 540 ). For example, if it is determined that certain data paths are expected to become congested after a certain period of time, additional alternate data paths expected to handle the additional load may be added before that point is reached or other alternate data paths that do not suffer from the same conditions are specified.
  • additional alternate media agents may be added by combining or otherwise altering storage policy data paths to help reduce the adverse impact of the anticipated problem (e.g., other alternate data paths not suffering from the same or similar conditions may be selected or added to help alleviate any detected or predicted problem).
  • data paths may be changed on a dynamic basis to balance load, maintain a substantially constant data load, or prevent a failover condition in accordance with user specifications or system requirements.
  • this may involve distributing work load across several communication paths as described in commonly assigned, co-pending case entitled Systems and Methods for Providing Multipath Storage Network, filed on Dec. 19, 2005, and which is hereby incorporated by reference in its entirety.
  • step 550 resource reallocation is considered if data path adjustment is not sufficient to correct or acceptably minimize any anticipated problem. This may involve, for example, allocating additional storage resources such as media agents, data paths, and storage devices, etc. from other storage operation cells, as described in commonly assigned, co-pending case entitled Systems and Methods for Migrating Components on a Hierarchical Storage Network, application Ser. No. 60/752,196, filed on Dec. 19, 2005, and which is hereby incorporated by reference in its entirety. If deemed helpful, the reallocation is performed at step 560 as described in that case. The system may then periodically return to step 510 and perform the process as part of an ongoing recursive effort to maintain or optimize system performance.
  • additional storage resources such as media agents, data paths, and storage devices, etc. from other storage operation cells, as described in commonly assigned, co-pending case entitled Systems and Methods for Migrating Components on a Hierarchical Storage Network, application Ser. No. 60/752,196, filed on Dec. 19, 2005, and which is hereby incorporated by reference
  • proposed reallocation scenarios may be simulated and evaluated with the expected results extended over time in order to choose the best solution to any resource shortcoming that best fits enterprise needs or user expectations.
  • any such resource reallocation may need to be approved by an administrator prior to reallocation, which may involve reviewing simulation results and approving reallocations on a component by component or proposal by proposal basis.
  • such resource reallocation may be performed substantially automatically.
  • the system monitors performance subsequent to resource reallocation to help confirm the reallocation is providing the desired effect. This may involve monitoring the operation of the actual reallocated resources and/or the system components or processes the reallocation was intended to benefit. If actual operation of the system is not in accordance with expectations and/or simulation results, the system may be quiesced, and the original configuration returned until an analysis may be performed to determine why expected results were not achieved.
  • a trouble ticketing system or other notification system may be activated to notify the administrator of the failed reallocation.
  • the level of performance may be examined to determine if the reallocation is having the expect level of desired effect. For example, if a particular reallocation is operating within a certain percentage of expectations (e.g., 80%), which may be user defined, the reallocation may be considered acceptable. If not, the reallocation may be considered unacceptable, and the system configuration may be returned to its prior state (automatically or upon user approval).
  • the system may monitor or log some or all resource reallocations and subsequent associated performance changes so that the changes may be continually evaluated, used as a basis or model for future changes, and as a basis for returning some or all of the system to prior configurations. Moreover, this information may act as a template for future system provisioning and deployment and evaluating the operation of selected system software or hardware components.
  • Systems and modules described herein may comprise software, firmware, hardware, or any combination(s) of software, firmware, or hardware suitable for the purposes described herein.
  • Software and other modules may reside on servers, workstations, personal computers, computerized tablets, PDAs, and other devices suitable for the purposes described herein.
  • Software and other modules may be accessible via local memory, via a network, via a browser or other application in an ASP context or via other means suitable for the purposes described herein.
  • Data structures described herein may comprise computer files, variables, programming arrays, programming structures, or any electronic information storage schemes or methods, or any combinations thereof, suitable for the purposes described herein.
  • User interface elements described herein may comprise elements from graphical user interfaces, command line interfaces, and other interfaces suitable for the purposes described herein. Screenshots presented and described herein can be displayed differently as known in the art to input, access, change, manipulate, modify, alter, and work with information.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

In accordance with some aspects of the present invention, systems and methods are provided for dynamically and/or automatically selecting and/or modifying data path definitions that are used in performing storage operations on data. Alternate data paths may be specified or selected that use some or all resources that communicate with a particular destination to improve system reliability and performance. The system may also dynamically monitor and choose data path definitions to optimize system performance, conserve storage media and promote balanced load distribution.

Description

    RELATED APPLICATIONS
  • Any and all applications for which a foreign or domestic priority claim is identified in the Application Data Sheet, or any correction thereto, are hereby incorporated by reference into this application under 37 CFR 1.57.
  • This application is also related to the following patents and pending applications, each of which is hereby incorporated herein by reference in its entirety:
  • U.S. Pat. No. 6,418,478, titled PIPELINED HIGH SPEED DATA TRANSFER MECHANISM, issued Jul. 9, 2002;
  • U.S. Pat. No. 7,035,880 titled MODULAR BACKUP AND RETRIEVAL SYSTEM USED IN CONJUNCTION WITH A STORAGE AREA NETWORK, filed Jul. 6, 2000;
  • U.S. Pat. No. 6,542,972 titled LOGICAL VIEW AND ACCESS TO PHYSICAL STORAGE IN MODULAR DATA AND STORAGE MANAGEMENT SYSTEM;
  • U.S. patent application Ser. No. 10/658,095 titled DYNAMIC STORAGE DEVICE POOLING IN A COMPUTER SYSTEM, filed Sep. 9, 2003, now U.S. Pat. No. 7,130,970, issued Oct. 31, 2006;
  • U.S. patent application Ser. No. 10/818,749, titled SYSTEM AND METHOD FOR DYNAMICALLY PERFORMING STORAGE OPERATIONS IN A COMPUTER NETWORK, filed Apr. 5, 2004, now U.S. Pat. No. 7,246,207, issued Jul. 17, 2007;
  • U.S. patent application Ser. No. 11/120,619, titled HIERARCHICAL SYSTEMS AND METHODS FOR PROVIDING A UNIFIED VIEW OF STORAGE INFORMATION, filed May 2, 2005, now U.S. Pat. No. 7,343,453, issued Mar. 11, 2008;
  • U.S. Provisional Application No. 60/752,203, titled SYSTEMS AND METHODS FOR CLASSIFYING AND TRANSFERRING INFORMATION IN A STORAGE NETWORK, filed Dec. 19, 2005;
  • U.S. application Ser. No. 11/313,224 titled SYSTEMS AND METHODS FOR PERFORMING MULTI-PATH STORAGE OPERATIONS, filed Dec. 19, 2005, now U.S. Pat. No. 7,620,710, issued Nov. 17, 2009;
  • U.S. Provisional Application No. 60/752,196 titled SYSTEMS AND METHODS FOR MIGRATING COMPONENTS ON A HIERARCHICAL STORAGE NETWORK, filed Dec. 19, 2005;
  • U.S. Provisional Application No. 60/752,202 titled SYSTEMS AND METHODS FOR UNIFIED RECONSTRUCTION OF DATA IN A STORAGE NETWORK, filed Dec. 19, 2005;
  • U.S. Provisional Application No. 60/752,201 titled SYSTEMS AND METHODS FOR RESYNCHRONIZING STORAGE OPERATIONS, filed Dec. 19, 2005; and
  • U.S. Provisional Application Ser. No. 60/752,197 titled SYSTEMS AND METHODS FOR HIERARCHICAL CLIENT GROUP MANAGEMENT, filed Dec. 19, 2005.
  • BACKGROUND OF THE INVENTION Field of the Invention
  • The inventions disclosed herein relate generally to performing storage operations on electronic data in a computer network. More particularly, aspects of the present invention relate to data transmission schemes used during a storage operation including data pathways and other components used in the transfer of data.
  • Over time, storage of electronic data has evolved through many forms. During the early development of the computer, data storage was limited to individual computers. Electronic data was stored in the Random Access Memory (RAM) or some other storage medium such as a hard drive or tape drive that was an actual physical part of the individual computer.
  • Later, with the advent of network computing, storage of electronic data gradually migrated from individual computers to stand-alone storage devices accessible via a network. Over time, these individual network storage devices evolved into more complex systems including networks of tape drives, optical libraries, Redundant Arrays of Inexpensive Disks (RAID), CD-ROM jukeboxes, and other devices. Common architectures included drive pools, which generally are logical collections of drives with associated media groups including the tapes or other storage media used by a given drive pool.
  • Serial, parallel, Small Computer System Interface (SCSI), or other cables directly connect such stand-alone storage devices to individual computers that are part of a network of other computers such as a Local Area Network (LAN) or a Wide Area Network (WAN). Generally, each individual computer on the network controlled the storage devices that were physically attached to that computer and could also access the storage devices of the other network computers to perform backups, transaction processing, file sharing, and other storage-related operations.
  • Network Attached Storage (NAS) is another storage scheme using stand-alone storage devices in a LAN or other such network. In NAS, a storage controller computer typically controls the storage device to the exclusion of other computers on the network, but the SCSI or other cabling directly connecting that storage device to the individual controller is eliminated. Instead, storage devices are directly attached to the network itself.
  • Yet another network storage scheme is modular storage architecture which is more fully described in U.S. Pat. Nos. 7,035,880 and 6,542,268. An example of such a software application is the Galaxy™ system, by CommVault Systems of Oceanport, N.J. The Galaxy™ system is a multi-tiered storage management solution which includes, among other components, a storage manager, one or more media agents, and one or more storage devices. The storage manager directs storage operations of client data to storage devices such magnetic and optical media libraries. Media agents are storage controller computers that serve as intermediary devices managing the flow of data from client information stores to individual storage devices. Each storage device may be uniquely associated with a particular media agent and this association may be tracked by the storage manager.
  • A common feature shared by all of the above-described network architectures is the substantially static relationship between storage controller computers and storage devices. In these traditional network architectures, storage devices are generally connected, virtually or physically, to a single storage controller computer. Generally, only the storage controller computer to which a particular device is physically connected has read/write access to that device. One computer typically cannot control the drive pool and media group be that is being controlled by another. Requests to store and retrieve data from such a drive pool and media group would have to be coordinated by the controlling computer. Typically, storage media reserved or being written to by one media agent cannot be written to be another media agent. Thus, often storage media being used pursuant to one storage policy cannot be used by another storage policy and vice versa often resulting in the inefficient use of storage resources.
  • In some prior art systems, storage policies may specify alternate data paths or resources in the case device failure or an otherwise unavailable data path. However, such systems typically specify a single alternate data path. Moreover, because backup operations are traditional performed on a client by client basis, each client may store information on different media, resulting in inefficient media use. Furthermore, in many systems, failover conditions often result in the use of additional media further resulting in inefficient use of resources. In addition, alternate data paths are defined in a static fashion, and thus conventional data protection schemes are unable to adapt to changing network conditions.
  • SUMMARY OF THE INVENTION
  • In accordance with certain aspects of the present invention, systems and methods are provided for dynamically or automatically selecting and/or modifying data path definitions that are used in performing storage operations. Alternate data paths may be specified or selected that use some or all resources that communicate with a particular destination to improve system reliability and performance. The system may also dynamically monitor and choose data path definitions to optimize system performance, conserve storage media, prevent resource exhaustion and promote balanced load distribution.
  • In one illustrative embodiment, a method for configuring a storage operation system includes defining a first storage operation path to be used in performing a storage operation. The first storage operation path may specify a destination and substantially all of the resources capable of communicating with the destination. The system may define a second storage operation path used in the storage operation when the first storage path is unavailable.
  • In an alternate embodiment, a storage operation system may include a management module for controlling or coordinating a storage operation to a destination, a plurality of storage devices, and at least two storage operation paths linking a client to one or more storage devices. The first storage operation path may specify many, most or substantially all of the resources capable of communicating with the destination, while the second storage operation path may be used in the storage operation when the first storage path is unavailable.
  • In yet another embodiment, a method for consolidating storage policies within a storage operation network is provided which may include, analyzing storage operation paths, which may defined in storage policies. Determining whether any of the storage operation paths have common element points and consolidating two or more of the storage policies having at least one common element into a single storage operation policy such that the single storage operation policy supports copy operations to or with the common element point such as a common destination.
  • Another embodiment includes a system for consolidating storage policies within a storage operation network. The system may include a management module for directing a storage operation to a destination, a plurality of storage devices and a plurality of storage operation paths. The storage operation paths may be defined within a plurality of storage policies and have a series of element points defining locations or resources along the path, ending with the destination. The management module may consolidate two or more of the storage policies having at least one common element point into a single storage policy such that the single storage policy supports copy operations to the common element point.
  • One embodiment of the present invention includes a method for consolidating storage policies within a storage operation network that includes analyzing a plurality of storage operation paths that are defined in storage policies. This may involve identifying certain inefficiencies in the storage operation paths and reconfiguring the storage operation paths to improve system performance. This may further involve monitoring the storage network for the inefficiencies in the storage network subsequent to redefining the plurality of storage operation paths to determine whether the reconfiguration has achieved the desired effect.
  • Another embodiment of the present invention includes a system for consolidating storage policies within a storage operation network. The system may include a management component for controlling or coordinating a storage operation to a destination using one of a plurality of storage operation paths defined within a plurality of storage policies. The management component may identify inefficiencies in the storage operation paths and reconfigure or redefine the storage operation paths to correct or improve or the modified inefficiencies. The management component may also monitor the storage network including any reconfigurations subsequent to redefining the storage operation paths to determine whether the reconfigurations provided the desired correct or improvement. If, not, additional analysis and reconfiguration may be performed.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Aspects of the invention are illustrated in the figures of the accompanying drawings which are meant to be exemplary and not limiting, in which like references are intended to refer to like or corresponding parts, and in which:
  • FIG. 1 is a block diagram of a network architecture for a system to perform storage operations on electronic data in a computer network according to an embodiment of the invention;
  • FIG. 2 is a block diagram of an exemplary media storage device for performing storage operations on electronic data in a computer network according to an embodiment of the invention;
  • FIG. 3 is a flow chart illustrating some of the steps of a storage operation in accordance with an embodiment of the invention;
  • FIG. 4 is a flow chart illustrating some steps of assigning storage policies to system resources and evaluating existing storage policies of in accordance with an embodiment of the invention; and
  • FIG. 5 is a flow chart illustrating some of the steps of a method of dynamically analyzing and managing storage policies and data paths in accordance with an embodiment of the invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • Detailed embodiments of the present invention are disclosed herein, however, it is to be understood that the disclosed embodiments are merely exemplary of the invention, which may be embodied in various forms. Therefore, specific functional details disclosed herein shall not be interpreted as limiting, but merely as a basis for teaching one skilled in the art to employ the present invention in any specific embodiment.
  • With reference to FIGS. 1 through 5, representative embodiments of the invention are presented. Turning now to FIG. 1, a block diagram of one network architecture suitable for performing storage operations on electronic data in a computer network according to an embodiment of the invention is shown. The embodiment, as shown, may include a storage management component such as manager 100 and one or more of the following: a client 85, an information store 90, a data agent 95, a media agent 105, an index cache 110, and a storage device 115. The system and elements thereof are exemplary of a three-tier backup system such as the CommVault Galaxy™ backup system, available from CommVault Systems, Inc. of Oceanport, N.J., and further described in U.S. Pat. No. 7,035,880 which is incorporated herein by reference in its entirety.
  • A data agent 95 is generally a software module that responsible for archiving, migrating, and recovering data of a client computer 85 stored in an information store 90 or other memory location. Each client computer 85 may have one or more data agent(s) 95 and the system can support multiple client computers 85. The system may include a plurality of data agents 95 each of which is intended to backup, migrate, and recover data associated with a different application. For example, different individual data agents 95 may be designed to handle Microsoft Exchange® data, Lotus Notes® data, Microsoft Windows 2000® file system data, Microsoft Active Directory Objects® data, and other types of data known in the art.
  • In the case where a client computer 85 has two or more types of data, a dedicated data agent 95 may be used for each data type to archive, migrate, and restore the client computer 85 data. For example, to backup, migrate, and restore all of the data on a Microsoft Exchange 2000® server, the client computer 85 would use one Microsoft Exchange 2000® Mailbox data agent 95 to backup the Exchange 2000® mailboxes, one Microsoft Exchange 2000® Database data agent 95 to backup the Exchange 2000® databases, one Microsoft Exchange 2000® Public Folder data agent 95 to backup the Exchange 2000® Public Folders, and one Microsoft Windows 2000® File System data agent 95 to backup the client computer's 85 file system. These data agents 95 would be treated as four separate data agents 95 by the system even though they reside on the same client computer 85.
  • In some embodiments, however, multipurpose or generic data agents, may be used that operate on multiple data without types. For example, one data agent may operate on Microsoft Exchange 2000® Mailbox and Microsoft Windows 2000® File System data, etc.
  • Storage manager 100, in one embodiment, may be implemented as a software module or application that coordinates and controls various aspects of the system shown in FIG. 1. For example, storage manager 100 may communicate with some or all elements of the system including client computers 85, data agents 95, media agents 105, and storage devices 115, to schedule, initiate, manage and coordinate system backups, migrations, and data recoveries.
  • In one embodiment, a media agent 105 is may be implemented as a software module that conducts data, as directed by storage manager 100, between the client computer 85 and one or more storage devices 115 such as a tape library, a magnetic media storage device, an optical media storage device, or other storage device known in the art. For example, as shown in FIG. 1, storage manager 100 may direct data agents 95 to copy data from one or more clients 85 to storage device 115 through media agents 105. In some embodiments, media agent 105 communicates with and controls the storage device 115.
  • For example, media agent 105 may instruct storage device 115 to use a robotic arm or other means to load or eject a media cartridge, to archive, migrate, or restore data to or from certain media present in device 115. Media agents 105 may also communicate with the storage devices 115 via a local bus such as a SCSI adaptor, or other suitable connection means. In other implementations, storage device 115 may communicate to the data agent 105 via a Storage Area Network (“SAN”).
  • Each media agent 105 may maintain an index cache 110 which stores the index data the system generates during backup, migration, and restore storage operations as further described herein. For example, storage operations for Microsoft Exchange® data generate index data containing the location and other information such as metadata regarding the data on the storage device 145 the Exchange data is stored on.
  • Index data provides the system with an efficient mechanism for locating user files or data for recovery operations. This index data is generally stored with the data backed up to the storage device 115. The media agent 105 that controls the storage operation may also write an additional copy of the index data to its index cache 110. The data in media agent 105 and index cache 110 is thus readily available to the system for use (in storage and retrieval operations and other activities) without having to be first retrieved from a storage device 115.
  • Storage manager 100 also maintains an index cache 110. Such index data may include logical associations between components of the system, user preferences, metadata regarding application data or user preferences, management tasks, and other useful data. For example, the storage manager 100 may use its index cache 110 to track the logical associations between media agents 105 and storage devices 115.
  • Index caches 110 typically reside on their corresponding storage component's hard disk or other fixed storage device. Like any cache, the index cache 110 has finite capacity and the amount of index data that can be maintained directly corresponds to the size of that portion of the disk that is allocated to the index cache 110. In one embodiment, the system may manage the index cache 110 on a least recently used (“LRU”) basis as known in the art. When the capacity of the index cache 110 is reached, the system may overwrite those files in the index cache 110 that have been least recently used with the new index data. In some embodiments, before data in the index cache 110 is overwritten, the data may be copied to an index cache copy and stored on a storage device 115. If a recovery operation requires index data that is no longer stored in the index cache 110, such as in the case of a cache miss, the system may recover the index data from the copy stored in storage device 115.
  • In some embodiments, components of the system may reside and execute on the same computer. In alternative embodiments, a client computer 85 component such as a data agent 95, a media agent 105, or a storage manager 100 may coordinate and direct local archiving, migration, and retrieval of application functions as further described in U.S. Pat. No. 7,035,880. Thus, client computer 85 component can function independently or together with other similar client computer 85 components.
  • Turning now to FIG. 2, a block diagram of an exemplary media library storage device 120 for performing storage operations on electronic data in a computer network according to an embodiment of the invention is presented. Media library device 120 represents one specific type of storage device 115 (FIG. 1) that may be used with an implementation if the invention.
  • Media library storage device 120 may contain any suitable magnetic, optical or other storage media 145 and associated drives 125, 130, 135, and 140. Media 145 may store electronic data containing backups of application data, user preferences, metadata, system information, and other useful information known in the art. Drives 125, 130, 135 and 140 are used to store and retrieve electronic data from media 145. In one embodiment, drives 125, 130, 135 and 140 may function as a drive pool, as further described in application Ser. No. 10/658,095 which is hereby incorporated herein by reference in its entirety. A drive pool is a logical concept that associates drives and storage media with a storage policy and a source device such as a client 85. Storage policies representing storage patterns and preferences are more fully discussed in U.S. Pat. No. 6,542,972 which is hereby incorporated by reference herein in its entirety.
  • A drive pool may be identified by a set of drives within a library storage device 120 as pointed to by one or more media agents 105. For example, a drive pool known as DP1 consisting of drives 125 and 130 in library 120 known as LIB1 may be associated by a storage policy, with a first media agent 105 MA1 in an index cache 110 entry as follows: LIB1/MA1/DP1. A second drive pool consisting of drives 130, 135, and 140 within the library storage device 120 associated with the same media agent 105 may be expressed in index cache 110 as follows: LIB1/MA1/DP2.
  • As further described herein, the present invention permits logical association of drive pools associated with different media agents 105 (FIG. 1). Multiple drive pools, media agents, and other system components can be associated in a single index cache 110 entry. Thus, for example, an index cache 110 entry for a storage policy, according to an embodiment of the present invention, may combine the two previous entries instead and thus may be logically represented as: ##STR1##
  • In addition and as further described herein, media 145 may be associated by the system with drive pools or storage policies, and not necessarily with individual drives 125, 130, 135 and 140. A media group may be a collection of media 145 or other storage media assigned to a specific storage policy. The media group may dynamically point to different drive pools as further described herein, including those with different recording formats as the system may update the recording format of the media group in a media group table stored in an index cache 110.
  • Aspects of the present invention, as further described herein, permit data associated with a particular storage policy copy to be stored on and share certain media 145. Data from each storage policy copy may be appended to media 145 shared by other storage policy copies. Thus, a storage policy copy may be shared between several media agents 105 in a dynamic drive pooling environment with media 145 also being shared by the different media agents 105 and storage policies. Media 145 can be located in virtually any storage device 115 and for a given storage policy copy may be spread across multiple storage devices 115. Thus, an index cache entry may associate multiple media sets 145 with multiple media agents, storage policies, drive pools, and other system components. For example, two different media sets from the previous example of index entries might be associated in a single index cache 110 entry as follows: TABLE-US-00001 storage policy1:media agent1:drive pool1:media set1:media set2 storage policy2:media agent2:drive pool2:media set1:media set2.
  • In addition to media sets, a single index cache 110 entry may also specify and associate multiple media agents 105, storage policies, drive pools, network pathways, and other components.
  • While the embodiments described above employ the use of two drive pools and two media agents, one skilled in the art will recognize that additional media agents and logical drive pools may be implemented across the storage policies without deviating from the scope and spirit of the present invention.
  • Tuning to FIG. 3, a flow chart 300 illustrating some of the steps involved in performing storage operations on electronic data in a computer network according to an embodiment of the invention is shown. Selection of desired storage components for storage operations may be performed manually or automatically in dynamic fashion. In operation, the system may initiate a storage operation in response to a scheduled procedure or as directed by a user, system administrator, or as otherwise directed by the system (step 310).
  • For example, the system may initiate a backup operation or a restore operation at a specific time of day or in response to a certain threshold being exceeded as specified in a storage policy. The system may select a media agent 105 (FIG. 1) according to selection logic or a specified data path as further described herein (step 320). The selection logic and data paths may determined by a set of criteria defined in the storage policies or according to system configuration or operational rules or guidelines. Examples of such criteria may include, load balancing within the network, bandwidth use and efficiency, media usage, available media space, etc.
  • In one illustrative embodiment, the selection logic includes the ability to conduct a LAN-free storage operation, such as using a SAN, when it is desired to optimize storage operations via load balancing. For example, an index entry in index cache 110 may associate certain media agents 105, storage devices 115, or other components with LAN-free storage operations either via user input, network topology detection algorithms known in the art, or other methods. As another example, the system may select a free media agent 105 to optimize storage operations via load balancing when a default media agent 105 or other media agent 105 specified in a storage policy is already performing other storage operations or otherwise occupied. The system may also select an appropriate drive pool in a network storage device according to selection logic further described herein (step 330). Once the system has selected an appropriate media agent and drive pool, the storage operation is performed, using the selected storage components (step 340).
  • Another embodiment of the present invention allows storage policies to be recognized or be defined in terms of sub-clients (e.g., processes or portions of data of a volume that are mutually exclusive) and have data protection operations performed at the sub-client level. For example, a storage policy may specify a path similar to those described above for each sub-client operating on a client. In some embodiments, storage policies associated with each sub-client specify a default data path and one or more alternate data paths. These illustrative paths, in one embodiment may be expressed as follows:
  • Default: media agent1: library1
  • Alternate: media agent 2: library 1
  • Alternate data paths are desirable as they provide additional means by which a storage operation may be completed and thus improve system reliability and promote robust operation. Thus, the system may automatically select certain available alternate data paths to facilitate load balancing and failover recovery. Such alternate data paths may be specified using some or all of the additional routing resources available in the system. For example, the alternate data path above may specify as alternates some or all of the media agents in the system that are capable of communicating with library 1 and may be expressed as follows:
  • Alternate: media agent 2; media agent 3; . . . media agent n: library 1
  • Where ‘n’ is the total number of media agents specified in the alternate data path.
  • This arrangement allows the system to take advantage of other available routing resources, providing the ability to select from multiple data paths to the desired destination. Using one approach, a storage policy may specify all of the media agents 105 in the system capable of communicating with a particular destination (e.g., library 1). This provides the greatest likelihood that a storage operation will be completed, assuming that at least some storage resources are functioning or not otherwise congested.
  • Using another approach, some of the available routing resources, such as media agents 105, may be specified as alternates, providing a greater likelihood that the storage operation will be completed, rather than relying on a single alternate. Such alternate resources may be selected based on the degree of utilization, capacity, bandwidth, physical location, the desired confidence factor or other considerations and may be specified manually or assigned automatically based on data protections goals specified for the system.
  • In addition, alternate data paths may be specified in many ways to provide robust routing options. For example, alternate data paths may be specified according to user preferences. A system administrator may specify certain alternate data paths and the priority and/or order in which the data paths are to be used. Another method for providing alternate data paths may involve using the “round robin” approach in which alternate data paths are selected from a group of available data paths such that each alternate data path is selected and used before any previously used data path is selected and used again. This approach is typically useful in promoting load balancing within the system as it tends to spread out data transfer operations across available data paths in a substantially uniform fashion. Other approaches may include specifying alternate data paths to emphasize the ability to complete a storage operation in the event of a failover condition.
  • In certain embodiments, alternate data paths may be specified such that data from one client or sub-client may be routed to a particular destination through substantially every available data path that may potentially link the client or sub-client to the destination. Similarly, alternate data paths may be defined such that data is restored to particular client, sub-client or computing device from some or all storage devices within the system. This arrangement provides significant flexibility within the system for performing and completing both storage and restore operations.
  • Similarly, in some embodiments, alternate destinations may be specified and used in failover or other emergency data protection operations. For example, a storage policy may specify a data path including library 1, library 2, and others, with the provision that the specified media agents have access to each of the specified libraries. In some embodiments, media agents specified in such data paths may share an index cache.
  • One benefit of the arrangement described above is the ability to conserve media within a storage system. In some embodiments, storage policies may not have the ability to share storage media due to certain conflicts within programming logic or the need for storage policies to resolve any such conflicts in mutually exclusive manner to ensure computational integrity. Thus, storage operations performed pursuant to different storage policies are generally required to write to different media, often resulting in the inefficient use of media.
  • For example, a client may communicate to a storage device 115 through a first media agent 105 pursuant to a first storage policy and a second client may communicate to the same storage device pursuant to a second storage policy and a second media agent. In this case, each communication or storage operation by each media agent may be written to different media in the storage device due to programming constraints. Moreover, when a failover condition occurs, further communications to the storage device may be written to a third media based on the alternate data path definitions, resulting in an even higher media usage rate.
  • An aspect of the present invention streamlines this process by specifying data paths on a sub-client basis and creating a complimentary storage policy based on this information to avoid the logical conflict described above, or any other logical conflict that may exist. Moreover, this arrangement allows multiple clients (and associated sub-clients) to use the same storage policy, significantly reducing the number of storage policies required to manage the system as well as simplifying the process involved in updating or changing the policies themselves. This also facilitates updating and/or changing the client associations with storage policies that control or otherwise specify particulars involved in data movement.
  • Additionally, two groups of clients may specify two sets of client or sub-client data paths (e.g., a default and alternate for each), but, may be governed by a single storage policy in accordance with one embodiment of the present invention. This may be accomplished by examining the data paths and combining or rearranging them into a suitable form for use in the storage policy. For example, a first group of sub-clients may specify the following data paths:
  • Default data path: Media agent 1: library 1
  • Alternate data path: Media agent 2: library 1
  • The second group of sub-clients may specify the following data paths:
  • Default data path: Media agent 2: library 1
  • Alternate data path: Media agent 1: library 1
  • These may be examined and modified (or combined and rearranged) to specify or point to a single storage policy with data paths as expressed below which takes into account the data path preferences of each sub-client while eliminating the need for two separate storage policies:
  • Default data path: Media agent 1: library 1
  • Alternate data path: Media agent 2: library 1
  • In operation, the system may consult this modified storage policy (default first and alternate second) to obtain data path preferences when moving data from the first set of clients. When moving data from the second set of clients, this storage policy may be consulted in reverse order, thus preserving the original preferences (i.e., Media agent 2 as the default with Media agent 1 as the alternate). Using this single storage policy arrangement, data from various storage operations may be written to the same media, rather than using separate media as explained above, promoting media conservation. Moreover, specifying data paths on a sub-client level allows multiple clients to write data to the same media and avoids the potential logical conflicts described above.
  • Another embodiment in accordance with the present invention includes the case where two client domains separated by a firewall, each domain containing multiple sub-clients. Assume, for example that each domain has set of sub-clients with different data paths as shown below:
  • Domain 1:
  • Default data path: Media agent 1: library 1
  • Alternate data path: Media agent 2: library 1
  • Domain 2:
  • Default data path: Media agent 3: library 1
  • Alternate data path: Media agent 4: library 1
  • As in the example above, media agent utilization will increase if two storage policies are used to manage this arrangement. Thus, in accordance with an embodiment of the present invention, these data paths may be modified (or combined) into one storage policy set forth below having four specified data paths rather than two storage polices with two data paths each, thus maintaining failover protection and promote minimum media utilization:
  • Media agent 1: library 1; Media agent 2: library 1
  • Media agent 3: library 1; Media agent 4: library 1
  • When moving data from the first domain, the first entry is consulted and vice versa for the second domain which allows information from both domains to be written to the same media, promoting efficient media utilization. This order of operations may be defined within the storage policy or may be specified by placing the appropriate pointers or other referential elements in an index or other entry that governs data path preferences.
  • Another benefit of the present invention includes the ability to use a single storage policy to govern multiple clients. This provides users with significant flexibility by allowing them to define a storage policy and “point” to that policy through referential elements to multiple clients, thereby simplifying system administration.
  • For example, in the case where one or more clients need to have changes or modifications made to an associated storage policy, with the provided arrangement, a single policy may be changed having a global effect rather than requiring a similar change be made to multiple individual polices. Moreover, clients may easily be assigned or moved from one storage policy to another merely by changing a pointer or other referential element. This eliminates having to copy, significantly modify, change or create a new storage policy from scratch. Further, storage policies are no longer defined and associated on an individual client by client basis.
  • In one embodiment, storage policies and associated storage domains may be associated with one another based on system configuration, user needs, or other considerations. This process may be performed either manually, automatically, or may be partially automated, requiring certain user input such as customization information, intended or expected use, etc. For example, at system setup a configuration program may walk an administrator through a configuration program and prompt the user for certain customization information. In alternative embodiments, this process may be predominantly or completely automated based on certain specific goals including, but not limited to, efficient media usage, degree of desired data protection, and substantially even and/or efficient load distribution.
  • Flowchart 400 of FIG. 4 illustrates some of the steps involved in assigning storage policies to system resources or in evaluating existing storage polices for possible consolidation as part of an ongoing effort to analyze and increase system efficiency.
  • As shown, at step 410 any existing storage polices or defined data paths for performing storage operations may be retrieved, examined and analyzed. This may involve, for example, retrieving path information from an index cache associated with a media agent or master storage manager or retrieving similar information from a metabase that may be associated with such components. The analysis may include examining data path information such as origination point (e.g., clients and/or sub-clients), destination point (storage device, library, media pool, etc.), transmission resources scheduled to be involved including media agents, data conduits and other transmission elements. In some embodiments, this may involve the creation of a system wide or more limited process-based netlist to obtain a basic understanding of system routing options and transmission patterns and preferences.
  • At step 420, the system may determine whether any identified clients or sub-clients have a common destination point. The destination points are typically defined as a storage device for receiving data from copy operations representing the last location of data at the completion of a particular copy operation. A list of origination points (e.g., clients and or media agents) and common destination points may be compiled as a starting point to determine similarities between various identified data paths that may be suitable for combination or rearrangement into one or more storage policy to improve overall system efficiency and/or reduce media consumption. Next, at step 430, media agents and other data transfer resources may be associated with the list to generate a more complete picture of the routes and resources involved/available in traversing the data paths between origination and destination points. At this point, the netlist may be substantially complete taking into account available routing and resource information.
  • At step 440, the system examines any pre-existing or identified storage polices and compares them with other storage polices and the information generated at step 430 to identify common elements that may be combined or folded into the existing storage policies. This may also involve identifying and comparing clients/sub-clients with common origination points and correlating them with storage polices having common destination points as a basis for potentially creating new storage polices. Other information of interest may include identifying common media agents and associated destination points, etc. for similar reasons.
  • Next, at step 450, it may be determined whether any identified common elements are precluded from writing information to the same storage device and/or storage media. If so, in some embodiments, these elements may be noted on a list of items not suitable for combination to into storage polices and may be identified as needing individual treatment. The gathered information may be analyzed to determine if the number of storage polices may be reduced by combining common elements, by combining or modifying existing storage polices, or by recasting storage polices with other identified data paths into more efficient storage pathways (step 460).
  • For example, the analysis may reveal four existing storage polices that have many common elements. Depending on the management goals of the storage system, these four storage polices may be combined into one comprehensive storage policy with a common destination if minimal media usage is desired or may be combined into two storage polices to minimize the possibility of alternate data path congestion.
  • Other analysis results may reveal several sub-client data path definitions that can be combined into a new storage policy to reduce media usage without substantially affecting storage device accessibility. Moreover, although some results may suggest the combination of significant numbers of storage policies or other common data paths, such suggestions may be examined to determine whether overall system performance would be adversely impacted, for example, beyond a preset performance threshold, and if so, may not be implemented even though such combinations may reduce overall media consumption.
  • In some embodiments of the invention, factors other than media consumption or possible congestion may be taken into account when determining how to create, change or modify storage polices to accommodate certain system management goals. Such considerations may include load balancing, optimization, service level performance or other operational goals including adjustments to account for changes that may occur over time.
  • A system administrator, for example, may wish to maintain a substantially even workload across the storage network and maintain that distribution on a going forward basis. Other goals may include maintaining operational performance within a certain percentage level to ensure a specified level of data protection or maximizing system efficiency during peak usage periods. Achieving these and other goals may involve the dynamic and periodic redefinition of data paths and associated storage policies
  • Turning now to FIG. 5, a flow chart 500 illustrating some of the steps involved with the dynamic analysis and potential redefinition of storage policies/selection of alternate data paths in accordance with aspects of the present invention is illustrated. At step 510, client/sub-client data paths within the system are analyzed similarly to step 410 described above in connection with FIG. 4. Next, at step 520, with the data paths identified, certain system performance and forecasting reports may be run as described in co-pending, commonly assigned cases entitled Systems and Methods for Allocation of Organizational Resources Application, and Hierarchical Systems and Methods for Providing a Unified View of Storage Information, Ser. No. 11/120,619, filed May 2, 2005, which are hereby incorporated by reference in its entirety. Such reports may forecast, based on past performance or other parameters, how resource utilization may grow or otherwise change and predict how capacity, efficiency, failure rates, and traffic load may impact storage operations over time.
  • Based on the forecasting information, the system analyzes, at step 530, data paths to identify which ones are susceptible to or likely to experience an adverse impact due to the changing conditions (e.g., based on predefined thresholds or resource capacity). Such data path definitions or storage policies may then be modified on a dynamic basis to accommodate or otherwise account for predicted conditions to minimize impact (step 540). For example, if it is determined that certain data paths are expected to become congested after a certain period of time, additional alternate data paths expected to handle the additional load may be added before that point is reached or other alternate data paths that do not suffer from the same conditions are specified.
  • If certain media agents that serve particular storage devices are expected to become overloaded or constantly busy at or near capacity, additional alternate media agents may be added by combining or otherwise altering storage policy data paths to help reduce the adverse impact of the anticipated problem (e.g., other alternate data paths not suffering from the same or similar conditions may be selected or added to help alleviate any detected or predicted problem).
  • Moreover, data paths may be changed on a dynamic basis to balance load, maintain a substantially constant data load, or prevent a failover condition in accordance with user specifications or system requirements. In certain embodiments this may involve distributing work load across several communication paths as described in commonly assigned, co-pending case entitled Systems and Methods for Providing Multipath Storage Network, filed on Dec. 19, 2005, and which is hereby incorporated by reference in its entirety.
  • Next at step 550, resource reallocation is considered if data path adjustment is not sufficient to correct or acceptably minimize any anticipated problem. This may involve, for example, allocating additional storage resources such as media agents, data paths, and storage devices, etc. from other storage operation cells, as described in commonly assigned, co-pending case entitled Systems and Methods for Migrating Components on a Hierarchical Storage Network, application Ser. No. 60/752,196, filed on Dec. 19, 2005, and which is hereby incorporated by reference in its entirety. If deemed helpful, the reallocation is performed at step 560 as described in that case. The system may then periodically return to step 510 and perform the process as part of an ongoing recursive effort to maintain or optimize system performance.
  • In some embodiments, prior to actual reallocation of resources, proposed reallocation scenarios may be simulated and evaluated with the expected results extended over time in order to choose the best solution to any resource shortcoming that best fits enterprise needs or user expectations. Moreover, in some embodiments, any such resource reallocation may need to be approved by an administrator prior to reallocation, which may involve reviewing simulation results and approving reallocations on a component by component or proposal by proposal basis. However, in other embodiments, such resource reallocation may be performed substantially automatically.
  • Next at step 570, the system monitors performance subsequent to resource reallocation to help confirm the reallocation is providing the desired effect. This may involve monitoring the operation of the actual reallocated resources and/or the system components or processes the reallocation was intended to benefit. If actual operation of the system is not in accordance with expectations and/or simulation results, the system may be quiesced, and the original configuration returned until an analysis may be performed to determine why expected results were not achieved.
  • In one embodiment, a trouble ticketing system or other notification system, as is known in the art, may be activated to notify the administrator of the failed reallocation. Moreover, in some embodiments, the level of performance may be examined to determine if the reallocation is having the expect level of desired effect. For example, if a particular reallocation is operating within a certain percentage of expectations (e.g., 80%), which may be user defined, the reallocation may be considered acceptable. If not, the reallocation may be considered unacceptable, and the system configuration may be returned to its prior state (automatically or upon user approval).
  • In some embodiments, the system may monitor or log some or all resource reallocations and subsequent associated performance changes so that the changes may be continually evaluated, used as a basis or model for future changes, and as a basis for returning some or all of the system to prior configurations. Moreover, this information may act as a template for future system provisioning and deployment and evaluating the operation of selected system software or hardware components.
  • Systems and modules described herein may comprise software, firmware, hardware, or any combination(s) of software, firmware, or hardware suitable for the purposes described herein. Software and other modules may reside on servers, workstations, personal computers, computerized tablets, PDAs, and other devices suitable for the purposes described herein. Software and other modules may be accessible via local memory, via a network, via a browser or other application in an ASP context or via other means suitable for the purposes described herein. Data structures described herein may comprise computer files, variables, programming arrays, programming structures, or any electronic information storage schemes or methods, or any combinations thereof, suitable for the purposes described herein. User interface elements described herein may comprise elements from graphical user interfaces, command line interfaces, and other interfaces suitable for the purposes described herein. Screenshots presented and described herein can be displayed differently as known in the art to input, access, change, manipulate, modify, alter, and work with information.
  • While the invention has been described and illustrated in connection with preferred embodiments, many variations and modifications as will be evident to those skilled in this art may be made without departing from the spirit and scope of the invention, and the invention is thus not to be limited to the precise details of methodology or construction set forth above as such variations and modification are intended to be included within the scope of the invention.

Claims (21)

1. (canceled)
2. A method for consolidating storage policies within a storage operation network, the method comprising:
automatically evaluating with one or more computer hardware processors, first and second storage policies to determine that the first and second storage policies use first and second storage operation paths to conduct data from a first client to a first storage device;
consolidating the first and second storage policies into a comprehensive storage policy, wherein the comprehensive storage policy associates at least the first and second storage operation paths to conduct data from the first client to the first storage device;
generating a forecast of at least one prediction of how one or more future network operating conditions may impact future storage operations; and
automatically adding at least a third storage operation path to the comprehensive storage policy based on the at least one prediction of how the one or more future network operating conditions may impact future storage operations, wherein the third storage operation path is an alternate data path that is different than the first and second storage operation paths.
3. The method of claim 2 wherein a first media agent transfers the data via the first storage operation path and a second media agent transfers the data via the second storage operation path.
4. The method of claim 2 wherein the first client comprises at least one sub client.
5. The method of claim 2 wherein the one or more future network operating conditions comprise at least one of the group consisting of: data transfer rate, network usage, load balancing, resource exhaustion, transmission congestion, or performance optimization.
6. The method of claim 2 wherein automatically evaluating the first and second storage policies is based at least in part whether the first and second storage policies share a common element.
7. The method of claim 2 wherein information about the first and second storage operation paths is obtained from an index cache.
8. The method of claim 2 wherein information about the first and second storage operation paths is obtained from a metabase.
9. The method of claim 2 wherein automatically evaluating the first and second storage policies is based at least in part on one of the group consisting of: origination point, destination point, transmission resources scheduled to be involved, and a process-based netlist.
10. The method of claim 2 wherein automatically evaluating the first and second storage policies is based at least in part on which storage operation paths are likely to experience an adverse impact due to changing conditions.
11. The method of claim 2 wherein dynamically adding the third storage operation path is based at least in part on preventing a predicted failover condition.
12. A storage operation system comprising:
a plurality of storage devices; and
a storage manager executing in one or more computer processors, the storage manager configured to:
automatically evaluate first and second storage policies to determine that first and second storage policies use first and second storage operation paths to conduct data from a first client to a first storage device;
consolidate the first and second storage policies into a comprehensive storage policy, wherein the comprehensive storage policy associates at least the first and second storage operation paths to conduct data from the first client to the first storage device; and
generate a forecast of at least one prediction of how one or more future network operating conditions may impact future storage operations; and
automatically add at least a third storage operation path to the comprehensive storage policy based on the at least one prediction of how the one or more future network operating conditions may impact future storage operations, wherein the third storage operation path is an alternate data path that is different than the first and second storage operation paths.
13. The storage operation system of claim 12 wherein the storage manager is configured to direct a first media agent to transfer the data via the first storage operation path and a second media agent to transfer the data via the second storage operation path.
14. The storage operation system of claim 12 wherein the first client comprises at least one sub client.
15. The storage operation system of claim 12 wherein the one or more future network operating conditions comprise at least one of the group consisting of: data transfer rate, network usage, load balancing, resource exhaustion, transmission congestion, or performance optimization.
16. The storage operation system of claim 12 wherein the storage manager is configured to automatically evaluate the first and second storage policies based at least in part on whether the first and second storage policies share a common element.
17. The storage operation system of claim 12 wherein the storage manager is configured to obtain information about the first and second storage operation paths from an index cache.
18. The storage operation system of claim 12 wherein the storage manager is configured to obtain information about the first and second storage operation paths from a metabase.
19. The storage operation system of claim 12 wherein the storage manager is configured to automatically evaluate the first and second storage policies based at least in part one of the group consisting of: origination point, destination point, transmission resources scheduled to be involved, and a process-based netlist.
20. The storage operation system of claim 12 wherein the storage manager is configured to automatically evaluate the first and second storage policies based at least in part on which storage operation paths are likely to experience an adverse impact due to changing conditions.
21. The storage operation system of claim 12 wherein the storage manager is configured to dynamically add the third storage operation path based at least in part on preventing a predicted failover condition.
US16/800,284 2005-12-19 2020-02-25 Systems and methods for granular resource management in a storage network Abandoned US20200267219A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/800,284 US20200267219A1 (en) 2005-12-19 2020-02-25 Systems and methods for granular resource management in a storage network

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US75219805P 2005-12-19 2005-12-19
US11/640,144 US8572330B2 (en) 2005-12-19 2006-12-15 Systems and methods for granular resource management in a storage network
US14/018,014 US9313143B2 (en) 2005-12-19 2013-09-04 Systems and methods for granular resource management in a storage network
US15/080,186 US9930118B2 (en) 2005-12-19 2016-03-24 Systems and methods for granular resource management in a storage network
US15/897,939 US20180278689A1 (en) 2005-12-19 2018-02-15 Systems and methods for granular resource management in a storage network
US201916570302A 2019-09-13 2019-09-13
US16/800,284 US20200267219A1 (en) 2005-12-19 2020-02-25 Systems and methods for granular resource management in a storage network

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US201916570302A Continuation 2005-12-19 2019-09-13

Publications (1)

Publication Number Publication Date
US20200267219A1 true US20200267219A1 (en) 2020-08-20

Family

ID=38429716

Family Applications (5)

Application Number Title Priority Date Filing Date
US11/640,144 Expired - Fee Related US8572330B2 (en) 2005-12-19 2006-12-15 Systems and methods for granular resource management in a storage network
US14/018,014 Expired - Fee Related US9313143B2 (en) 2005-12-19 2013-09-04 Systems and methods for granular resource management in a storage network
US15/080,186 Active 2027-02-17 US9930118B2 (en) 2005-12-19 2016-03-24 Systems and methods for granular resource management in a storage network
US15/897,939 Abandoned US20180278689A1 (en) 2005-12-19 2018-02-15 Systems and methods for granular resource management in a storage network
US16/800,284 Abandoned US20200267219A1 (en) 2005-12-19 2020-02-25 Systems and methods for granular resource management in a storage network

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US11/640,144 Expired - Fee Related US8572330B2 (en) 2005-12-19 2006-12-15 Systems and methods for granular resource management in a storage network
US14/018,014 Expired - Fee Related US9313143B2 (en) 2005-12-19 2013-09-04 Systems and methods for granular resource management in a storage network
US15/080,186 Active 2027-02-17 US9930118B2 (en) 2005-12-19 2016-03-24 Systems and methods for granular resource management in a storage network
US15/897,939 Abandoned US20180278689A1 (en) 2005-12-19 2018-02-15 Systems and methods for granular resource management in a storage network

Country Status (1)

Country Link
US (5) US8572330B2 (en)

Families Citing this family (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8346733B2 (en) 2006-12-22 2013-01-01 Commvault Systems, Inc. Systems and methods of media management, such as management of media to and from a media storage library
US7603518B2 (en) 2005-12-19 2009-10-13 Commvault Systems, Inc. System and method for improved media identification in a storage device
US7961594B2 (en) * 2002-10-23 2011-06-14 Onaro, Inc. Methods and systems for history analysis for access paths in networks
US7174433B2 (en) 2003-04-03 2007-02-06 Commvault Systems, Inc. System and method for dynamically sharing media in a computer network
WO2004090789A2 (en) 2003-04-03 2004-10-21 Commvault Systems, Inc. System and method for extended media retention
US8560671B1 (en) * 2003-10-23 2013-10-15 Netapp, Inc. Systems and methods for path-based management of virtual servers in storage network environments
US7472238B1 (en) 2004-11-05 2008-12-30 Commvault Systems, Inc. Systems and methods for recovering electronic information from a storage medium
WO2007038617A2 (en) * 2005-09-27 2007-04-05 Onaro Methods and systems for validating accessibility and currency of replicated data
US8572330B2 (en) 2005-12-19 2013-10-29 Commvault Systems, Inc. Systems and methods for granular resource management in a storage network
WO2007092615A2 (en) * 2006-02-09 2007-08-16 Monosphere Inc. Storage capacity planning
US10180809B2 (en) * 2006-05-17 2019-01-15 Richard Fetik Secure application acceleration system, methods and apparatus
US20080065749A1 (en) * 2006-09-08 2008-03-13 Simge Kucukyavuz System and method for connectivity between hosts and devices
US7539783B2 (en) 2006-09-22 2009-05-26 Commvault Systems, Inc. Systems and methods of media management, such as management of media to and from a media storage library, including removable media
US20080147878A1 (en) * 2006-12-15 2008-06-19 Rajiv Kottomtharayil System and methods for granular resource management in a storage network
US7831566B2 (en) 2006-12-22 2010-11-09 Commvault Systems, Inc. Systems and methods of hierarchical storage management, such as global management of storage operations
US8826032B1 (en) 2006-12-27 2014-09-02 Netapp, Inc. Systems and methods for network change discovery and host name resolution in storage network environments
US8332860B1 (en) 2006-12-30 2012-12-11 Netapp, Inc. Systems and methods for path-based tier-aware dynamic capacity management in storage network environments
US20080221974A1 (en) * 2007-02-22 2008-09-11 Alexander Gilgur Lazy Evaluation of Bulk Forecasts
US9042263B1 (en) * 2007-04-06 2015-05-26 Netapp, Inc. Systems and methods for comparative load analysis in storage networks
US8706976B2 (en) 2007-08-30 2014-04-22 Commvault Systems, Inc. Parallel access virtual tape library and drives
JP5026212B2 (en) * 2007-09-28 2012-09-12 株式会社日立製作所 Computer system, management apparatus and management method
US8195620B2 (en) * 2007-10-19 2012-06-05 International Business Machines Corporation Storage system with improved multiple copy targeting
EP2248003A1 (en) * 2007-12-31 2010-11-10 Netapp, Inc. System and method for automatic storage load balancing in virtual server environments
US20100070466A1 (en) 2008-09-15 2010-03-18 Anand Prahlad Data transfer techniques within data storage devices, such as network attached storage performing data migration
US9244779B2 (en) 2010-09-30 2016-01-26 Commvault Systems, Inc. Data recovery operations, such as recovery from modified network data management protocol data
US9477597B2 (en) 2011-03-25 2016-10-25 Nvidia Corporation Techniques for different memory depths on different partitions
US8701057B2 (en) 2011-04-11 2014-04-15 Nvidia Corporation Design, layout, and manufacturing techniques for multivariant integrated circuits
US9529712B2 (en) * 2011-07-26 2016-12-27 Nvidia Corporation Techniques for balancing accesses to memory having different memory types
US10459924B2 (en) * 2011-09-26 2019-10-29 Nec Corporation Information processing system, information processing method, information processing device and communication terminal, and method and program for controlling same
AU2013202553B2 (en) 2012-03-30 2015-10-01 Commvault Systems, Inc. Information management of mobile device data
GB2504719A (en) 2012-08-07 2014-02-12 Ibm Grid based data mobility
US9003086B1 (en) 2012-10-27 2015-04-07 Twitter, Inc. Dynamic distribution of replicated data
WO2014099682A1 (en) * 2012-12-20 2014-06-26 Amazon Technologies, Inc. Virtual tape library system
US10013166B2 (en) 2012-12-20 2018-07-03 Amazon Technologies, Inc. Virtual tape library system
US9069799B2 (en) 2012-12-27 2015-06-30 Commvault Systems, Inc. Restoration of centralized data storage manager, such as data storage manager in a hierarchical data storage system
US11238009B1 (en) * 2013-03-13 2022-02-01 Veritas Technologies Llc Techniques for automated policy analysis
CA2973866C (en) * 2013-04-19 2019-10-01 James Carey Video identification and analytical recognition system
US9483213B1 (en) 2013-04-29 2016-11-01 Amazon Technologies, Inc. Virtual media changers
US20160041996A1 (en) 2014-08-11 2016-02-11 Netapp, Inc. System and method for developing and implementing a migration plan for migrating a file system
US10860529B2 (en) 2014-08-11 2020-12-08 Netapp Inc. System and method for planning and configuring a file system migration
US9998531B2 (en) * 2013-09-18 2018-06-12 International Business Machines Corporation Computer-based, balanced provisioning and optimization of data transfer resources for products and services
US9934099B1 (en) * 2014-05-29 2018-04-03 EMC IP Holding Company LLC Dynamically pausing large backups
US10552874B2 (en) * 2014-12-18 2020-02-04 Facebook, Inc. Prompting a user to purchase items for use in an application in a feed of content provided by an online system
US9928144B2 (en) 2015-03-30 2018-03-27 Commvault Systems, Inc. Storage management of data using an open-archive architecture, including streamlined access to primary data originally stored on network-attached storage and archived to secondary storage
US10404568B2 (en) * 2015-07-31 2019-09-03 Cisco Technology, Inc. Agent manager for distributed transaction monitoring system
US10101913B2 (en) 2015-09-02 2018-10-16 Commvault Systems, Inc. Migrating data to disk without interrupting running backup operations
US20180041426A1 (en) * 2016-08-08 2018-02-08 Layer3 TV, Inc. Dynamic route selection for routing data in a content delivery system
US10078552B2 (en) 2016-12-29 2018-09-18 Western Digital Technologies, Inc. Hierarchic storage policy for distributed object storage systems
US10742735B2 (en) 2017-12-12 2020-08-11 Commvault Systems, Inc. Enhanced network attached storage (NAS) services interfacing to cloud storage
KR102079439B1 (en) * 2018-02-22 2020-02-19 주식회사 에이티센스 Ecg measurement system having patch-type ecg measurement apparatus
US10678439B2 (en) * 2018-04-02 2020-06-09 Micron Technology, Inc. Optimization of memory systems based on performance goals
US10963172B2 (en) * 2018-08-09 2021-03-30 Apple Inc. Systems and methods for providing a back pressure free interconnect
US11599283B2 (en) 2019-10-29 2023-03-07 Western Digital Technologies, Inc. Power reduction in distributed storage systems
EP4172776B1 (en) * 2020-07-23 2024-03-20 Huawei Technologies Co., Ltd. Device and method for merging backup policies
US11593223B1 (en) 2021-09-02 2023-02-28 Commvault Systems, Inc. Using resource pool administrative entities in a data storage management system to provide shared infrastructure to tenants
US12026382B2 (en) 2021-10-29 2024-07-02 Pure Storage, Inc. Storage path routing in a container system

Family Cites Families (229)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4267568A (en) 1975-12-03 1981-05-12 System Development Corporation Information storage and retrieval system
US4084231A (en) 1975-12-18 1978-04-11 International Business Machines Corporation System for facilitating the copying back of data in disc and tape units of a memory hierarchial system
GB2035014B (en) 1978-11-06 1982-09-29 British Broadcasting Corp Cyclic redundancy data check encoding method and apparatus
US4417321A (en) 1981-05-18 1983-11-22 International Business Machines Corp. Qualifying and sorting file record data
US4641274A (en) 1982-12-03 1987-02-03 International Business Machines Corporation Method for communicating changes made to text form a text processor to a remote host
EP0128945B1 (en) 1982-12-09 1991-01-30 Sequoia Systems, Inc. Memory backup system
US4686620A (en) 1984-07-26 1987-08-11 American Telephone And Telegraph Company, At&T Bell Laboratories Database backup method
GB8622010D0 (en) 1986-09-12 1986-10-22 Hewlett Packard Ltd File backup facility
US5193154A (en) 1987-07-10 1993-03-09 Hitachi, Ltd. Buffered peripheral system and method for backing up and retrieving data to and from backup memory device
US5005122A (en) 1987-09-08 1991-04-02 Digital Equipment Corporation Arrangement with cooperating management server node and network service node
JPH0743676B2 (en) 1988-03-11 1995-05-15 株式会社日立製作所 Back-up data dump control method and device
US4912637A (en) 1988-04-26 1990-03-27 Tandem Computers Incorporated Version management tool
US4995035A (en) 1988-10-31 1991-02-19 International Business Machines Corporation Centralized management in a computer network
US5093912A (en) 1989-06-26 1992-03-03 International Business Machines Corporation Dynamic resource pool expansion and contraction in multiprocessing environments
EP0405926B1 (en) 1989-06-30 1996-12-04 Digital Equipment Corporation Method and apparatus for managing a shadow set of storage media
US5454099A (en) 1989-07-25 1995-09-26 International Business Machines Corporation CPU implemented method for backing up modified data sets in non-volatile store for recovery in the event of CPU failure
US5133065A (en) 1989-07-27 1992-07-21 Personal Computer Peripherals Corporation Backup computer program for networks
US5321816A (en) 1989-10-10 1994-06-14 Unisys Corporation Local-remote apparatus with specialized image storage modules
US5504873A (en) 1989-11-01 1996-04-02 E-Systems, Inc. Mass data storage and retrieval system
US5276860A (en) 1989-12-19 1994-01-04 Epoch Systems, Inc. Digital data processor with improved backup storage
US5276867A (en) 1989-12-19 1994-01-04 Epoch Systems, Inc. Digital data storage system with improved data migration
JPH0410041A (en) 1990-04-27 1992-01-14 Toshiba Corp Data saving system
GB2246218B (en) 1990-07-18 1994-02-09 Stc Plc Distributed data processing systems
US5239647A (en) 1990-09-07 1993-08-24 International Business Machines Corporation Data storage hierarchy with shared storage level
US5544347A (en) 1990-09-24 1996-08-06 Emc Corporation Data storage system controlled remote data mirroring with respectively maintained data indices
US5212772A (en) 1991-02-11 1993-05-18 Gigatrend Incorporated System for storing data in backup tape device
US5287500A (en) 1991-06-03 1994-02-15 Digital Equipment Corporation System for allocating storage spaces based upon required and optional service attributes having assigned piorities
US5333315A (en) 1991-06-27 1994-07-26 Digital Equipment Corporation System of device independent file directories using a tag between the directories and file descriptors that migrate with the files
US5347653A (en) 1991-06-28 1994-09-13 Digital Equipment Corporation System for reconstructing prior versions of indexes using records indicating changes between successive versions of the indexes
US5410700A (en) 1991-09-04 1995-04-25 International Business Machines Corporation Computer system which supports asynchronous commitment of data
EP0541281B1 (en) 1991-11-04 1998-04-29 Commvault Systems, Inc. Incremental-computer-file backup using signatures
US5263154A (en) 1992-04-20 1993-11-16 International Business Machines Corporation Method and system for incremental time zero backup copying of data
US5241668A (en) 1992-04-20 1993-08-31 International Business Machines Corporation Method and system for automated termination and resumption in a time zero backup copy process
US5241670A (en) 1992-04-20 1993-08-31 International Business Machines Corporation Method and system for automated backup copy ordering in a time zero backup copy session
EP0680634B1 (en) 1993-01-21 1997-05-14 Apple Computer, Inc. Apparatus and method for backing up data from networked computer storage devices
EP0681721B1 (en) 1993-02-01 2005-03-23 Sun Microsystems, Inc. Archiving file system for data servers in a distributed network environment
JPH0721135A (en) 1993-07-02 1995-01-24 Fujitsu Ltd Data processing system with duplex monitor function
US5642496A (en) 1993-09-23 1997-06-24 Kanfi; Arnon Method of making a backup copy of a memory over a plurality of copying sessions
US5544345A (en) 1993-11-08 1996-08-06 International Business Machines Corporation Coherence controls for store-multiple shared data coordinated by cache directory entries in a shared electronic storage
EP0728333A1 (en) 1993-11-09 1996-08-28 Arcada Software Data backup and restore system for a computer network
US5495607A (en) 1993-11-15 1996-02-27 Conner Peripherals, Inc. Network management system having virtual catalog overview of files distributively stored across network domain
US5491810A (en) 1994-03-01 1996-02-13 International Business Machines Corporation Method and system for automated data storage system space allocation utilizing prioritized data set parameters
US5673381A (en) 1994-05-27 1997-09-30 Cheyenne Software International Sales Corp. System and parallel streaming and data stripping to back-up a network
US5638509A (en) 1994-06-10 1997-06-10 Exabyte Corporation Data storage and protection system
US5574906A (en) 1994-10-24 1996-11-12 International Business Machines Corporation System and method for reducing storage requirement in backup subsystems utilizing segmented compression and differencing
US6272465B1 (en) 1994-11-02 2001-08-07 Legerity, Inc. Monolithic PC audio circuit
US5559957A (en) 1995-05-31 1996-09-24 Lucent Technologies Inc. File system for a data storage device having a power fail recovery mechanism for write/replace operations
US5699361A (en) 1995-07-18 1997-12-16 Industrial Technology Research Institute Multimedia channel formulation mechanism
US5813009A (en) 1995-07-28 1998-09-22 Univirtual Corp. Computer based records management system method
US5619644A (en) 1995-09-18 1997-04-08 International Business Machines Corporation Software directed microcode state save for distributed storage controller
US5819020A (en) 1995-10-16 1998-10-06 Network Specialists, Inc. Real time backup system
US5778395A (en) 1995-10-23 1998-07-07 Stac, Inc. System for backing up files from disk volumes on multiple nodes of a computer network
US5729743A (en) 1995-11-17 1998-03-17 Deltatech Research, Inc. Computer apparatus and method for merging system deltas
US5761677A (en) 1996-01-03 1998-06-02 Sun Microsystems, Inc. Computer system method and apparatus providing for various versions of a file without requiring data copy or log operations
KR970076238A (en) 1996-05-23 1997-12-12 포만 제프리 엘 Servers, methods and program products thereof for creating and managing multiple copies of client data files
US5889935A (en) 1996-05-28 1999-03-30 Emc Corporation Disaster control features for remote data mirroring
US5812398A (en) 1996-06-10 1998-09-22 Sun Microsystems, Inc. Method and system for escrowed backup of hotelled world wide web sites
GB9612214D0 (en) * 1996-06-11 1996-08-14 Mallinckrodt Veterinary Inc Vaccine
US6886035B2 (en) * 1996-08-02 2005-04-26 Hewlett-Packard Development Company, L.P. Dynamic load balancing of a network of client and server computer
US5758359A (en) 1996-10-24 1998-05-26 Digital Equipment Corporation Method and apparatus for performing retroactive backups in a computer system
US5875478A (en) 1996-12-03 1999-02-23 Emc Corporation Computer backup using a file system, network, disk, tape and remote archiving repository media system
US6131095A (en) 1996-12-11 2000-10-10 Hewlett-Packard Company Method of accessing a target entity over a communications network
AU5929398A (en) 1997-01-23 1998-08-18 Overland Data, Inc. Virtual media library
US6658526B2 (en) 1997-03-12 2003-12-02 Storage Technology Corporation Network attached virtual data storage subsystem
US5924102A (en) 1997-05-07 1999-07-13 International Business Machines Corporation System and method for managing critical files
US6094416A (en) 1997-05-09 2000-07-25 I/O Control Corporation Multi-tier architecture for control network
US5887134A (en) 1997-06-30 1999-03-23 Sun Microsystems System and method for preserving message order while employing both programmed I/O and DMA operations
DE69802294T2 (en) 1997-08-29 2002-05-16 Hewlett-Packard Company, Palo Alto SYSTEMS FOR DATA BACKUP AND RECOVERY
EP0899662A1 (en) 1997-08-29 1999-03-03 Hewlett-Packard Company Backup and restore system for a computer network
US5950205A (en) 1997-09-25 1999-09-07 Cisco Technology, Inc. Data transmission over the internet using a cache memory file system
US6275953B1 (en) 1997-09-26 2001-08-14 Emc Corporation Recovery from failure of a data processor in a network server
US6052735A (en) 1997-10-24 2000-04-18 Microsoft Corporation Electronic mail object synchronization between a desktop computer and mobile device
US6021415A (en) 1997-10-29 2000-02-01 International Business Machines Corporation Storage management system with file aggregation and space reclamation within aggregated files
US6418478B1 (en) * 1997-10-30 2002-07-09 Commvault Systems, Inc. Pipelined high speed data transfer mechanism
US6260068B1 (en) 1998-06-10 2001-07-10 Compaq Computer Corporation Method and apparatus for migrating resources in a multi-processor computer system
JPH11143754A (en) 1997-11-05 1999-05-28 Hitachi Ltd Version information and constitution information display method and device therefor, and computer readable recording medium for recording version information and constitution information display program
JP3665460B2 (en) 1997-12-05 2005-06-29 富士通株式会社 Route selection system, method, and recording medium by response time tuning of distributed autonomous cooperation type
US6088697A (en) 1997-12-18 2000-07-11 International Business Machines Corporation Dynamic change management in an extended remote copy operation
US6131190A (en) 1997-12-18 2000-10-10 Sidwell; Leland P. System for modifying JCL parameters to optimize data storage allocations
US6076148A (en) 1997-12-26 2000-06-13 Emc Corporation Mass storage subsystem and backup arrangement for digital data processing system which permits information to be backed up while host computer(s) continue(s) operating in connection with information stored on mass storage subsystem
US6154787A (en) 1998-01-21 2000-11-28 Unisys Corporation Grouping shared resources into one or more pools and automatically re-assigning shared resources from where they are not currently needed to where they are needed
US6260069B1 (en) 1998-02-10 2001-07-10 International Business Machines Corporation Direct data retrieval in a distributed computing system
EP0945800B1 (en) 1998-03-02 2003-07-16 Hewlett-Packard Company, A Delaware Corporation Data backup system
US6026414A (en) 1998-03-05 2000-02-15 International Business Machines Corporation System including a proxy client to backup files in a distributed computing environment
US7277941B2 (en) * 1998-03-11 2007-10-02 Commvault Systems, Inc. System and method for providing encryption in a storage network by storing a secured encryption key with encrypted archive data in an archive storage device
US6161111A (en) 1998-03-31 2000-12-12 Emc Corporation System and method for performing file-handling operations in a digital data processing system using an operating system-independent file map
US6167402A (en) 1998-04-27 2000-12-26 Sun Microsystems, Inc. High performance message store
US6195703B1 (en) * 1998-06-24 2001-02-27 Emc Corporation Dynamic routing for performance partitioning in a data processing network
US6438595B1 (en) * 1998-06-24 2002-08-20 Emc Corporation Load balancing using directory services in a data processing system
US6421711B1 (en) 1998-06-29 2002-07-16 Emc Corporation Virtual ports for data transferring of a data storage system
US6216202B1 (en) 1998-06-30 2001-04-10 Emc Corporation Method and apparatus for managing virtual storage devices in a storage system
US6269431B1 (en) 1998-08-13 2001-07-31 Emc Corporation Virtual storage and block level direct access of secondary storage for recovery of backup data
GB2341249A (en) 1998-08-17 2000-03-08 Connected Place Limited A method of generating a difference file defining differences between an updated file and a base file
JP3004008B1 (en) 1998-10-20 2000-01-31 三菱電機株式会社 Update history management device and update history management method
US6516314B1 (en) 1998-11-17 2003-02-04 Telefonaktiebolaget L M Ericsson (Publ) Optimization of change log handling
US6487561B1 (en) 1998-12-31 2002-11-26 Emc Corporation Apparatus and methods for copying, backing up, and restoring data using a backup segment size larger than the storage block size
US6212512B1 (en) 1999-01-06 2001-04-03 Hewlett-Packard Company Integration of a database into file management software for protecting, tracking and retrieving data
US6324581B1 (en) 1999-03-03 2001-11-27 Emc Corporation File server system using file system storage, data movers, and an exchange of meta data among data movers for file locking and direct access to shared file systems
JP3763992B2 (en) 1999-03-30 2006-04-05 富士通株式会社 Data processing apparatus and recording medium
US6389432B1 (en) 1999-04-05 2002-05-14 Auspex Systems, Inc. Intelligent virtual volume access
US6516348B1 (en) 1999-05-21 2003-02-04 Macfarlane Druce Ian Craig Rattray Collecting and predicting capacity information for composite network resource formed by combining ports of an access server and/or links of wide arear network
US6519679B2 (en) 1999-06-11 2003-02-11 Dell Usa, L.P. Policy based storage configuration
US7035880B1 (en) 1999-07-14 2006-04-25 Commvault Systems, Inc. Modular backup and retrieval system used in conjunction with a storage area network
US7395282B1 (en) 1999-07-15 2008-07-01 Commvault Systems, Inc. Hierarchical backup and retrieval system
US6538669B1 (en) 1999-07-15 2003-03-25 Dell Products L.P. Graphical user interface for configuration of a storage system
US7457233B1 (en) * 1999-07-15 2008-11-25 Juniper Networks, Inc. Method and apparatus for fast reroute in a connection-oriented network
US7389311B1 (en) 1999-07-15 2008-06-17 Commvault Systems, Inc. Modular backup and retrieval system
US6343324B1 (en) 1999-09-13 2002-01-29 International Business Machines Corporation Method and system for controlling access share storage devices in a network environment by configuring host-to-volume mapping data structures in the controller memory for granting and denying access to the devices
US6564228B1 (en) 2000-01-14 2003-05-13 Sun Microsystems, Inc. Method of enabling heterogeneous platforms to utilize a universal file system in a storage area network
US6760723B2 (en) 2000-01-31 2004-07-06 Commvault Systems Inc. Storage management across multiple time zones
US6542972B2 (en) * 2000-01-31 2003-04-01 Commvault Systems, Inc. Logical view and access to physical storage in modular data and storage management system
US7003641B2 (en) 2000-01-31 2006-02-21 Commvault Systems, Inc. Logical view with granular access to exchange data managed by a modular data and storage management system
US6658436B2 (en) 2000-01-31 2003-12-02 Commvault Systems, Inc. Logical view and access to data managed by a modular data and storage management system
US6721767B2 (en) 2000-01-31 2004-04-13 Commvault Systems, Inc. Application specific rollback in a computer system
US20020103889A1 (en) * 2000-02-11 2002-08-01 Thomas Markson Virtual storage layer approach for dynamically associating computer storage with processing hosts
US20020120741A1 (en) 2000-03-03 2002-08-29 Webb Theodore S. Systems and methods for using distributed interconnects in information management enviroments
US6629189B1 (en) 2000-03-09 2003-09-30 Emc Corporation Method and apparatus for managing target devices in a multi-path computer system
US6792472B1 (en) * 2000-03-31 2004-09-14 International Business Machines Corporation System, method and computer readable medium for intelligent raid controllers operating as data routers
US6356801B1 (en) 2000-05-19 2002-03-12 International Business Machines Corporation High availability work queuing in an automated data storage library
US6691209B1 (en) * 2000-05-26 2004-02-10 Emc Corporation Topological data categorization and formatting for a mass storage system
US7404005B1 (en) * 2000-06-09 2008-07-22 International Business Machines Corporation Method, system, and program for selecting one of multiple paths to communicate with a device
US6330642B1 (en) 2000-06-29 2001-12-11 Bull Hn Informatin Systems Inc. Three interconnected raid disk controller data processing system architecture
US6925476B1 (en) 2000-08-17 2005-08-02 Fusionone, Inc. Updating application data including adding first change log to aggreagate change log comprising summary of changes
NO313399B1 (en) 2000-09-14 2002-09-23 Fast Search & Transfer Asa Procedure for searching and analyzing information in computer networks
US6611849B1 (en) 2000-09-29 2003-08-26 Palm Source, Inc. System for synchronizing databases on multiple devices utilizing a home base
US6591188B1 (en) * 2000-11-01 2003-07-08 Navigation Technologies Corp. Method, system and article of manufacture for identifying regularly traveled routes
US7106691B1 (en) 2000-11-01 2006-09-12 At&T Corp. Method for tracking source and destination internet protocol data
US6985956B2 (en) * 2000-11-02 2006-01-10 Sun Microsystems, Inc. Switching system
US7313614B2 (en) * 2000-11-02 2007-12-25 Sun Microsystems, Inc. Switching system
US7068597B1 (en) 2000-11-27 2006-06-27 3Com Corporation System and method for automatic load balancing in a data-over-cable network
US6823477B1 (en) 2001-01-23 2004-11-23 Adaptec, Inc. Method and apparatus for a segregated interface for parameter configuration in a multi-path failover system
US20020124137A1 (en) 2001-01-29 2002-09-05 Ulrich Thomas R. Enhancing disk array performance via variable parity based load balancing
US7231391B2 (en) 2001-02-06 2007-06-12 Quest Software, Inc. Loosely coupled database clusters with client connection fail-over
US20020159458A1 (en) 2001-04-27 2002-10-31 Foster Michael S. Method and system for reserved addressing in a communications network
US7007189B2 (en) * 2001-05-07 2006-02-28 Sun Microsystems, Inc. Routing scheme using preferred paths in a multi-path interconnection fabric in a storage network
US6883108B2 (en) * 2001-05-07 2005-04-19 Sun Microsystems, Inc. Fault-tolerant routing scheme for a multi-path interconnection fabric in a storage network
US7343410B2 (en) * 2001-06-28 2008-03-11 Finisar Corporation Automated creation of application data paths in storage area networks
US7024517B1 (en) 2001-09-27 2006-04-04 Emc Corporation System and method for configuring data storage in accordance with workload requirements
US20030079018A1 (en) * 2001-09-28 2003-04-24 Lolayekar Santosh C. Load balancing in a storage network
US20030145086A1 (en) * 2002-01-29 2003-07-31 O'reilly James Scalable network-attached storage system
JP2003248611A (en) 2002-02-26 2003-09-05 Hitachi Ltd Storage management integration system and its storage management control method
US7072304B2 (en) 2002-02-27 2006-07-04 Nortel Networks Limited Network path selection based on bandwidth
JP4196579B2 (en) * 2002-04-10 2008-12-17 株式会社日立製作所 Storage operation management method and system
US7546364B2 (en) 2002-05-16 2009-06-09 Emc Corporation Replication of remote copy data for internet protocol (IP) transmission
JP2003345631A (en) * 2002-05-28 2003-12-05 Hitachi Ltd Computer system and allocating method for storage area
TWI256556B (en) 2002-07-08 2006-06-11 Via Tech Inc Distributed concurrent version management system and method
GB2410106B (en) 2002-09-09 2006-09-13 Commvault Systems Inc Dynamic storage device pooling in a computer system
US7162496B2 (en) 2002-09-16 2007-01-09 Commvault Systems, Inc. System and method for blind media support
US8370542B2 (en) 2002-09-16 2013-02-05 Commvault Systems, Inc. Combined stream auxiliary copy system and method
US7307948B2 (en) * 2002-10-21 2007-12-11 Emulex Design & Manufacturing Corporation System with multiple path fail over, fail back and load balancing
US7546333B2 (en) * 2002-10-23 2009-06-09 Netapp, Inc. Methods and systems for predictive change management for access paths in networks
US7401064B1 (en) 2002-11-07 2008-07-15 Data Advantage Group, Inc. Method and apparatus for obtaining metadata from multiple information sources within an organization in real time
US7275103B1 (en) * 2002-12-18 2007-09-25 Veritas Operating Corporation Storage path optimization for SANs
US7480512B2 (en) * 2004-01-16 2009-01-20 Bones In Motion, Inc. Wireless device, program products and methods of using a wireless device to deliver services
KR101168423B1 (en) 2003-02-05 2012-07-25 가부시키가이샤 자나비 인포메틱스 Path search method of navigation apparatus and display method of traffic information
US7290168B1 (en) * 2003-02-28 2007-10-30 Sun Microsystems, Inc. Systems and methods for providing a multi-path network switch system
US7383264B2 (en) 2003-03-27 2008-06-03 Hitachi, Ltd. Data control method for duplicating data between computer systems
US7174433B2 (en) 2003-04-03 2007-02-06 Commvault Systems, Inc. System and method for dynamically sharing media in a computer network
US7158985B1 (en) 2003-04-09 2007-01-02 Cisco Technology, Inc. Method and apparatus for efficient propagation of large datasets under failure conditions
US6839724B2 (en) 2003-04-17 2005-01-04 Oracle International Corporation Metamodel-based metadata change management
US20040243699A1 (en) * 2003-05-29 2004-12-02 Mike Koclanes Policy based management of storage resources
US20050038954A1 (en) * 2003-06-04 2005-02-17 Quantum Corporation Storage drive having universal format across media types
US7467168B2 (en) 2003-06-18 2008-12-16 International Business Machines Corporation Method for mirroring data at storage locations
US7454569B2 (en) 2003-06-25 2008-11-18 Commvault Systems, Inc. Hierarchical system and method for performing storage operations in a computer network
JP4492084B2 (en) 2003-10-07 2010-06-30 株式会社日立製作所 Storage path control method
JP4066932B2 (en) * 2003-11-10 2008-03-26 株式会社日立製作所 Computer resource allocation method based on prediction
CA2544063C (en) * 2003-11-13 2013-09-10 Commvault Systems, Inc. System and method for combining data streams in pilelined storage operations in a storage network
WO2005050381A2 (en) 2003-11-13 2005-06-02 Commvault Systems, Inc. Systems and methods for performing storage operations using network attached storage
US7440982B2 (en) 2003-11-13 2008-10-21 Commvault Systems, Inc. System and method for stored data archive verification
CA2546304A1 (en) 2003-11-13 2005-05-26 Commvault Systems, Inc. System and method for performing an image level snapshot and for restoring partial volume data
US20050114595A1 (en) * 2003-11-26 2005-05-26 Veritas Operating Corporation System and method for emulating operating system metadata to provide cross-platform access to storage volumes
US7634582B2 (en) * 2003-12-19 2009-12-15 Intel Corporation Method and architecture for optical networking between server and storage area networks
US7469262B2 (en) 2003-12-29 2008-12-23 Oracle International Corporation Customizable metadata merging framework
US7406509B2 (en) * 2004-01-07 2008-07-29 Network Appliance, Inc. Dynamic switching of a communication port in a storage system between target and initiator modes
JP4477370B2 (en) 2004-01-30 2010-06-09 株式会社日立製作所 Data processing system
US8006056B2 (en) 2004-01-30 2011-08-23 Hewlett-Packard Development Company, L.P. Storage system including capability to move a virtual storage device group without moving data
JP2005217815A (en) 2004-01-30 2005-08-11 Hitachi Ltd Path control method
US7533181B2 (en) 2004-02-26 2009-05-12 International Business Machines Corporation Apparatus, system, and method for data access management
US8230085B2 (en) * 2004-04-12 2012-07-24 Netapp, Inc. System and method for supporting block-based protocols on a virtual storage appliance executing within a physical storage appliance
US20050228875A1 (en) * 2004-04-13 2005-10-13 Arnold Monitzer System for estimating processing requirements
JP4476683B2 (en) 2004-04-28 2010-06-09 株式会社日立製作所 Data processing system
US7340652B2 (en) 2004-04-30 2008-03-04 International Business Machines Corporation Invalidation of storage control unit cache metadata
US7346751B2 (en) * 2004-04-30 2008-03-18 Commvault Systems, Inc. Systems and methods for generating a storage-related metric
US8266406B2 (en) * 2004-04-30 2012-09-11 Commvault Systems, Inc. System and method for allocation of organizational resources
US7617321B2 (en) 2004-05-07 2009-11-10 International Business Machines Corporation File system architecture requiring no direct access to user data from a metadata manager
JP4335090B2 (en) 2004-05-14 2009-09-30 シャープ株式会社 Mobile terminal device
US8423634B2 (en) * 2004-05-24 2013-04-16 Neustar, Inc. System and method for determining cost of website performance
EP1782246B1 (en) 2004-07-07 2020-02-12 Sciencelogic, LLC Self configuring network management system
US7131027B2 (en) * 2004-07-09 2006-10-31 Hitachi, Ltd. Method and apparatus for disk array based I/O routing and multi-layered external storage linkage
JP4545529B2 (en) 2004-08-27 2010-09-15 株式会社日立製作所 Snapshot creation method and program, and storage system
US8060650B2 (en) * 2004-10-27 2011-11-15 Hewlett-Packard Development Company, L.P. Diagnosing a path in a storage network
US7472238B1 (en) * 2004-11-05 2008-12-30 Commvault Systems, Inc. Systems and methods for recovering electronic information from a storage medium
US7536291B1 (en) 2004-11-08 2009-05-19 Commvault Systems, Inc. System and method to support simulated storage operations
US7529745B2 (en) 2004-11-19 2009-05-05 International Business Machines Corporation Method of verifying metadata of a migrated file
KR100611578B1 (en) 2004-11-23 2006-08-10 한국전자통신연구원 A resource allocation device for providing the differentiated service, and a method thereof
JP2006178811A (en) 2004-12-24 2006-07-06 Hitachi Ltd Storage system, and path control method for the system
JP2006209636A (en) 2005-01-31 2006-08-10 Hitachi Ltd Method for maintaining snapshot
JP4671720B2 (en) 2005-03-11 2011-04-20 株式会社日立製作所 Storage system and data migration method
US7489639B2 (en) * 2005-03-23 2009-02-10 International Business Machines Corporation Root-cause analysis of network performance problems
US7461230B1 (en) 2005-03-31 2008-12-02 Symantec Operating Corporation Maintaining spatial locality of write operations
JP4738941B2 (en) 2005-08-25 2011-08-03 株式会社日立製作所 Storage system and storage system management method
JP4686606B2 (en) * 2005-09-08 2011-05-25 インターナショナル・ビジネス・マシーンズ・コーポレーション Method, computer program, and system for dynamic distribution of input / output workload among removable media devices attached via multiple host bus adapters
US20070088702A1 (en) * 2005-10-03 2007-04-19 Fridella Stephen A Intelligent network client for multi-protocol namespace redirection
US7689736B2 (en) 2005-11-07 2010-03-30 Dot Hill Systems Corporation Method and apparatus for a storage controller to dynamically determine the usage of onboard I/O ports
US20070130344A1 (en) * 2005-11-14 2007-06-07 Pepper Timothy C Using load balancing to assign paths to hosts in a network
US7657550B2 (en) 2005-11-28 2010-02-02 Commvault Systems, Inc. User interfaces and methods for managing data in a metabase
WO2007062258A2 (en) 2005-11-28 2007-05-31 Storagedna, Inc. Distributed file system with file fragmentation
US7636743B2 (en) 2005-12-19 2009-12-22 Commvault Systems, Inc. Pathname translation in a data replication system
US7606844B2 (en) * 2005-12-19 2009-10-20 Commvault Systems, Inc. System and method for performing replication copy storage operations
US7617253B2 (en) 2005-12-19 2009-11-10 Commvault Systems, Inc. Destination systems and methods for performing data replication
US8572330B2 (en) 2005-12-19 2013-10-29 Commvault Systems, Inc. Systems and methods for granular resource management in a storage network
US7620710B2 (en) 2005-12-19 2009-11-17 Commvault Systems, Inc. System and method for performing multi-path storage operations
CA2632935C (en) 2005-12-19 2014-02-04 Commvault Systems, Inc. Systems and methods for performing data replication
US7543125B2 (en) * 2005-12-19 2009-06-02 Commvault Systems, Inc. System and method for performing time-flexible calendric storage operations
US7651593B2 (en) 2005-12-19 2010-01-26 Commvault Systems, Inc. Systems and methods for performing data replication
US7617262B2 (en) 2005-12-19 2009-11-10 Commvault Systems, Inc. Systems and methods for monitoring application data in a data replication system
US7904681B1 (en) 2006-06-30 2011-03-08 Emc Corporation Methods and systems for migrating data with minimal disruption
US7882077B2 (en) 2006-10-17 2011-02-01 Commvault Systems, Inc. Method and system for offline indexing of content and classifying stored data
US20080147878A1 (en) * 2006-12-15 2008-06-19 Rajiv Kottomtharayil System and methods for granular resource management in a storage network
US7734669B2 (en) 2006-12-22 2010-06-08 Commvault Systems, Inc. Managing copies of data
US9098495B2 (en) 2008-06-24 2015-08-04 Commvault Systems, Inc. Application-aware and remote single instance data management
US8307177B2 (en) 2008-09-05 2012-11-06 Commvault Systems, Inc. Systems and methods for management of virtualization data
US8578120B2 (en) 2009-05-22 2013-11-05 Commvault Systems, Inc. Block-level single instancing
US20100333116A1 (en) 2009-06-30 2010-12-30 Anand Prahlad Cloud gateway system for managing data storage to cloud storage sites
US8364652B2 (en) 2010-09-30 2013-01-29 Commvault Systems, Inc. Content aligned block-based deduplication
US8484419B2 (en) * 2010-11-24 2013-07-09 International Business Machines Corporation Systems and methods for backing up storage volumes in a storage system
US9020900B2 (en) 2010-12-14 2015-04-28 Commvault Systems, Inc. Distributed deduplicated storage system
US20120150818A1 (en) 2010-12-14 2012-06-14 Commvault Systems, Inc. Client-side repository in a networked deduplicated storage system
US20160092115A1 (en) * 2014-09-29 2016-03-31 Hewlett-Packard Development Company, L. P. Implementing storage policies regarding use of memory regions

Also Published As

Publication number Publication date
US20160277499A1 (en) 2016-09-22
US9930118B2 (en) 2018-03-27
US20070198722A1 (en) 2007-08-23
US20140012994A1 (en) 2014-01-09
US8572330B2 (en) 2013-10-29
US20180278689A1 (en) 2018-09-27
US9313143B2 (en) 2016-04-12

Similar Documents

Publication Publication Date Title
US20200267219A1 (en) Systems and methods for granular resource management in a storage network
US20110004683A1 (en) Systems and Methods for Granular Resource Management in a Storage Network
US11132139B2 (en) Systems and methods for migrating components in a hierarchical storage network
US8661216B2 (en) Systems and methods for migrating components in a hierarchical storage network
US9251190B2 (en) System and method for sharing media in a computer network
US7606844B2 (en) System and method for performing replication copy storage operations
US8103829B2 (en) Hierarchical systems and methods for performing storage operations in a computer network

Legal Events

Date Code Title Description
AS Assignment

Owner name: COMMVAULT SYSTEMS, INC., NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOTTOMTHARAYIL, RAJIV;GOKHALE, PARAG;MULLER, MARCUS S.;SIGNING DATES FROM 20070411 TO 20070416;REEL/FRAME:051921/0328

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT, ILLINOIS

Free format text: SECURITY INTEREST;ASSIGNOR:COMMVAULT SYSTEMS, INC.;REEL/FRAME:058496/0836

Effective date: 20211213

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION