WO2016046977A1 - ストレージ装置及びその保守運用方式 - Google Patents
ストレージ装置及びその保守運用方式 Download PDFInfo
- Publication number
- WO2016046977A1 WO2016046977A1 PCT/JP2014/075702 JP2014075702W WO2016046977A1 WO 2016046977 A1 WO2016046977 A1 WO 2016046977A1 JP 2014075702 W JP2014075702 W JP 2014075702W WO 2016046977 A1 WO2016046977 A1 WO 2016046977A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- pool
- amount
- storage
- remaining
- capacity
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/0614—Improving the reliability of storage systems
- G06F3/0616—Improving the reliability of storage systems in relation to life time, e.g. increasing Mean Time Between Failures [MTBF]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/08—Error detection or correction by redundancy in data representation, e.g. by using checking codes
- G06F11/10—Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
- G06F11/1076—Parity data used in redundant arrays of independent storages, e.g. in RAID systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0629—Configuration or reconfiguration of storage systems
- G06F3/0631—Configuration or reconfiguration of storage systems by allocating resources to storage systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0638—Organizing or formatting or addressing of data
- G06F3/064—Management of blocks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0638—Organizing or formatting or addressing of data
- G06F3/0644—Management of space entities, e.g. partitions, extents, pools
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0646—Horizontal data movement in storage systems, i.e. moving data in between storage devices or systems
- G06F3/0647—Migration mechanisms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0653—Monitoring storage devices or systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0655—Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
- G06F3/0659—Command handling arrangements, e.g. command buffers, queues, command scheduling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/067—Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/0671—In-line storage system
- G06F3/0683—Plurality of storage devices
- G06F3/0688—Non-volatile semiconductor memory arrays
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/08—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers from or to individual record carriers, e.g. punched card, memory card, integrated circuit [IC] card or smart card
Definitions
- the present invention relates to a storage apparatus equipped with a semiconductor storage medium having a rewritable life and its maintenance operation method.
- SSD Solid State Drive
- NAND type flash memory NAND type flash memory
- a flash memory records data by applying and applying electrons to a floating gate blocked by an insulating film. Since the insulating film is worn by forced passage of electrons, the number of data rewrites is limited. The number of rewritable times is decreasing with process miniaturization. Accordingly, in storage using a flash memory such as an SSD, in addition to managing the capacity, an operation in consideration of the rewrite life is required.
- Patent Document 1 discloses a technique for proactively maintaining data by copying data to a spare before failure.
- the concept of coping with the service life is the same as the failure replacement.
- the other is a method of extending the life by adding a RAID configuration group composed of a flash memory drive or a plurality of flash memory drives and distributing the load of rewrites.
- a wear leveling technology in the flash memory drive to use the flash chip mounted inside evenly.
- the number of times of rewriting is equalized between multiple flash memory drives (FM packages).
- FM packages flash memory drives
- the first method is exchange frequency. Since write data for a RAID group is written in a striped manner on all the constituent drives, the drive groups in the same RAID group reach the end of their lives almost simultaneously. For this reason, even if replacement is performed one by one, after all, drive replacement occurs one after another. In addition, rebuilding or copying to another drive occurs at the time of replacement, which affects system performance, but the user cannot actively cope with it. In addition, when the replacement occurs near the life of the apparatus (for example, 5 years), it is replaced with a drive that has an unnecessarily long life in maintenance.
- the present invention has been made in view of the above, and an object of the present invention is to make it possible to know the capacity necessary for resolving the shortage of life, and to enable maintenance operators to perform maintenance replacement with little waste. To provide maintenance information. Another object of the present invention is to enable the operation manager to manage the rewrite life by the existing capacity operation management.
- a storage apparatus includes one or more storage controllers and a plurality of non-volatile memory (NVM) drives, and the storage controller stores data using the NVM drives.
- the operation manager can manage the capacity and the life in an integrated manner. Problems, configurations, and effects other than those described above will become apparent from the following description of embodiments.
- FIG. 1 is a block diagram showing the configuration of a storage apparatus according to an embodiment of the present invention.
- FIG. 2 is a diagram showing the appearance of the storage apparatus according to the embodiment of the present invention.
- FIG. 3 is a schematic diagram showing a relationship between a virtual logical volume, a capacity pool, a RAID group, and a drive in the embodiment of the present invention.
- FIG. 4 is a diagram showing a pool capacity management screen of the storage apparatus according to the embodiment of the present invention.
- FIG. 5 is a diagram showing a drive management information table in the embodiment of the present invention.
- FIG. 6 is a diagram showing a RAID group management information table in the embodiment of the present invention.
- FIG. 7 is a diagram showing a pool management information table in the embodiment of the present invention.
- FIG. 1 is a block diagram showing the configuration of a storage apparatus according to an embodiment of the present invention.
- FIG. 2 is a diagram showing the appearance of the storage apparatus according to the embodiment of the present invention.
- FIG. 8 is a diagram illustrating the relationship between the aging lifetime, the rewritable lifetime, and the remaining rewritable capacity in the embodiment of the present invention.
- FIG. 9 is a flowchart showing a monitoring control flow for the storage apparatus according to the first embodiment of the present invention to manage the pool life.
- FIG. 10 is a flowchart of maintenance work.
- FIG. 11 is an RG management table showing the correspondence between RAID groups that can be configured in the storage apparatus and rewritable capacity (maintenance drive capacity).
- FIG. 12 is a flowchart showing an operation when the storage apparatus 1 according to the present embodiment performs pool life management.
- FIG. 13 is an example of a GUI showing the transition of the capacity pool free capacity provided by the storage apparatus 1 in this embodiment through the operation management terminal 31.
- FIG. 14 is a flowchart showing the form of the maintenance method in the present embodiment.
- management table an expression such as “management table”, but the various types of information may be expressed using a data structure other than a table. Further, the “management table” can be referred to as “management information” to indicate that it does not depend on the data structure.
- the program is executed by a processor such as an MP (Micro Processor) or a CPU (Central Processing Unit), and performs a predetermined process. Since the processor performs processing while appropriately using storage resources (for example, a memory) and a communication interface device (for example, a communication port), the subject of the processing may be a processor.
- the processor may have dedicated hardware in addition to the CPU.
- the computer program may be installed on each computer from a program source.
- the program source may be provided by, for example, a program distribution server or a storage medium.
- each element for example, the controller can be identified by a number or the like, but other types of identification information such as a name may be used as long as the information can identify each element.
- identification information such as a name may be used as long as the information can identify each element.
- the same reference numerals are given to the same parts.
- the present invention is not limited to the present embodiment, and any application examples that meet the idea of the present invention can be applied. Included in the scope. Further, unless specifically limited, each component may be plural or singular.
- FIG. 1 is a block diagram showing the configuration of a storage system in which a nonvolatile memory drive (NVM drive) with a limited number of rewrites is mounted in this embodiment.
- the NVM drives D0 to Dn are, for example, SSDs, add-in cards, or modules having a unique form.
- the storage device 1 includes a controller housing 50 and a drive housing 51. However, when the storage device 1 is realized by a general-purpose server device, the respective cases may not be distinguished.
- the storage device 1 is connected to an operation management terminal 31 through a management network 30, and device settings and information acquisition can be performed through the operation management terminal 31.
- the operation management terminal 31 may be included as a part of the storage apparatus 1.
- the storage apparatus 1 is connected to the host computers 20A to 20C through the storage network 40.
- This is realized by the storage network 40, the Fiber Channel, the Ethernet (registered trademark) network, or the PCI-Express (registered trademark).
- an operating system and an application operating on the operating system operate and issue an I / O to the storage apparatus 1.
- the storage device is realized by a general-purpose server device, the host computer and the storage device may be physically the same, and the storage network 40 may be omitted at that time. In some cases, the host computer may be implemented in software like a virtual machine.
- the NVM drives D0 to Dn are connected to the storage controller package through the back-end connection mechanism 500 and the back-end connection wiring 510 provided in the drive housing 51 and the I / F cards 90A and 90B provided in the controller housing 50. Connected to 70A and 70B.
- the back-end interconnection mechanism 500 is, for example, a SAS (Serial Attached SCSI) expander or a PCI-Express switch
- the back-end connection wiring 510 is, for example, a SAS cable or a PCI-Express cable.
- the I / F cards 90A and 90B are, for example, an HBA (Host Bus Adapter) equipped with a SAS protocol chip, a PCI-Express Extension board, or the like.
- the controller housing 50 includes storage controller packages 70A and 70B in which a storage control program operates, the drive connection I / F cards 90A and 90B, and an I / F card 80A for connecting to the host computers 20A to 20C. 81A, 80B, 81B are provided. Both controllers are connected by a plurality of internal data transfer paths 61 and 62 in order to transfer data to each other.
- each package is shown as two systems, A-system and B-system, but two or more packages may be provided and connected to each other, or an I / F card (for example, 91A or 91B), a plurality of storage controller housings 50 may be coupled to each other. Or you may be comprised only by A type
- the operation management terminal 31 or the storage device 1 is connected to the management server of the maintenance center 99 via the network 95.
- information such as the remaining rewriting capacity and remaining life of a pool configured with NVM drives described later can be acquired, and the state of the storage apparatus 1 can be monitored and managed.
- the management server 991 can also transmit remote maintenance work from the maintenance center and maintenance cost billing information to the operation management terminal 31. Note that the management server 991 may be directly connected to the storage apparatus 1.
- FIG. 2 is an external view shown to assist in understanding the storage apparatus in the present embodiment.
- the controller housing 50 of the storage apparatus 1 is mounted at the lower part of the rack shown in the figure, and the drive housings 51, 52, and 58 are shown thereon.
- the back end connection wiring 510 and the cables for configuring the storage network 40 are wired on the back surface.
- FIG 3 shows the virtual logical volumes 120A to 120C and the capacity pool 110, the page group 115, the RAID group group 100, and the NVM drive Dn (D0 to D3, etc.) provided by the storage apparatus 1 in this embodiment to the host computers 20A to 20C. It is the figure which showed the relationship.
- the storage apparatus 1 in the present embodiment uses a technique called “Thin Provisioning” in which the virtual logical volume 120 is provided to the host computer 20 and an actual capacity is allocated only to the accessed part.
- the storage apparatus 1 configures a RAID using several NVM drives Dn. This is called a RAID group.
- a RAID group For example, in the RAID group 100A, 3D + 1P is configured using four drives D0, D1, D2, and D3.
- the capacity pool 100 is composed of one or more RAID groups.
- the capacity pool 110 can add a RAID group at any point in time, thereby increasing the capacity of the pool. Further, the capacity pool 110 is managed in fixed length units divided by several KB or several MB called pages 115.
- the host computers 20A to 20C are accessed using the virtual logical volumes 120A to 120C provided by the storage apparatus 1.
- the virtual logical volumes 120A to 102C are empty volumes having a virtual capacity.
- the storage apparatus 1 allocates entities in units of pages 115 from the capacity pool 110 to the LBA area on the virtual logical volume 120. More specifically, the storage apparatus 1 has page mapping information for managing the correspondence between the LBA of the virtual logical volume and the LBA of the capacity pool in units of pages, and updates this information (not shown).
- the storage apparatus 1 selects the page so that the load is distributed to a plurality of RAID groups, and also selects the level of the number of rewrites.
- the capacity pool 110 is managed in units of pages 115, the number of allocated pages is almost synonymous with the used capacity of the capacity pool, and the number of unallocated and available pages is synonymous with the free pool capacity.
- FIG. 4 is an example of a GUI (Graphic User Interface) showing the transition of the capacity pool free capacity provided by the storage apparatus 1 according to the embodiment of the present invention through the operation management terminal 31.
- the horizontal axis of the screen 200 represents the elapsed usage time (for example, the number of days or the number of months) since the start of the capacity pool operation, and the vertical axis represents the free capacity of the pool.
- a solid line 211 indicates a change in the pool capacity. For example, a time point (A) indicates that the capacity is added to the pool and the free capacity is restored.
- a dotted line 215 is a line indicating a threshold value for preventing free space depletion.
- this threshold for example, 20% or less of the total pool capacity
- the storage device 1 uses the operation management terminal 31 or the device SNMP (Simple Network Management Protocol) to inform the operation manager to the pool capacity. Notify (or warn) that addition (addition of a RAID group) is necessary.
- SNMP Simple Network Management Protocol
- FIG. 5 is an example of an NVM drive management table provided to the operation manager by the storage apparatus 1 according to the embodiment of the present invention through the operation management terminal 31.
- the NVM drive management table 250 is a table for managing the capacity, operating time, remaining life time, and the like for each NVM drive.
- the drive of number 2 belongs to the capacity 1.6 TB, RAID group (RG) 0, the rewrite life consumption rate (consumption life) of the drive is 85%, the drive operation time is 34, It can be seen that the specified BW (Bite Written), which is a rewritable capacity for 278 hours, is 12 PB.
- FIG. 6 is an example of a RAID group management information table (RG management information table 260) provided by the storage apparatus 1 in this embodiment through the operation management terminal 31.
- the RG management information table 260 manages the capacity for each RG, the remaining rewritable capacity, the health state of the NVM drive (“good”: high soundness / low deterioration, “warning”: low soundness / high deterioration), and the like. It is a table.
- the RAID group 0 has a capacity of 4.8 TB, belongs to the capacity pool PVOL0, the remaining rewritable capacity (remaining BW) is 3.2 PB, and the soundness state Indicates “good”.
- the specified BW for each RG may be managed in the same manner as the NVM drive management table 250. Further, the soundness state may be determined by an absolute value of the capacity of the remaining BW, or may be determined by a relative value of a ratio (%) of the remaining BW to the specified BW.
- FIG. 7 is an example of a pool management information table, similar to FIG.
- the pool management table 280 is a table for managing the capacity for each pool, the remaining rewritable capacity, the deterioration state of the flash memory (low deterioration: “good”, high deterioration: “warning”), and the like.
- the capacity of the pool PVOL0 is 19.2 TB
- the remaining rewritable capacity (remaining BW) is 15.0 PB
- 24 PB is expected to have insufficient capacity for life
- the state of the NVM drive is “warning”.
- the state of “good” or “warning” is shown, but a state such as “attention required” between “good” and “warning” may be provided.
- the prescribed BW for each pool (the total of the prescribed BWs of RGs belonging to the pool) may be managed.
- the soundness state may be determined based on the absolute value of the remaining BW capacity, or may be determined based on a relative value of the ratio (%) of the remaining BW to the specified BW.
- FIG. 8 schematically shows the relationship between the aging life (product life) of the NVM drive and the rewrite life that varies depending on the write amount to the NVM drive.
- the rewrite life varies depending on the light interval, pattern, and temperature, but is omitted because it is not the essence of the description.
- 8A and 8B show a drive as an example, but the concept is the same even if it is replaced with a RAID group or a pool. Further, the operation management terminal 31 and the operation management server 81 may be able to display the same information as in FIG.
- the NVM drive has a rewritable capacity (specified BW) specified at the time of design, and if rewriting is performed more than expected, it will reach the rewriting life before the aged life (Fig. 8 (1)). If done less, the NVM drive will achieve its aging life (product life) (FIG. 8 (2)).
- the aging life corresponds to the shelf life of the product, and in the case of an enterprise product, for example, it is 5 years after the start of use.
- FIG. 8 (1) represents a case where more data was written than expected (that is, the remaining rewriting capacity was rapidly reduced).
- the assumed line 310 is a line representing a decrease when the rewriting life is used at a uniform pace so as to be approximately equal to the aging life.
- the actual transition 320 is as the name indicates, and the extrapolation in the case of continuing to be used at the inclination is indicated by a dotted line.
- the difference 330 between the points (a) and (b) represents the rewriting capacity that has been consumed more than expected at the current time 350. If it continues to be used at this pace, the drive in this figure reaches its rewriting life at a point 360 that intersects the horizontal axis. Further, the difference 340 between the points (c) and (d) represents the expected value of the rewriting capacity that is insufficient to achieve the aging life.
- FIG. 8 (2) shows a case where writing is less than expected, contrary to FIG. 8 (1).
- the difference 331 represents the rewriting capacity that is surplus than expected at the current time 350. If it continues to be used at this pace, the surplus to the assumption will gradually increase.
- the difference 341 is a rewriting capacity that becomes surplus when the aged life is reached.
- the difference 330 that indicates a shortage is larger than the difference 331 that indicates a surplus, and therefore falls below the assumed line even when leveling. That is, it is determined that the aged life cannot be satisfied.
- the difference 341 representing surplus is larger than the difference 340 when compared at the time of aging, so if it is leveled, it exceeds the expected line. That is, it is determined that the aging life is satisfied.
- the shortage at the present time does not necessarily mean that there will be a shortage in the future.
- the excess and deficiency at the present time is obtained by simple comparison of the acquired values, the calculation cost is small, but the future excess and deficiency requires statistical processing of the transition of the value in order to obtain the expected value, and the calculation cost is large. Based on this, life management in the storage apparatus 1 will be described with reference to FIG.
- FIG. 9 is a flowchart showing a monitoring control flow for the storage apparatus 1 in this embodiment to manage the pool life.
- the processing in this flowchart is executed by the storage controller 70 of the storage apparatus 1 at certain time intervals or based on certain triggers such as capacity addition to the pool.
- the storage apparatus 1 first acquires the remaining rewritable capacity of each drive included in the pool by using means such as SMART (Self-Monitoring Analysis and Reporting Technology) or LOG SENSE command (S1000). When the remaining rewritable capacity cannot be acquired directly, the storage device 1 can perform the remaining rewritable based on information that the storage controller can acquire, such as the transition of drive life rate, WA (Write Amplification) ratio, or the number of error blocks. Infer capacity. The storage device 1 may obtain the remaining rewritable capacity of each drive from the accumulated destage capacity and the like.
- the storage apparatus 1 calculates the excess or deficiency of the current rewrite life from the difference from the assumed line calculated from the specifications of the drive (S1200).
- the storage system 1 obtains the overwriting / shortening of the rewriting life of the RAID group from the overwriting / shortening of the rewriting life of each drive (S1300). Specifically, the value of the drive having the shortest life in the RAID group is set as the value of the RAID group. This is because a RAID group is degenerated if one unit fails. Note that the lifetime may be obtained not by degeneracy but by occlusion.
- the storage apparatus 1 calculates the sum of overwriting / shortening of the rewrite life of each RAID group, and obtains the overwriting / shortening of the rewrite life of the pool (S1400). Then, it is determined whether or not the pool rewrite life is currently short, that is, whether or not the sum of the pool rewrite life excess and deficiency is negative (S1500). Note that after S1500, a “pool state warning / notice of caution” such as S2000 or S2050, which will be described later, may be performed. When it is determined that there is no shortage of the pool rewrite life at the current time (S1500: No), the storage apparatus 1 updates the monitoring result as “good” in the pool management table 280 (S1650), and ends the processing.
- the storage apparatus 1 obtains the expected value of each drive as to how much the rewrite life capacity of the pool is short in the period until the apparatus replacement (process from S1600) )to go into. It should be noted that what should be obtained here is the shortage at the time of device replacement, and is not when the aging life of each drive is reached. This is because the entire storage device is replaced without waiting for the aging life of the drive added during the operation. Note that the period from the introduction of the storage apparatus 1 to the replacement, that is, the product life years of the storage apparatus is set to the same as the standard apparatus warranty period (for example, 5 years) unless otherwise specified.
- the storage device 1 obtains the decreasing slope of the rewrite capacity of each drive from the past statistical information. Then, the required write capacity at the time of replacement of the device itself is obtained by extrapolation (extrapolation method: a method for obtaining a numerical value expected outside the data range based on some known numerical data). Then, the excess / deficiency amount of the rewrite capacity for each drive is calculated from the difference from the assumed line (S1600). Next, the storage system 1 calculates the excess or deficiency of each RAID group by the same method as in S1300 (S1700). Finally, the total excess / deficiency of the pool is obtained (S1800).
- the storage apparatus 1 determines that the capacity pool has insufficient rewrite life at the present time and is expected to be insufficient in the future. Therefore, in order to prompt the drive to be added to the capacity pool, the storage apparatus 1 sets the status of the pool to “warning” in the pool management table 280, and displays the value obtained in S1400 as the short-lived rewriting capacity (S2000). .
- the storage device 1 determines that there is a low need for adding a drive to the storage device 1 immediately though there is a shortage at the present time. Therefore, in the pool management table 280, the status of the pool is set to “Caution” (S2050), and the storage apparatus 1 ends the process. Note that, as shown in FIG. 8 (2), the rewrite life may be longer than the product life, so the storage apparatus 1 is started to be used with a small pool capacity in advance, and the above processing is appropriately executed to An operation for executing addition may be used.
- FIG. 10 is a flowchart of maintenance work
- FIG. 11 is an RG management table 600 showing the correspondence between RAID groups that can be configured by the storage apparatus and rewritable capacity (maintenance drive capacity). It is assumed that the storage system administrator has signed a maintenance service contract with the storage system maintenance provider.
- the operation manager When the pool life becomes “warning”, the operation manager requests maintenance work from the maintenance company at any time or at a planned timing. Therefore, a maintenance work request is input to the operation management terminal 31, and the operation management terminal 31 accepts maintenance work request information and transmits it to the maintenance management server 991 of the maintenance center 90 (S5000).
- the maintenance management server 991 (maintenance provider side) refers to the pool management table 280 shown in FIG. 7 or the information on the maintenance dedicated screen corresponding to the pool management table 280 to calculate the short-life rewriting capacity of the pool (S5100). Then, the maintenance company obtains a configuration that satisfies the shortage capacity from the specified rewritable capacity (specified BW) of the maintenance drive and the RAID group configuration that can be configured (S5200).
- the maintenance provider loads the required number of maintenance drives into the storage apparatus 1 and configures a RAID group (here, 3D + 1P) from the maintenance work terminal or the operation management terminal 31. Then, the maintenance provider adds the configured RAID group NVM drive to the “warning” pool.
- the storage apparatus 1 receives information about the added RAID group (S5300).
- the storage apparatus 1 executes the monitoring flow shown in FIG. 9 and diagnoses the state of the pool again.
- the pool shortage capacity before the addition is 24 PB, and since a RAID group having a rewrite life of 36 PB was added this time, the total excess or shortage of the pool is changed from negative ( ⁇ ) ⁇ 24 PB to positive (+) +12 PB. It becomes. Therefore, the storage apparatus 1 turns off the “warning” display on the operation management terminal 31.
- the maintenance provider confirms that the status of the pool is no longer “warning” and finishes the maintenance (S5400).
- the storage apparatus 1 determines the excess or deficiency based on the current acquired value with a small amount of calculation, then determines the future excess or deficiency as necessary, and determines the capacity. Since it can be shown quantitatively, it is possible to know the capacity to be added to the pool and to maintain it with the minimum necessary capacity.
- the maintenance management server 991 (maintenance worker side) determines the RAID configuration from the rewritable capacity of the maintenance drive and the RG management table 600 of the configurable RAID group.
- the storage apparatus 1 instead of the maintenance management server 991, the storage apparatus 1 itself may determine these and present a configuration to be added to the pool management table 280. In this case, operation errors caused by maintenance workers can be reduced, and more reliable maintenance is possible.
- ⁇ Modification 2 of Example 1> the configuration of the storage device added to the pool after configuring the RAID is shown, but there is a storage device that can be added to the pool in units of drives without configuring a RAID group in advance.
- the storage device 1 stores information on the insufficient drive capacity or the number of insufficient drives in the pool management table 280 in place of the short-lived rewritable capacity, and stores it in the operation management terminal 31 or the maintenance management server 991. You may be notified.
- an NVM drive for adding an RG may be mounted in the storage apparatus 1 in advance, and the RG may be automatically configured and added to the pool.
- Example 2 In the first embodiment, the method for presenting and maintaining the pool life shortage required during the operation period of the apparatus as the shortage rewriting capacity has been described. However, the operation manager must still perform management from the viewpoint of life (FIG. 7) and management from the capacity utilization viewpoint (FIG. 4). Therefore, in the second embodiment, the shortage of the pool life is reflected in the pool capacity transition as shown in the screen 200 of FIG. 4 so that the life can be managed only by capacity management.
- the configuration of the apparatus in the present embodiment is the same as that in the first embodiment, but since the process at the time of shortage of life and the notification method are different, only the difference is described.
- FIG. 12 is a flowchart illustrating an operation when the storage apparatus 1 according to the present embodiment performs pool life management. Note that S1000 (acquisition of remaining rewritable capacity of each drive) to S1400 (calculation of excess / deficiency of pool rewritable life) is the same as that in the first embodiment (FIG. 9), and thus description thereof is omitted.
- the storage apparatus 1 calculates the storage capacity necessary to cover the shortage of the rewrite life, and performs the process of not using the capacity, that is, reducing it from the pool capacity. Do. As a result, the capacity added in the first embodiment is secured from the existing pool prior to the maintenance. In the following description, it is assumed that the use of RAID groups included in the pool is leveled.
- the storage apparatus 1 apportions the insufficient rewrite capacity to each RAID group according to the capacity ratio of the RAID groups included in the pool (S3000).
- the pool PVOL0 is composed of RAID groups RG0, RG1, and RG2, and the capacity of each RAID group is 4.8 TB, 4.8 TB, and 9.6 TB.
- the storage apparatus 1 identifies the drive that has determined the remaining rewritable capacity of the RAID group, and the ratio between the capacity of the drive and the specified rewritable capacity (specified BW) is allocated to the allocated insufficient rewritable capacity. To obtain the capacity that should not be used in the RAID group in order to cover the shortage. That is, “apportioned insufficient rewriting amount ⁇ (drive capacity ⁇ drive regulation BW)” is obtained (S3100).
- the insufficient rewriting amount prorated in RG0 is 6 PB.
- the number of unused valid pages is synonymous with the free capacity of the pool, so reducing the number of valid pages corresponding to the shortage of the pool life by the above processing reduces the available pool capacity. Equivalent to decreasing.
- FIG. 13 is an example of a GUI showing the transition of the capacity pool free capacity provided by the storage apparatus 1 in the present embodiment through the operation management terminal 31.
- the screen 400 in addition to the dotted line 211 representing the transition of the capacity pool due to the page allocation to the virtual logical volume, a line 411 reflecting the shortage of life is shown. In a situation where there is no shortage of the lifetime, the dotted line 211 and the line 411 coincide.
- the line 411 shows a position that has decreased from before. Therefore, as long as the operation manager manages the transition of the pool capacity as shown in the screen 400, the operation manager can manage the lifetime as well.
- FIG. 14 is a flowchart showing a maintenance method according to this embodiment.
- the difference from the first embodiment is that the pool remaining amount is reduced, and the maintenance opportunity is not always the lifetime.
- the pool remaining amount may be reduced simply by capacity allocation to the virtual logical volume 120.
- the operation manager has concluded a maintenance service contract with the maintenance provider, and the capacity corresponding to the lifetime maintenance is assumed to be free maintenance.
- capacity addition for virtual logical volume allocation is charged (purchasing a pool).
- the operation manager requests maintenance when the remaining capacity of the pool decreases, because capacity is added to the pool through planned maintenance. Therefore, the maintenance work request is input to the operation management terminal 31, and the operation management terminal 31 receives the request information for the maintenance work and transmits it to the maintenance management server 991 of the maintenance center 90 (S6000). Alternatively, the operation management terminal 31 may automatically request maintenance based on the rules set by the operation manager on the operation management terminal 31.
- the maintenance management server 991 (maintenance provider side) obtains (confirms) the screen 400 of FIG. 13 or the maintenance management information corresponding thereto, and calculates (confirms) the pool decrease state and the decrease in the life ( S6100).
- the maintenance management server 991 determines the capacity to be added to the pool and the RAID group configuration based on the contents of the maintenance contract concluded or information such as the maintenance procedure manual determined by the operation manager ( S6200). Then, the maintenance provider adds the determined RAID group configuration to the pool of the storage device 1. The storage apparatus 1 receives information about the added RAID group (S6300).
- the maintenance management server 991 (maintenance provider side) obtains (confirms) the result that the added capacity is safely reflected as the free capacity pool. Then, the maintenance management server 991 (maintenance provider side) calculates the maintenance fee for additional capacity by discounting the maintenance cost corresponding to the life elimination from the added pool capacity, and the operation management terminal 31 (operation Send maintenance information billing information to the administrator. The operation management terminal 31 (operation manager side) receives the maintenance cost billing information from the maintenance management server 991 (maintenance provider side) (S6400). The processing in the storage apparatus 1 including the operation management terminal 31 and the maintenance management server 991 of the maintenance center 99 is thus completed.
- the operation manager can operate by monitoring only the capacity fluctuation without separately managing the life of the pool and the capacity fluctuation due to the pure use of the pool. Can be simplified.
- the maintenance provider can unify lifetime maintenance and capacity maintenance as a capacity addition to the pool, maintenance with less cost waste is possible.
- the storage capacity is specified by the storage device, an appropriate cost can be charged to the operation manager.
- ⁇ Modification 1 of Example 2> In the screen 13 of the second embodiment, the decrease due to the lifetime and the capacity allocation to the virtual logical volume are shown separately. However, the operation manager no longer needs to consider the lifetime itself, so the lifetime is not displayed. You may do it (that is, only the line 411). In this case, an operation similar to that of a storage device in which only a conventional HDD is mounted is possible. In addition, since the maintenance provider can charge the operation administrator for additional capacity in accordance with the amount of writing, the maintenance provider can provide low I / O even for users who use high I / O (users with long life shortages). It is possible to provide a fair maintenance service to users (users who have little or no shortage of life).
- an additional pool capacity calculated by (3 data drives ⁇ 3RG) ⁇ (5-year life / drive) may be added. Since the above calculation is about 2 drives, it is only necessary to add 1 RG (4 drives) at most in maintenance work. Therefore, there is an effect that the maintenance man-hour, time and cost (personnel cost and material cost (drive cost, etc.)) can be reduced by at least about 70% compared to the conventional case. Also, by adding drives in units of RGs, the access load is rebalanced in the pool (load distribution), so that it is possible to improve the access performance from the host device.
- the shortage of the life of the NVM drive can be covered by adding the pool capacity, so that it is possible to flexibly cope with a change in operation.
- since it is not necessary to provide a plurality of models for the NVM drive it is possible to manufacture one kind of mass product, and it is possible to reduce the cost and the product cost.
- the capacity of the NVM drive that can still be used can be reduced when the life of the storage device is exhausted, the amount of resources to be discarded can be reduced, and an environmentally friendly product can be provided.
- this invention is not limited to the above-mentioned Example, Various modifications are included.
- the above-described embodiments have been described in detail for easy understanding of the present invention, and are not necessarily limited to those having all the configurations described. Further, a part of the configuration of one embodiment can be replaced with the configuration of another embodiment, and the configuration of another embodiment can be added to the configuration of one embodiment. Further, it is possible to add, delete, and replace other configurations for a part of the configuration of each embodiment.
- Each of the above-described configurations, functions, processing units, processing means, and the like may be realized by hardware by designing a part or all of them with, for example, an integrated circuit.
- Each of the above-described configurations, functions, and the like may be realized by software by interpreting and executing a program that realizes each function by the processor.
- Information such as programs, tables, and files for realizing each function may be stored in a memory, a recording device such as a hard disk or SSD, or a recording medium such as an IC card, SD card, or DVD.
- a recording device such as a hard disk or SSD
- a recording medium such as an IC card, SD card, or DVD.
- control lines and information lines indicate what is considered necessary for the explanation, and not all the control lines and information lines on the product are necessarily shown. Actually, it may be considered that almost all the components are connected to each other.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Quality & Reliability (AREA)
- Microelectronics & Electronic Packaging (AREA)
- Debugging And Monitoring (AREA)
Abstract
Description
1つは、故障交換を待つことである。一般にストレージ装置ではRAID(Redundant Array of Independent/Inexpensive Disks)技術等によって冗長化がなされているので、故障してもすぐにデータ損失には繋がらない。そこで、寿命に到達したドライブは、故障扱いとして閉塞し、新たなドライブと交換することで延命する方法である。閉塞した場合、RAIDは新たなドライブを用いてリビルド処理を行うので、終了するまでの間の処理性能の低下や可用性の低下を許容しなければならない。このため、故障前にデータをスペアにコピーして予防的に保守する技術が特許文献1として知られている。しかしながら、寿命に対する対処の考え方としては故障交換と同じである。
実施例1について、図1~図9を用いて説明する。
図1は本実施形態における、書換回数に制約のある不揮発メモリドライブ(NVMドライブ)を搭載したストレージシステムの構成を示すブロック図である。NVMドライブD0~Dnは、例えばSSDや、アドインカードや独自の形態を持つモジュールである。ストレージ装置1は、コントローラ筐体50とドライブ筐体51から構成される。ただし、ストレージ装置1が汎用的なサーバ機器によって実現されるような場合、各筐体は区別されないこともある。ストレージ装置1は管理用ネットワーク30を通じて運用管理端末31と接続されており、装置の設定や情報の取得はこの運用管理端末31を通じて実施することができる。運用管理端末31はストレージ装置1の一部として含まれていてもよい。
ストレージ装置1のコントローラ筐体50は図に示すラックの下部に搭載されており、その上にドライブ筐体51、52、58が示されている。図に表れていないが、背面には先に示したバックエンド接続配線510やストレージネットワーク40を構成するためのケーブルが配線される。
そして、プールの書換寿命が現時点で不足しているか、すなわちプール書換寿命の過不足の総和が負となるか否かを判断する(S1500)。なお、S1500の後に、後述するS2000またはS2050のような「プール状態の警告/要注意の通知」等をおこなってもよい。
現時点でのプールの書換寿命の不足がないと判断された場合(S1500:No)、ストレージ装置1は監視の結果をプール管理テーブル280に「良」として更新し(S1650)、処理を終了する
次にストレージ装置1は、S1300と同様の方法で各RAIDグループの過不足を算出する(S1700)。そして、最終的にプールの過不足総量を求める(S1800)。なお、前述のようにドライブ毎の書換可能容量の想定ラインと書換実績による残書換可能容量を、RG毎またはプール毎に設定及び算出することで、マクロ視点での記憶容量管理も可能とできる。
保守提供者は、当該プールの状態が「警告」でなくなったことを確認して保守を終える(S5400)。
実施例1では、保守管理サーバ991(保守作業者側)は、保守ドライブの書換可能容量と構成可能なRAIDグループのRG管理テーブル600から、RAID構成を決定していた。しかしながら、保守管理サーバ991の代わりにストレージ装置1自体がこれらを判断し、プール管理テーブル280に追加すべき構成を提示するようにしてもよい。この場合、保守作業者によるオペレーションミスを減らすことができ、より確実な保守が可能となる。
実施例1では、RAIDを構成してからプールに追加するストレージ装置の構成を示したが、あらかじめRAIDグループを構成することなく、ドライブ単位でプールに追加できるストレージ装置がある。このような装置の場合、ストレージ装置1は、プール管理テーブル280に、対寿命不足書換可能容量に代わり、不足ドライブ容量もしくは不足ドライブ台数の情報を格納し、運用管理端末31や保守管理サーバ991に通知してもよい。この場合、RAIDグループ構成の制約を受けないため、更に無駄の少ない保守が可能となる。また、保守作業者によるオペレーションミスを減らすことができ、より確実な保守が可能となる。なお、保守作業者によるNVMドライブの追加ではなく、予めストレージ装置1にRG追加用のNVMドライブを搭載して、それを自動的にRG構成してプールに追加するようにしてもよい。
実施例1では、装置の稼働期間中に必要なプール寿命不足分を、不足書換容量として提示し、保守する方法について述べた。しかし、運用管理者は、依然として寿命視点での管理(図7)と容量利用視点での管理(図4)とを個別に行わなければならない。そこで実施例2では、プール寿命の不足分を図4の画面200に示すようなプール容量推移に反映することで、容量管理のみで寿命も管理できるようにする。本実施例における装置の構成形態は実施例1に同じであるが、寿命不足時の処理と通知方法が異なるため、その差分についてのみ記述する。
そして、ストレージ装置1は、定期的な監視によってプール容量の推移画面の更新を行い、処理を終了する(S3400)。
実施例2の画面13では、寿命による減少分と仮想論理ボリュームへの容量割当分とを分けて示したが、運用管理者はもはや寿命そのものを考慮する必要がなくなるので、寿命分をあえて表示しないようにしてもよい(すなわち線411だけとする)。
この場合、従来のHDDのみ搭載したストレージ装置と同様の運用が可能となる。
また、保守提供者は、書込み量に応じた容量追加の保守費用を運用管理者へ請求できるので、高I/Oで利用するユーザ(寿命不足が大きいユーザ)に対しても、低I/Oで利用するユーザ(寿命不足が小さいまたは無いユーザ)に対しても、公平な保守サービスを提供することができる。
Claims (15)
- ホスト計算機と接続して記憶領域を提供するストレージ装置において、
前記ストレージ装置は、
前記ホスト計算機からのデータを記憶する記憶ドライブを1つ以上有する記憶部と、
前記記憶ドライブの記憶領域を管理する制御部と
を備え、
前記記憶ドライブの内、所定のデータ書換可能量を有する不揮発メモリ記憶ドライブを複数用いてRAIDグループを構成し、当該RAIDグループを1つ以上使用して構成されるプールを1つ以上備え、
前記制御部は、
前記プールを所定のサイズに分割した記憶領域を前記ホスト計算機へ提供し、
前記ホスト計算機から前記プールへのデータ書換量を算出し、
前記算出されたデータ書換量を前記プールでの書換可能量から減算して当該プールの残書換可能量を算出する
ことを特徴とするストレージ装置。 - 請求項1記載のストレージ装置であって、前記制御部は、
前記不揮発メモリ記憶ドライブ毎のデータ書換量から、当該不揮発メモリ記憶ドライブの残書換可能量を算出し、
前記RAIDグループを構成する不揮発メモリ記憶ドライブの残書換可能量から、RAIDグループの残書換可能量を算出し、
前記プールを構成するRAIDグループの残書換可能量から当該プールの残書換可能量を算出する
ことを特徴とするストレージ装置。 - 請求項2記載のストレージ装置であって、前記制御部は、
前記不揮発メモリ記憶ドライブでのデータ書換可能量と、予め設定され格納された前記ストレージ装置の製品寿命年数とに基づき、製品使用年数に対するプールの残書換可能量の想定曲線を生成し、
製品使用年数と、製品使用年数に対するプールの残書換可能量の推移に基づき、当該製品使用年数に対するプールの残書換可能量の実績曲線を生成し、
製品寿命到達時点でのプールの残書換可能量を、前記製品使用年数に対するプールの残書換可能量の実績曲線に基づき算出し、
前記プールの残書換可能量の実績曲線が想定曲線を下回り、前記製品寿命到達時点でのプールの残書換可能量が0を下回るときに、前記プールへ前記不揮発メモリ記憶ドライブの追加または、複数の不揮発メモリ記憶ドライブで構成されるRAIDグループの追加で、前記プールの書換可能量を増加させる
ことを特徴とするストレージ装置。 - 請求項3記載のストレージ装置であって、前記制御部は、前記プールへの前記不揮発メモリ記憶ドライブ追加の代わりに、前記ホスト計算機へ提供していない未使用の記憶領域を、当該ホスト計算機へ提供せず書換可能量の追加として使用し、残書換可能量を増加させる
ことを特徴とするストレージ装置。 - 請求項4記載のストレージ装置であって、前記ストレージ装置を管理する管理端末が当該ストレージ装置に接続し、当該ストレージ装置及び当該管理端末を管理する管理サーバが当該管理端末に接続し、
前記管理端末及び前記管理サーバは、それぞれ情報を表示する表示部を備え、
前記ストレージ装置は、前記算出したプールの残書換可能量、不揮発メモリ記憶ドライブの残書換可能量、プールの残書換可能量の想定曲線、プールの残書換可能量の実績曲線、追加または確保するプールの書換可能量のいずれかの1つ以上の情報を前記管理端末または前記管理サーバに送信し、
前記管理端末または前記管理サーバは、受信した情報を表示部に表示させる
ことを特徴とするストレージ装置。 - 請求項5記載のストレージ装置であって、
前記プールの残書換可能量の実績曲線が想定曲線を下回った場合か、前記製品寿命到達時点でのプールの残書換可能量が0を下回る場合か、前記プールの残書換可能量が予め前記ストレージ装置に格納された警告閾値の量より下回った場合かのいずれかである場合、前記制御部が警告を前記管理端末または前記管理サーバに通知する
ことを特徴とするストレージ装置。 - 請求項3記載のストレージ装置であって、前記不揮発メモリ記憶ドライブまたは前記プールまたは前記RAID毎に、書換可能量及び残書換可能量が管理され、当該書換可能量及び当該残書換可能量に基づき算出された健全性が管理される
ことを特徴とするストレージ装置。 - 請求項5記載のストレージ装置であって、前記管理サーバは、
前記受信した情報に基づき、前記ストレージ装置にプール容量を追加するか否かを判断し、
追加が必要であると判断された場合、追加が必要なプール容量及びRAIDグループ構成を算出して、前記管理端末または前記ストレージ装置に送信する
ことを特徴とするストレージ装置。 - 請求項8記載のストレージ装置であって、前記管理サーバは、
前記算出した追加が必要プール容量及びRAIDグループ構成から必要な保守費用を算出して、当該保守費用の情報を前記管理端末または前記ストレージ装置に送信する
ことを特徴とするストレージ装置。 - ホスト計算機と接続して記憶領域を提供するストレージ装置の保守運用方式において、
前記ストレージ装置は、
前記ホスト計算機からのデータを記憶する記憶ドライブを1つ以上有する記憶部と、
前記記憶ドライブの記憶領域を管理する制御部と
を備え、
前記記憶ドライブの内、所定のデータ書換可能量を有する不揮発メモリ記憶ドライブを複数用いてRAIDグループを構成し、当該RAIDグループを1つ以上使用して構成されるプールを1つ以上備え、
前記制御部は、
前記プールを所定のサイズに分割した記憶領域を前記ホスト計算機へ提供し、
前記ホスト計算機から前記プールへのデータ書換量を算出し、
前記算出されたデータ書換量を前記プールでの書換可能量から減算して当該プールの残書換可能量を算出する
ことを特徴とするストレージ装置の保守運用方式。 - 請求項10記載のストレージ装置の保守運用方式であって、前記制御部は、
前記不揮発メモリ記憶ドライブ毎のデータ書換量から、当該不揮発メモリ記憶ドライブの残書換可能量を算出し、
前記RAIDグループを構成する不揮発メモリ記憶ドライブの残書換可能量から、RAIDグループの残書換可能量を算出し、
前記プールを構成するRAIDグループの残書換可能量から当該プールの残書換可能量を算出する
ことを特徴とするストレージ装置の保守運用方式。 - 請求項11記載のストレージ装置の保守運用方式であって、前記制御部は、
前記不揮発メモリ記憶ドライブでのデータ書換可能量と、予め設定され格納された前記ストレージ装置の製品寿命年数とに基づき、製品使用年数に対するプールの残書換可能量の想定曲線を生成し、
製品使用年数と、製品使用年数に対するプールの残書換可能量の推移に基づき、当該製品使用年数に対するプールの残書換可能量の実績曲線を生成し、
製品寿命到達時点でのプールの残書換可能量を、前記製品使用年数に対するプールの残書換可能量の実績曲線に基づき算出し、
前記プールの残書換可能量の実績曲線が想定曲線を下回り、前記製品寿命到達時点でのプールの残書換可能量が0を下回るときに、前記プールへ前記不揮発メモリ記憶ドライブの追加または、複数の不揮発メモリ記憶ドライブで構成されるRAIDグループの追加で、前記プールの書換可能量を増加させる
ことを特徴とするストレージ装置の保守運用方式。 - 請求項12記載のストレージ装置の保守運用方式であって、前記ストレージ装置を管理する管理端末が当該ストレージ装置に接続し、当該ストレージ装置及び当該管理端末を管理する管理サーバが当該管理端末に接続し、
前記管理端末及び前記管理サーバは、それぞれ情報を表示する表示部を備え、
前記ストレージ装置は、前記算出したプールの残書換可能量、不揮発メモリ記憶ドライブの残書換可能量、プールの残書換可能量の想定曲線、プールの残書換可能量の実績曲線、追加または確保するプールの書換可能量のいずれかの1つ以上の情報を前記管理端末または前記管理サーバに送信し、
前記管理端末または前記管理サーバは、受信した情報を表示部に表示させる
ことを特徴とするストレージ装置の保守運用方式。 - 請求項13記載のストレージ装置の保守運用方式であって、
前記プールの残書換可能量の実績曲線が想定曲線を下回った場合か、前記製品寿命到達時点でのプールの残書換可能量が0を下回る場合か、前記プールの残書換可能量が予め前記ストレージ装置に格納された警告閾値の量より下回った場合かのいずれかである場合、前記制御部が警告を前記管理端末または前記管理サーバに通知する
ことを特徴とするストレージ装置の保守運用方式。 - 請求項14記載のストレージ装置の保守運用方式であって、前記管理サーバは、
前記受信した情報に基づき、前記ストレージ装置にプール容量を追加するか否かを判断し、
追加が必要であると判断された場合、追加が必要なプール容量及びRAIDグループ構成を算出して、前記管理端末または前記ストレージ装置に送信するとともに、
前記算出した追加が必要プール容量及びRAIDグループ構成から必要な保守費用を算出して、当該保守費用の情報を前記管理端末または前記ストレージ装置に送信する
ことを特徴とするストレージ装置の保守運用方式。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/125,740 US10120592B2 (en) | 2014-09-26 | 2014-09-26 | Storage device and maintenance/operation system therefor |
PCT/JP2014/075702 WO2016046977A1 (ja) | 2014-09-26 | 2014-09-26 | ストレージ装置及びその保守運用方式 |
JP2016549877A JP6368789B2 (ja) | 2014-09-26 | 2014-09-26 | ストレージ装置及びその保守運用方式 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2014/075702 WO2016046977A1 (ja) | 2014-09-26 | 2014-09-26 | ストレージ装置及びその保守運用方式 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016046977A1 true WO2016046977A1 (ja) | 2016-03-31 |
Family
ID=55580536
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/075702 WO2016046977A1 (ja) | 2014-09-26 | 2014-09-26 | ストレージ装置及びその保守運用方式 |
Country Status (3)
Country | Link |
---|---|
US (1) | US10120592B2 (ja) |
JP (1) | JP6368789B2 (ja) |
WO (1) | WO2016046977A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11262917B2 (en) | 2020-03-24 | 2022-03-01 | Hitachi, Ltd. | Storage system and SSD swapping method of storage system |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9886203B2 (en) * | 2015-10-22 | 2018-02-06 | International Business Machines Corporation | Shifting wearout of storage disks |
US10528276B2 (en) | 2015-10-22 | 2020-01-07 | International Business Machines Corporation | Shifting wearout of storage disks |
US9921760B2 (en) | 2015-10-22 | 2018-03-20 | International Business Machines Corporation | Shifting wearout of storage disks |
CN107506314B (zh) * | 2016-06-14 | 2021-05-28 | 伊姆西Ip控股有限责任公司 | 用于管理存储系统的方法和装置 |
US11132133B2 (en) | 2018-03-08 | 2021-09-28 | Toshiba Memory Corporation | Workload-adaptive overprovisioning in solid state storage drive arrays |
CN111104047B (zh) * | 2018-10-25 | 2023-08-25 | 伊姆西Ip控股有限责任公司 | 管理冗余磁盘阵列的方法、设备和计算机可读存储介质 |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014141411A1 (ja) * | 2013-03-13 | 2014-09-18 | 株式会社日立製作所 | ストレージシステムおよびストレージシステム制御方法 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05100801A (ja) | 1991-10-09 | 1993-04-23 | Hitachi Ltd | デイスクサブシステムの高信頼化方式 |
JP2010061578A (ja) * | 2008-09-05 | 2010-03-18 | Jds:Kk | 半導体記憶装置の使用限界予想解析用プログラム又はシステム |
CN102449607B (zh) | 2009-07-22 | 2015-05-27 | 株式会社日立制作所 | 具有多个闪存封装的存储系统 |
-
2014
- 2014-09-26 WO PCT/JP2014/075702 patent/WO2016046977A1/ja active Application Filing
- 2014-09-26 JP JP2016549877A patent/JP6368789B2/ja active Active
- 2014-09-26 US US15/125,740 patent/US10120592B2/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014141411A1 (ja) * | 2013-03-13 | 2014-09-18 | 株式会社日立製作所 | ストレージシステムおよびストレージシステム制御方法 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11262917B2 (en) | 2020-03-24 | 2022-03-01 | Hitachi, Ltd. | Storage system and SSD swapping method of storage system |
Also Published As
Publication number | Publication date |
---|---|
US10120592B2 (en) | 2018-11-06 |
JP6368789B2 (ja) | 2018-08-01 |
US20170003891A1 (en) | 2017-01-05 |
JPWO2016046977A1 (ja) | 2017-04-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6368789B2 (ja) | ストレージ装置及びその保守運用方式 | |
US11693568B2 (en) | Workload-adaptive overprovisioning in solid state storage drive arrays | |
US10324633B2 (en) | Managing SSD write quotas in data storage systems | |
US8489844B2 (en) | Storage system providing heterogeneous virtual volumes and storage area re-allocation | |
US8407409B2 (en) | Metrics and management for flash memory storage life | |
US9606728B2 (en) | Controlling a storage system | |
WO2014141411A1 (ja) | ストレージシステムおよびストレージシステム制御方法 | |
WO2012137242A1 (en) | Storage system and data control method therefor | |
WO2013160958A1 (en) | Information storage system and method of controlling information storage system | |
TW201314437A (zh) | 快閃碟陣列及控制器 | |
US10013322B2 (en) | Storage apparatus and storage apparatus control method | |
WO2015052833A1 (ja) | ストレージ装置、ストレージシステム、及びストレージ装置制御方法 | |
JP2008040713A (ja) | 記憶媒体グループの故障を抑えるストレージシステム | |
JP6017032B2 (ja) | ストレージシステム及びストレージ制御方法 | |
CN110413198B (zh) | 用于管理存储系统的方法、设备和计算机程序产品 | |
US8527700B2 (en) | Computer and method for managing storage apparatus | |
JP2011065624A (ja) | 動的ページ再配置ストレージシステム管理 | |
US20230013322A1 (en) | Solid state drive management method and solid state drive | |
US20210223955A1 (en) | Using recurring write quotas to optimize utilization of solid state storage in a hybrid storage array | |
JP2017162260A (ja) | メモリ制御装置、ストレージ装置、ストレージ装置の制御プログラム、及び、ストレージ装置の制御方法 | |
US8578125B2 (en) | Allocation of storage space for critical data sets | |
WO2018154669A1 (ja) | ストレージシステム | |
JP2017199043A (ja) | ストレージ装置とシステム及び方法とプログラム | |
US11726834B2 (en) | Performance-based workload/storage allocation system | |
US9495112B1 (en) | Service level based data storage |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14902571 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15125740 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2016549877 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14902571 Country of ref document: EP Kind code of ref document: A1 |