EP1505504A2 - Remote copy system - Google Patents
Remote copy system Download PDFInfo
- Publication number
- EP1505504A2 EP1505504A2 EP03021756A EP03021756A EP1505504A2 EP 1505504 A2 EP1505504 A2 EP 1505504A2 EP 03021756 A EP03021756 A EP 03021756A EP 03021756 A EP03021756 A EP 03021756A EP 1505504 A2 EP1505504 A2 EP 1505504A2
- Authority
- EP
- European Patent Office
- Prior art keywords
- storage unit
- unit system
- control information
- storage subsystem
- journal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2053—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
- G06F11/2056—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring
- G06F11/2071—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring using a plurality of controllers
- G06F11/2074—Asynchronous techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2053—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
- G06F11/2056—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring
- G06F11/2058—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant by mirroring using more than 2 mirrored copies
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2201/00—Indexing scheme relating to error detection, to error correction, and to monitoring
- G06F2201/855—Details of asynchronous mirroring using a journal to transfer not-yet-mirrored changes
Definitions
- the present invention relates to a storage unit for storing data used by computers. Especially, this invention is concerned with a remote copy technique in which a storage subsystem incorporating a controller for transmission/reception of data to/from a computer and a disk device for storing data is interconnected to another storage subsystem existing at a location remotely distant from the storage subsystem and data is copied between the storage subsystems without routing through the computer so as to be duplicated.
- a technique for asynchronous remote copy is available, according to which a storage unit (hereinafter referred to as a storage subsystem) receiving data from a general computer system, a server, a personal computer on a network or another host computer system (hereinafter referred to as a host) transfers the received data asynchronously to a second storage subsystem installed at a remote location while assuring the data sequence to permit the second storage subsystem to write-in the transferred data.
- Asynchronous transfer referred to herein means that after receiving data from the host and returning to the host a response informing that processing of the received data is completed, the first storage subsystem executes the data transfer to the second storage subsystem.
- a technique for synchronous remote copy is also available, according to which synchronously with an update process of data between a host and a first storage subsystem connected thereto, data is transferred from the first storage subsystem to a second storage subsystem installed near the first storage subsystem or at a location remote therefrom.
- Synchronous transfer referred to herein means that after receiving data from the host and transferring the received data to the second storage subsystem, the first storage subsystem returns a response to the host.
- JP-A-2000-305856 and JP-A-2003-122509 disclose a technique for realizing duplication of data among three or more storage subsystems while assuring the sequence or order of data update by using synchronous remote copy and asynchronous remote copy in combination.
- remote copy between respective two data storing bases (sites) is used plural times in combination to execute remote copy among n sites.
- each site needs a memory capacity which is n times greater at the least than that in the case where each site does not execute the remote copy process, giving rise to extremely raised costs.
- Every site needs a storage subsystem mounted with a remote copy program, so that a plurality of highly functional and expensive storage subsystems are required.
- an intermediate site positioned intermediately of primary and secondary sites and required to participate in execution of the remote copy process in cooperation with both the primary and secondary sites is loaded particularly heavily, with the result that the number of I/O's processible by a storage subsystem of the intermediate site is limited.
- a first storage unit system and a second storage unit system are connected to each other through a third storage unit system.
- the first storage unit system responds to a write request received from a computer to transmit to the third storage unit system a journal having write data received from the computer and address information indicative of a storage position to which the write data is written, thereby writing the journal to the third storage unit system.
- the second storage unit system receives control information issued by the first storage unit system to read and acquire the journal from the third storage system on the basis of the control information. Then, the second storage unit system follows the address information contained in the journal to write the write data contained in the journal to a disk inside the second storage unit system.
- Fig. 1 there is illustrated, in block diagram form, an example of construction of a data processing system (hereinafter also referred to as a remote copy system) according to a first embodiment of the invention.
- a data processing system hereinafter also referred to as a remote copy system
- suffixes "a", "b” and "c"
- suffixes a, b and c will sometimes be omitted.
- the data processing system comprises a primary site 101a (hereinafter will sometimes be called a first site) including a primary host computer (hereinafter simply called a primary host or first host) 102a and a primary storage subsystem 104a (hereinafter will sometimes be called a first storage subsystem), a secondary site (hereinafter will sometimes be called a second site) 101b including a secondary host (hereinafter will sometimes be called a second host) 102b and a secondary storage subsystem (hereinafter will sometimes be called a second storage subsystem) 104b, and an intermediate site (hereinafter will sometimes be called a third site) 101c including an intermediate storage subsystem (hereinafter will sometimes be called a third storage subsystem) 104c.
- a primary site 101a including a primary host computer (hereinafter simply called a primary host or first host) 102a and a primary storage subsystem 104a
- a secondary site hereinafter will sometimes be called a second site
- 101b including a secondary host (hereinafter will
- the storage subsystem 104 will be described by way of example of a disk array device but the storage subsystem 104 is not limited particularly to the disk array device. Further, in each site 101, the host 102 and storage subsystem 104 are not limited particularly in number, provided that the number is one or more. Furthermore, each host 102 is such a computer as a workstation, microcomputer or mainframe computer.
- the storage subsystem 104 has one or more interfaces 110 and is connected to the host 102 via a storage area network (SAN) 103.
- the interface 110 referred to herein is an interface adapted for storage unit, such as SCSI, fiber channel, FICON or ESCON and the kind of interface 110 and the construction of SAN 103 are not limited particularly. In the following description of the present embodiment, an instance will be described in which the fiber channel is used as interface 110.
- the storage subsystems 104 present at different sites 101 are also connected to each other in a similar way through interface 110 (hereinafter, the interface used for connecting the storage subsystems 104 existing at different sites 101 will also be called a remote copy link 116).
- the primary storage subsystem 104a and intermediate storage subsystem 104c may be arranged in the same room or the same building but for security (aiming at preventing both the subsystems from suffering from the same fault at a time), they may be arranged distantly from each other. Since synchronous remote copy is executed between the primary and intermediate storage subsystems 104a and 104c and update of date is carried out synchronously, the distance between the primary and intermediate storage subsystems 104a and 104c may preferably be shorter. The distance should, however, be determined in accordance with an environment such as conditions of location of the subsystems and is not limited particularly. To add, it is said that the propagation delay in communication is typically 5 ⁇ s/km and the distance amounting up to about 100 km will not affect the host 102a considerably.
- the secondary storage subsystem 104b may preferably be installed at a location remotely distant from the primary and intermediate storage subsystems 104a and 104c.
- the storage subsystems 104 may be connected together via a broadband circuit such as ATM, SONET, EWDM or IP by being routed through the fiber channel added with an extender device.
- a broadband circuit such as ATM, SONET, EWDM or IP
- a management console 117 is software for monitoring/managing the storage subsystems 104 and is mounted on a host 102 packaging CPU 202 and memory 205. By taking an event of disasters into account, the management console 117 is installed on a fourth site (management center) 101d and operated on the host in the fourth site.
- the management console 117 is connected to the respective storage subsystems 104 through a management network 111 so as to operate to monitor the status of storage subsystem 104, carry out changing of construction of the data processing system and control the remote copy process.
- the management network 111 is constructed of a network using an IP protocol but it may be a network using a different protocol.
- the remote copy link 116 is a network other than the management network 111 but an IP network, for instance, may be used as management network to permit the management network 111 to be used also as the remote copy link 116.
- the primary storage subsystem 104a and secondary storage subsystem 104b constitute a pair for remote copy.
- a primary volume 109 inside a single primary storage subsystem 104a and a secondary volume 115 inside one or more secondary storage subsystems are related to each other to form remote copy pairs.
- Control of the remote copy is implemented with a program 105 on the storage subsystem 104.
- the programs 105 of the primary and secondary storage subsystems 104a and 104b have equivalent functions and therefore the primary volume 109 and secondary volume 115 may coexist inside the same storage subsystem 104 and in an event of disasters or during maintenance, switching between the primary volume 109 and secondary volume 115 may be executed.
- the primary storage subsystem 104 has, in addition to constituents the ordinary storage subsystem 104 has, remote copy program 105a, configuration information 106a, control information 107a for remote copy queue (RCQ) 114, update information 108 for remote copy and data subjected to remote copy.
- the secondary storage subsystem 104b has, in addition to constituents the ordinary storage subsystem 104 has, remote copy program 105b, configuration information 106b and control information 107b for remote copy queue 114.
- the configuration information 106 has, in addition to construction, product name and address information of each of the primary, intermediate and secondary storage subsystems 104a, 104c and 104b the remote copy system has, a storage destination address of an update request to be stored in the intermediate storage subsystem 104c (hereinafter called a journal) and a storage destination address of queue control information.
- Original information of the configuration information 106 is held by the management console 117 and a replication of the configuration information 106 is transmitted from the management console 117 to the individual storage subsystems 104 via the intermediate storage subsystem 104c or the management network 111 so as to be held in the individual storage subsystems.
- each of the primary and secondary storage subsystems 104a and 104b has a copy of the configuration information 106 so that the remote copy program 105 inside each storage subsystem 104 may consult the configuration information 106 stored in the storage subsystem 104 of its own to execute remote copy.
- Each of the primary and secondary storage subsystems 104a and 104b acquires, from the configuration information 106, position information for an update request stored in the remote copy queue area 114 on the intermediate storage subsystem 104c and position information for information of control of remote copy queue area 114 stored in a control information area 113 also on the intermediate storage subsystem 104c. Then, each of the primary and secondary storage subsystems 104a and 104b uses the acquired position information to transmit a request for read or write to the remote copy queue or control information area inside the intermediate storage subsystem 104c.
- the primary storage subsystem or the secondary storage subsystem updates or consults the control information 107 to write the update request, data or message for remote copy to the intermediate storage subsystem or read it from the intermediate storage subsystem, thereby ensuring that the information as above can be transmitted/received between the primary and secondary storage subsystems.
- storage subsystems 104 serving as owners are determined for individual areas inside intermediate storage subsystem 104c (for example, individual storage areas inside the control information area 113 or remote copy queue area 114) and the update of data is carried out from a storage subsystem 104 representing a solely determined owner.
- the intermediate storage subsystem 104c When the intermediate storage subsystem 104c has a function of exclusive control of storage area, it is desirable that the function be used to maintain consistency of data stored in the control information storage area 113 and remote copy queue area 114. Enumerated as a typical exclusive function is a Reserve command of SCSI. With the Reserve command used, exclusive control in the unit of logical unit (LU) can be implemented and therefore the intermediate storage subsystem 104c can execute the exclusive control process by allotting different LU's to the individual storage subsystems 104. In this case, the intermediate storage subsystem 104c preferably sets the LU size in accordance with necessary sizes of the individual areas. In the present invention, the unit of the storage area to which the exclusive control is applied and the type of exclusive control are not limited particularly and for example, the exclusive control may be done well in the unit smaller than the LU (for example, in the unit of block).
- the control information 107 for remote copy queue includes information for control of the remote copy queue 114.
- Each of the primary and secondary storage subsystems 104a and 104b consults the control information 107 for remote copy queue owned by the unit of its own to change the remote copy queue 114 and updates the control information 113 the intermediate storage subsystem 104c has and then informs the status after the change to a storage subsystem 104 acting as a partner of remote copy pair (from the primary storage subsystem to the secondary storage subsystem or vice versa).
- the update information 108 owned by the primary storage subsystem 104a referred to herein is information concerning remote copy in connection with a write request the primary storage subsystem 104a receives from the primary host 102a and includes write time, write position and size of each write data.
- the update information 108 is created by the primary storage subsystem 104a when the primary storage subsystem 104a receives a write request from the primary host 102a and it corresponds to data also included in an update request transmitted by the primary storage subsystem to the intermediate storage subsystem 104c. It is to be noted that the update information 108 is erased after it is confirmed that the request for update of remote copy issued by the primary storage subsystem 104a is acquired by the secondary storage subsystem 104b. In the event that a fault takes place in the intermediate storage subsystem 104c, the update information 108 is used to perform synchronization of the primary storage subsystem 104a with the secondary storage subsystem 104c.
- the program 105 described so far is installed from a different unit to a storage medium owned by each storage subsystem by using a portable storage medium such as compact disk or opto-magnetic disk or through the medium of the management network 111 and is executed by the CPU each storage subsystem has.
- FIG. 2 An example of construction of the storage subsystem 104 is illustrated in Fig. 2.
- Each storage subsystem 104 includes a controller 201 and one or more disk devices 210.
- the controller 201 has a host adapter 203 for connection to the host, a memory 205, disk adapters 207 for connection to the disk devices, a processor 202 and a network controller 204 for connection to the management system network.
- the number of individual constituent components owned by each storage subsystem is not limited particularly but from the standpoint of the performance and reliability, the individual constituent components may preferably be multiplexed.
- the host adapter 203 performs protocol control in respect of the interface 110 such as fiber channel.
- the disk adapter 207 performs a protocol process in respect of a disk interface 209 such as fiber channel.
- the memory 205 Stored on the memory 205 are data and programs used for the remote copy process. More particularly, in order to implement remote copy, the memory 205 stores the program 105 for remote copy, configuration information 106 for remote copy, control information 107, update information 108 and data 211. In addition to the information as above, programs and control information necessary for control of the storage subsystem 104 and data 211 received from the host 102 are stored on the memory 205. Preferably, for assurance of high reliability, the memory 205 may be doubled and a power supply to the memory 205 may also be doubled.
- the network controller 204 performs protocol control of the management system network 111 and carries out communication between the management console 117 and the storage subsystem 104.
- the disk drive 210 receives a read or write command from the controller 201 through the disk interface 209 and performs a read or write process in accordance with the command.
- the disk interface 209 may preferably be doubled.
- a plurality of disk drives 210 are used in combination to provide a redundant configuration and logical devices (hereinafter called volumes) are formed in the plurality of disk drives. The present embodiment will be described hereunder by using volumes representing logical devices.
- the processor 202 performs a process concerning the storage subsystem 104.
- the processor 202 connects to the host adapter 203, disk adapter 207 and network controller 204 inside the controller 201 through an internal bus 208 to control them. Further, the processor 202 also connects to the memory 205 through the internal bus 208 to execute the program 105 or update the control information 107 by using the configuration information 106 or control information 107 inside the memory 205.
- the present embodiment has been described by way of example of the storage system having a simplified internal construction but the internal construction of the storage subsystem 104 is not limited particularly, provided that a function comparable to that described above is achievable.
- a storage subsystem using a switch in place of the internal bus 208 as disclosed in JP-A-10-333836 may be used.
- FIG. 3 An example of configuration of the storage area arranged in the disk drive of intermediate storage subsystem 104c is depicted in Fig. 3.
- three areas are provided including a configuration information area 112 in which the configuration information 106 is stored, a control information area 113 in which the control information 107 is stored and a remote copy queue area 114 in which update requests are stored.
- the configuration information 106 includes, in addition to construction, product name and address information of each of the primary storage subsystem, intermediate storage subsystem and secondary storage subsystem the remote copy system has, storage positions of update requests inside the remote copy queue area 114 and storage destination addresses of the queue control information 107 inside the control information area 113.
- the control information 107 includes information for control/management of the remote copy queue 144.
- the control information area has two kinds of areas of which one is a primary control information area 301 for storing (primary ⁇ secondary) control information 107 addressed from primary storage subsystem 104a to secondary storage subsystem 104b and the other is a secondary control information area 302 for storing (secondary ⁇ primary) control information 107 addressed from secondary storage subsystem 104b to primary storage subsystem 104a.
- the primary storage subsystem 104a has the ownership of the primary control information area 301 and the secondary storage subsystem 104b has the ownership of the secondary control information area 302, thereby preventing data inconsistency from taking place.
- the configuration information 106 owned by each storage subsystem includes, in addition to addresses and sizes of the individual primary control information area 301 and secondary control information area 302, information indicative of owners of the individual areas, so that remote copy program 105 can confirm the owners by consulting the configuration information so as to perform control such that inconsistency of data does not take place.
- the intermediate storage subsystem 104c allots the primary and secondary control information areas every logical unit (LU) to permit exclusive control to be conducted LU by LU.
- the information stored in the remote copy queue 114 is mainly sorted into two kinds of information of message and data and to store these kinds of information, the remote copy queue 144 has two kinds of data areas of message area and data area. In the following, if the two kinds of areas need not be discriminated particularly from each other, these areas will be described implicitly as being queue 114.
- both pieces of information are stored in the queue 114, including (primary ⁇ secondary) information addressed from primary storage subsystem 104a to secondary storage subsystem 104b and (secondary ⁇ primary) information addressed from secondary storage subsystem 104b to primary storage subsystem 104a.
- the queue 114 there are at least four areas in the queue 114, that is, primary message and data areas 303 and 305 for storing messages and data, respectively, issued from the primary storage subsystem 104a to the secondary storage subsystem 104b, secondary message and data areas 304 and 306 for storing messages and data, respectively, issued from the secondary storage subsystem 104b to the primary storage subsystem 104a.
- the number of storage areas present on the remote copy queue 114 depends on the construction of the remote copy system. For example, in a remote copy system having a plurality of secondary storage subsystems, the number of storage areas present on the remote copy queue 114 will sometimes exceed four.
- the queue area 114 has the area, in which information (messages or data) addressed from primary storage subsystem 104a to secondary storage subsystem 104b is stored, of the owner represented by the primary storage subsystem and the information, addressed from secondary storage subsystem 104b to primary storage subsystem 104a, of the owner represented by the secondary storage subsystem.
- FIG. 4A Examples of structure of data stored in the disk drive of intermediate storage subsystem 104c are depicted in Figs. 4A and 4B.
- the disk drive 210 of intermediate storage subsystem 104c is an ordinary disk drive and the intermediate storage subsystem 104c is likewise an ordinary storage subsystem, information written to the disk drive 210 can be overwritten from another unit. Accordingly, data written to a storage area of disk drive 210 is overwritten by a logical fault (a bug in program) due to the owner or by another unit belonging to an owner different from that of the storage area, leading to the possibility that data becomes erroneous (inconsistent). If the error as above can be detected by means of the intermediate storage subsystem 104c, the error can be dealt with but it is sometimes difficult for the intermediate storage subsystem to detect the data overwrite and logical error by the host 102 or storage subsystem 104 not belonging to the owner.
- a logical fault a bug in program
- protection data is added to data written to the intermediate storage subsystem 104c to assure the security of the data transmitted/received between the primary and secondary storage subsystems 104a and 104b through the intermediate storage subsystem 104c.
- the intermediate storage subsystem 104c is a storage subsystem having block devices such as SCSI. Typically, the block has a size of 512 bytes and the disk device the intermediate storage subsystem 104c has is accessed in the unit of block.
- FIG. 4A Shown in Fig. 4A is an example in which the last area of each block 401 the intermediate storage subsystem 104c has is defined as a protection data area 402.
- the protection data 402 is written to the end of each block, so that source data to be stored in the last area of a block conventionally is stored in a head position (403) of the succeeding block.
- a position (404) heading the succeeding block conventionally follows the storage area 403 and with this method used, the storage position of data is successively shifted backwards.
- the thus provided protection data area causes the data stored within one block conventionally to be stored over a plurality of blocks. Then, in order for data of one block to be stored in the intermediate storage subsystem 104, a plurality of blocks must be updated.
- the block device is accessed in the unit of block and therefore, it is necessary, in this case, to read and write all of related blocks to the memory 205 inside the controller from the disk device, update the data and write it to the disk device 210.
- the data to be stored in the intermediate storage subsystem 104c is written only once and the same data is not read and written, proving that employment of the method shown in Fig. 4A hardly matters.
- FIG. 4B Shown in Fig. 4B is an example in which the block size of intermediate storage subsystem 104c is increased by the amount of protection data and protection data 402 is stored in the thus increased area. With this method used, data added with the protection data 402 can be stored in the disk device without changing the block address.
- protection data 402 is depicted in Fig. 5. As described in connection with Figs. 4A and 4B, the protection data 402 is added to each block stored in the intermediate storage subsystem 104c.
- the protection data 402 is mainly sorted into two kinds of data, of which one is logical information having identifier 501 of request and sequence number 502 in the request and the other is an error detection code 503 for checking an error in block data.
- the former is used to detect a logical error and data falsification by others and the latter is used to detect a data failure of block per se.
- the error detection code 503 may include parity, check sum, ECC, CRC or Hamming code but in the present invention, any type of error detection code can be used, provided that error detection and packaging are possible.
- the protection data 402 is added to data by means of host adapter 203 of the storage subsystem 104 which creates the data and the data is checked by using the protection data by means of host adapter 203 of the storage subsystem 104 which is on the receiving side.
- FIG. 6 there is illustrated an example of a logical structure of data stored in the intermediate storage subsystem 104c.
- Fig. 6 shows information concerning an update request transmitted from primary storage subsystem 104a to secondary storage subsystem 104b in a remote copy group.
- Configuration information 106 stored in the configuration information area 112 includes information of static configuration and the configuration information 106 is updated when the configuration of the remote copy group is changed.
- Each storage subsystem 104 acquires the configuration information 106 from the configuration information storage area 112 of intermediate storage subsystem 104c to develop it in the unit of its own and acquires information concerning remote copy and configuration information concerning the remote copy queue area 114. Then, each storage unit subsystem executes the remote copy and queue control by utilizing the information as above.
- the configuration information 106 is distributed to the individual storage subsystems 104 through the medium of intermediate storage subsystem 104c but alternatively, the configuration information 106 may be transmitted to each storage subsystem 104 directly from the management console 117.
- Shown in Fig. 6 is the configuration information 106 concerning a single remote copy group.
- the remote copy group is constructed of one or more volumes and defined as a group which keeps consistency between individual volumes. Such an operation as suspend, resynchronization and the like applied to the remote copy group can be executed in the unit of volume pair as well as in the unit of group.
- the remote copy group is constructed of a single primary storage subsystem 104a and one or more secondary storage subsystems 104b and is assigned with an identifier (group ID) 601.
- group ID group ID
- the configuration information 106 includes group ID 601, ID 602 of the primary storage subsystem belonging to the remote copy group, ID 604 of the secondary storage subsystem also belonging to the remote copy group and the number 603 of the secondary storage subsystems 104b. Further, the configuration information 106 is stored with the number 605 of areas such as primary message areas and primary data areas (hereinafter called queues) existing in the queue area 114 and separate information 606 of each queue.
- queues primary message areas and primary data areas
- the separate information of queue includes identifier 607 of queue, queue type 608 for identifying which one of message and data a queue is concerned with, head position 609 of the queue in the queue area 114 of intermediate storage subsystem 104c, size 610 of the queue, primary information 611 concerning control information 107 transmitted from the primary storage subsystem 104a to the secondary storage subsystem 104b and secondary information 612 concerning control information 107 transmitted from the secondary storage subsystem 104b to the primary storage subsystem 104a.
- an identifier of the storage subsystem in which the queue area is arranged and a logical address of LU of the queue are also included as the head position 609.
- a plurality of secondary storage subsystems are provided, a plurality of pieces of the information as above are also needed.
- the information 611 or 612 concerning the control information 107 includes identifier 613 of the storage subsystem which is a source of issue of control information, head position 614 indicative of a storage position of the control information in the control information area 113 and size 615 of the control information.
- each storage subsystem 104 specifies a storage area in which the control information 107 is stored to exchange the control information 107.
- control information 107 corresponding to a single queue are one piece of primary control information 631 and one or more pieces of secondary control information 632.
- Each control information 107 is updated by a predetermined sole storage subsystem 104 but the control information can be consulted from a plurality of storage subsystems 104.
- Each control information 107 has identifier 616 of remote copy group, identifier 617 of queue, possessive unit identifier 618 indicative of a unit possessing the control information 107, head position 619 which indicates a head position of data undergoing the remote copy for the purpose of indicating a progress condition of the remote copy process, size 620 of the data during the remote copy and heartbeat 621.
- the group ID 616, queue ID 617 and possessive unit identifier 618 are used to check whether the control area is logically correct.
- Stored in the head position 619 and size 620 are a head position of a storage area in the queue area in which the data subjected to the remote copy process is stored and a size of the storage, respectively.
- each storage subsystem 104 periodically writes information to the heartbeat 621 to thereby inform another unit that the unit of its own is in operation. In other words, by confirming whether or not the heartbeat 621 is updated, it can be decided from the different unit whether the storage subsystem corresponding to the heartbeat is in operation.
- the information written to the heartbeat area 621 has temporally changeable values such as defined by counter values or a timer.
- Data1 to Data6 are stored in the queue area 114. It is also assumed that in Fig. 6, the primary control information 631 on intermediate storage subsystem 104c has its head position 619 indicative of Data1 and its size 620 indicative a size of from Data1 to Data4, so that a request for processing the heading Data1 to the end Data4 (data during process) proceeds. It is further assumed that the secondary control information 632 has its head position 619 indicative of Data3 and its size 620 indicative of a size of Data3, so that a request for processing the Data3 (data during process) proceeds.
- control information 107a owned by the primary storage subsystem 104a has its head position 626 indicative of Data1 and its size indicative of a data size of from Data1 to Data6. Namely, it is assumed that the control information 107a indicates data of from Data1 to Data6 after all and the primary storage subsystem 104a recognizes the Data1 to Data6 as data participating in the remote copy. When the primary storage subsystem 104a writes the next data to the queue area 114, data is written from a position next to Data6 indicated by next position 628 of the control information 107a.
- the secondary storage subsystem 104b has already acquired Data1 and Data2 from the intermediate storage subsystem 104c and the control information 107b owned by the secondary storage subsystem 104b has its head position 629 indicative of Data3 and its size 630 indicative of a data amount of Data3. Accordingly, the secondary storage subsystem 104b recognizes a request for processing the Data3 (data during process).
- the secondary storage subsystem 104b determines that the primary storage subsystem recognizes data from Data1 at the head 624 to Data4 as the request during process (data during process). Accordingly, by consulting the primary control information 631 stored in the intermediate storage subsystem 104c, the secondary storage subsystem 104b can recognize that the Data1 and Data2 which have already been processed and Data3 during process as well as Data4 should be processed.
- the Data5 and Data6 constitute information added from the primary storage subsystem 104a to the queue area 114 after the primary control information 631 on the intermediate storage subsystem 104c has been updated. Since write of data to the queue area and update of the primary control information 631 are executed asynchronously with each other, the primary control information 631 has not yet been updated in respect of the addition of Data5 and Data6 to the queue area 114.
- the latest control information is present in the form of control information 107a inside the primary storage subsystem 104a and when a chance of updating the primary control information 631 comes along from the primary storage subsystem 104a to the intermediate storage subsystem 104c after a lapse of constant time, the primary control information 631 will be updated to the latest information.
- the primary control information 631 and data in the queue area are updated by the primary storage subsystem 104a synchronously or asynchronously with a write request from the host 102.
- the next position 628 of the control information 107a the primary storage subsystem 104a has is updated inside the primary storage subsystem 104a in synchronism with forwarding of an update request from the primary storage subsystem 104a to the queue area 114.
- the head position 626 of control information 107a the primary storage subsystem 104a has is updated by the primary storage subsystem 104a asynchronously with the write request from the host on the basis of the head position 619 of secondary control information 632 the intermediate storage subsystem 104c has.
- the update can be performed for a plurality of write operations in order to decrease traffic.
- the size 627 of control information 107a of the primary storage subsystem 104a can be updated concurrently with update of each of the head position 626 and the next position 628.
- the primary storage subsystem 104a periodically accesses the data stored in the control information area 113 of intermediate storage subsystem 104c. At that time, the primary storage subsystem 104a acquires (reads) the secondary control information 632 and by consulting the head position 619 of secondary control information 632, changes the head position 626 of control information 107a the primary storage subsystem has. For example, in the above example, the head position 626 of control information 107a indicates the Data1 at present but the head position 619 of secondary control information 632 indicates the Data3 and therefore, when accessing the control information area 113 of intermediate storage subsystem 104c, the primary storage subsystem acquires the secondary control information 632 to update the head position 626 of control information 107a of the own unit to Data3.
- the update information 108 stored in the primary storage subsystem 104a can also be erased. Namely, since the primary storage subsystem can recognize, by consulting the secondary control information 632, that the Data1 and Data2 have already been acquired by the secondary storage subsystem, the update information 108 for the Data1 and Data2 can be erased. Any timing can be used for erase, provided that it is after the time that the head position 626 is updated.
- the primary storage subsystem 104a also updates the primary size 627 of control information 107a by calculating it from the head position 626 and the next position 628 the control information 107a of primary storage subsystem 104a has.
- the primary size of control information 107a can indicate the amounts of data of from Data3 to Data6.
- the control information 107a the primary storage subsystem has can be updated to the latest data through the above process and hence, the primary storage subsystem writes it to the control information area 113 of intermediate storage subsystem so as to update the primary control information 631.
- the secondary storage subsystem 104b acquires the primary control information 631 from the intermediate storage subsystem 104c at intervals of constant times and acquires an update request, that is, one of update requests indicated by the primary control information 631 which has not been acquired yet, from the queue area 114. Thereafter, the secondary storage subsystem 104b updates the control information 107b and secondary control information 632 the own unit has.
- the primary control information 631 initially indicates Data1 to Data4 as data during process and the head position 619 of secondary control information 632 indicates Data3. Accordingly, by consulting the primary control information 631, the secondary storage subsystem 104b recognizes that the unit of its own has not yet acquired Data3 and Data4. Then, after acquiring the Data3 and Data4 from the queue area 114, the secondary storage subsystem updates the control information 107b the own unit 104b has and the secondary control information 632 the intermediate storage subsystem 104c has.
- control information is updated such that both the head position 629 of control information 107b and the head position 619 of secondary control information 632 indicate Data4 and both the size 630 of control information 107b and the size 620 of secondary control information 632 indicate 0 (zero).
- the secondary storage subsystem 104b performs only update of the control information.
- the control information is updated such that both the head position 629 of control information 107b and the head position 619 of secondary control information 632 indicate Data3 and both the size 630 of control information 107b and the size 620 of secondary control information 632 indicate a data amount of totaled Data3 and Data4.
- Fig. 7 there is illustrated an example of a data structure of each request 721 issued from the primary or secondary storage subsystem and stored in the queue.
- Header information 722 is stored at the head of a request and end information 724 is stored at the end.
- Stored in the header information 721 are an attribute of the request and separate information thereof including attribute 701, group ID 702 and queue ID 703.
- the attribute of request is used for checking logical errors.
- the separate information of request is sorted into two kinds of information including time series information and positional information.
- the time series information includes serial number of request in the queue (request ID) 704 and time stamp 705 added to the primary storage subsystem 104a or host 102 and is used to align requests in time series sequence inside the secondary storage subsystem 104b and check a dropout of request.
- the positional information includes ID 706 of volume, address 707 inside the volume and size 708 of request and on the basis of the positional information as above, data contained in a request is stored in the secondary storage subsystem 104b. It is to be noted that the positional information contained in the request is the same as that contained in the write request received from the host.
- the size 708 indicates a size of write data received from the host and this size 708 is added with fixed sizes of the header information 721 and end information 724 to provide a size which coincides with a size of request.
- Data 723 is write data received from the host.
- the end information 724 contains an error detection code 709 in addition to header information 722.
- the error detection code 709 is calculated by way of the header information 722 and data 723 and detects an error in the whole of the request.
- a request issued to the intermediate storage subsystem 104c in correspondence with reception of a write request from the host provides a write request applied to the consecutive areas and only one write to the intermediate storage subsystem 104c suffices.
- each request contains both of the control information and write data received from the host 102 as described above and is also called a journal.
- a manager acquires, through the management console 117, information of units constituting the remote copy system, information of storage subsystem 104 such as volumes inside each storage subsystem, information of applications operating on the host 102 existing in the remote copy system and information of a volume used by the host 102 (801).
- the manager determines volume pairs for remote copy and a consistency group assembling volume pairs on the basis of the information as above collected by the manager console and prepares configuration information 106 for remote copy of each storage subsystem 104 and inputs it to the management console 117 (802).
- the management console 117 holds it original information and its copy is transmitted to each storage subsystem 104 via the intermediate storage subsystem 104c or management network 111 (803).
- FIG. 9 there is illustrated an example of an update process executed when the primary storage subsystem 104a receives a write request from the host 102 in the present embodiment.
- the primary storage subsystem 104a receives a write request from the host 102 (901). Then, the primary storage subsystem 104a calculates related volume and address from the write request and performs a check process as below. Firstly, the primary storage subsystem examines whether a remote copy attribute is designated in respect of the volume to which the write request is made (902) and when the remote copy attribute is not designated, the primary storage subsystem receives update data (hereinafter called write data) from the host 102 (911) and reports the host 102 that the process responsive to the write request is ended (909), thus ending the process.
- write data update data
- the primary storage subsystem 104a checks the status of a remote copy pair corresponding to the write object volumes so as to check whether the pair is set up normally (paired status) or whether the pair is split (suspended) (903).
- difference information is recorded (910), update data is received from the host 102 (911) and end of the update process is reported to the host 102 (909), thus ending the process.
- the difference information is held in a format of bit map in which position information of an arbitrary size corresponds to one bit.
- an update log such as reception time of each write request, address of write object storage area, size of write data and write data must be held and as a result, a larger storage capacity than that required for managing the difference information in the form of a bit map is needed. Normally, the suspend lasts for a long time and therefore the difference information is often held in the bit map form.
- the bit map can be stored on the memory 205 of storage subsystem 104 but it may otherwise be stored on the disk 210.
- the primary storage subsystem 104a check the presence or absence of a storage area on the intermediate storage subsystem 104c. Specifically, the primary storage subsystem 104a calculates, from the configuration information 106a and control information 107a the primary storage subsystem has, a remaining storage capacity of the queue area 114 existing in the intermediate storage subsystem 104c (904).
- a sufficiently large queue area 114 is prepared in the intermediate storage subsystem 104c in order that the queue area 114 is not occupied fully with a load of host 102, that is, write request and write data from the host. But an unexpected load will sometimes take place and in such an event, the queue area of intermediate storage subsystem 104c will be exhausted.
- the primary storage subsystem 104a when the queue area of intermediate storage subsystem 104c has no empty capacity, the primary storage subsystem 104a returns a busy report to the host 102 and rejects the reception of the write request (912). Instead of returning the busy to the host 102, the primary storage subsystem 104a may determine that a fault occurs in the intermediate storage subsystem in the step (912) and may change the remote copy state to the suspend state. In this case, the primary storage subsystem 104a carries out a fault coping process as shown in Fig. 11 to shift the remote copy state to suspend, receives update data (write data) from the host to update the bit map and gives an ordinary end report to the host.
- the primary storage subsystem 104a When it is determined in the step 904 that there is an empty capacity in the intermediate storage subsystem 104c, that is, when it is determined in the process of from the steps 902 to 904 that the remote copy process is possible, the primary storage subsystem 104a carries out a process in step 905 and ensuing steps. Firstly, the primary storage subsystem 104a receives write data corresponding to the write request from the host 102 (905).
- the primary storage subsystem 104a prepares a request for update of remote copy corresponding to the write request (906), prepares an update request as shown in Fig. 7 in combination with write data and writes this update request to the next and ensuing positions of the queue area 114 of intermediate storage subsystem 104c (907). At that time, if write of the update request exceeds the last end of queue area 114, data following a superfluous part is written, starting with the head of queue area 114. To add, in writing the update request to the queue area 114, the primary storage subsystem 104a must be controlled such that the head position 629 indicated by the control information 107b the secondary storage subsystem 104b has is not exceeded.
- the primary storage subsystem 104a changes the next position 628 and size 627 of control information 107a by a size of the update request written to the queue area 114 (908) and reports the host 102 of the end (909) to thereafter end the process.
- FIG. 10 there is illustrated an example of an update process of control information executed between the primary storage subsystem 104a and intermediate storage subsystem 104c.
- the primary storage subsystem 104a acquires secondary control information 632 which represents control information of the secondary storage subsystem 104b (1001).
- the primary storage subsystem consults the heartbeat inside the secondary control information 632 to check whether the secondary storage subsystem 104b is in operation. More specifically, the primary storage subsystem 104a conserves the latest one of the heartbeats of secondary storage subsystem acquired in the past in the memory 205 inside the primary storage subsystem 104a and when newly obtaining secondary control information 632 from the intermediate storage subsystem 104c, the primary storage subsystem compares the acquired heartbeat information in secondary control information 632 with the heartbeat information stored in the memory to perform the checking. If the two pieces of heartbeat information have the same value, the secondary storage subsystem is determined not to be in operation.
- the secondary storage subsystem 104b is not determined not to be in operation by determining that the heartbeat information is not updated once but the operating condition of secondary storage subsystem 104b is decided by expanding decision operation to the case where the heartbeat information is not updated through consecutive several check operations or to the case where the heartbeat information is not updated over a constant time or more. For example, when the update period of secondary control information 632 is one second, the secondary storage subsystem is determined not to be in operation if the heartbeat information is not updated over five or more seconds.
- the primary storage subsystem 104a performs a fault coping process as shown in Fig. 11 (1006).
- the primary storage subsystem 104a consults the secondary control information 632 stored in the control information area 113 of intermediate storage subsystem 104c to update control information 107a the own unit has.
- the primary head position 626 of control information 107a is made to be coincident with the secondary head position 619 of secondary control information 632 and the primary size 627 of control information 107a is so set as to cover a size ranging from the primary head position 626 to the primary next position 628.
- the primary storage subsystem 104a uses its new control information 107a to update the primary control information 631 stored in the control information area 113 of intermediate storage subsystem 104c (1003).
- the primary storage subsystem discards the update information 108.
- the update request newly updated by the secondary storage subsystem 104b that is, the update request existing between the head position before the change in the step 1003 and the head position after the change has already been acquired by the secondary storage subsystem. Accordingly, the primary storage subsystem need not hold the update information corresponding to this update request and therefore, the primary storage subsystem make this update information 108 ready to be discarded and after a desired time, that update information is discarded (1004).
- the primary storage subsystem 104a waits for a constant time (for example, one second) before it carries out the process of from step 1001 to step 1004 at intervals of constant times (1005) and again repeats the process starting with the step 1001.
- a constant time for example, one second
- FIG. 11 there is illustrated an example of a process executed by the primary storage subsystem in the event that a fault occurs in the remote copy system.
- the primary storage subsystem 104a When detecting a fault (1101), the primary storage subsystem 104a specifies a portion at which the fault takes place (1102) and reports the management console 117 of the fault (1103). The management console 117 receiving the fault report blocks the faulty portion on the basis of a command from the manager (1104). The faulty portion is blocked not only by the management console 117 but also by the primary storage subsystem 104 which has detected the fault.
- the primary storage subsystem After blocking the faulty portion, the primary storage subsystem acquires the status of a communication path of remote copy and checks whether there is an alternating path (1105). In the presence of the alternating path, the process for remote copy can continue and therefore the primary storage subsystem 104 switches the path for remote copy to the alternating path and thereafter ends the fault process.
- the primary storage subsystem In the absence of the alternating path, the primary storage subsystem changes the status of remote copy pair. Even when the primary storage subsystem 104a determines in the step 904 that no empty area exists in the queue area of intermediate storage subsystem 104c and deems the nonexistence of empty area in the queue area as the occurrence of a fault, such a case is handled as the absence of alternating path.
- the primary storage subsystem 104 If the primary storage subsystem 104 is detecting a fault in the case of nonexistence of the alternating path, it changes the paired state by itself (1106) and registers position information of write data not stored in the secondary storage subsystem 104b (1107). Alternatively, the management console 117 may command the primary storage subsystem to perform a process for changing the pair state and preparing the difference information (bit map).
- the pair state is changed from the normal state (paired state) before fault detection to the state (suspend) in which the pair is split.
- the bit map is prepared on the memory 205 of primary storage subsystem 104a or on the disk drive 210. It is to be noted that at the time that preparation of the bit map is started, the request number concerning update request in each group is also stored in the storage subsystem.
- the primary storage subsystem 104a prepares a bit map headed by an unreflected remote copy update request. For example, when the suspend state takes place at the time that the remote copy system is in the state shown in Fig. 6, the primary storage subsystem 104a prepares a bit map to store the difference information concerning Data1 to Data6 in the bit map and after initializing the bit map, the primary storage subsystem prepares a bit map for write requests received from the host 102 after then (910).
- difference information is stored in the bit map on the primary storage subsystem 104a through the process of step 1107 each time that the primary storage subsystem 104a receives a write request and write data from the host 102a.
- the process can be proceeded with while the primary storage subsystem receives an IO request of the host 102 and continues a process.
- the process shown in the step 1107 also consumes much time if carried out for all of remote copy pairs at a time but can be performed for each remote copy pair within a sufficiently short time.
- a distribution process can be proceeded with among remote copy pairs and hence the bit map update process can be performed while continuing the IO process of host 102.
- FIG. 12 there is illustrated an example of a process in which the secondary storage subsystem 104b acquires the control information and updates it.
- the secondary storage subsystem 104b acquires the primary control information 631 from the control information area 113 of intermediate storage subsystem 104c (1201).
- the secondary storage subsystem carries out the fault coping process shown in Fig. 13 (1209).
- the secondary storage subsystem When the state is normal, that is, when no fault is detected, the secondary storage subsystem first examines the presence or absence of an update request (1203). The secondary storage subsystem can grasp, from a difference between the request during the remote copy process indicated by the primary head position 619 and primary size 620 of the acquired primary control information 631 and the request during remote copy process indicated by the head position 629 and size 630 of the control information 107b the unit of its own has, an update request added to the queue area 114 after the time that the unit of its own updated the control information 107b previously.
- the secondary storage subsystem ends the control information update process and waits for a constant time (1208).
- the secondary storage subsystem 104b acquires that update request from the queue area (1204).
- the secondary storage subsystem 104b acquires all update requests from the queue 114 at a time.
- the contents of the update request is analyzed inside the secondary storage subsystem 104b and the data transmitted from the primary storage subsystem is reflected on the volume in the secondary storage subsystem 104b in accordance with the update request (1205). Namely, data 723 contained in the update request is stored in a storage area indicated by an address 707 contained in the update request.
- control information 107b the secondary storage subsystem has is updated by an amount of the update request corresponding to the updated data (specifically, the head position 629 and secondary size 630 are updated) (1206) and the secondary control information 632 stored in the control information area 113 of intermediate storage subsystem 104c is updated by using the contents of the control information 107b after update (1207).
- the secondary storage subsystem 104b divides the update request for the purpose of acquiring and processing it. Firstly, the secondary storage subsystem 104b prepares an area for update request acquisition and acquires the update request starting with the head position. This area is an area preparable by the secondary storage subsystem 104b, having 10MB, for instance.
- the secondary storage subsystem acquires an update request of only the initial 1MB portion from the intermediate storage subsystem and analyzes header information corresponding to that portion to reflect only data stored in the head portion upon the volume inside the secondary storage subsystem. Then, only in connection with the head portion of the processed update request, the secondary head position 629 and secondary size 630 of control information 107b are changed to update the control information 113 on the intermediate storage subsystem 104c. It will be appreciated that, in connection with the remainder of the update request, the secondary storage subsystem will process it during acquisition of the next update request.
- FIG. 13 there is illustrated an example of a process executed in the secondary storage subsystem 104b in the event of the occurrence of a fault.
- the secondary storage subsystem 104b acquires, as much as possible, update requests remaining on the intermediate storage subsystem 104c.
- the primary storage subsystem 104a is synchronous with the intermediate storage subsystem 104c and therefore when a fault occurs only in the primary storage subsystem 104a and access to the intermediate storage subsystem 104c is possible, the secondary storage subsystem 104b can prevent data extinguishments by acquiring the update requests stored in the intermediate storage subsystem 104c.
- the secondary storage subsystem 104b first detects the fault occurring in the primary storage subsystem from, for example, a heartbeat of primary control information 631 acquired from the intermediate storage subsystem 104c (1301). Then, the secondary storage subsystem 104b examines the status of intermediate storage subsystem 104c (1302).
- the bit map is created starting from the time that the fault takes place.
- the secondary storage subsystem 104b has acquired update requests up to Data2 to end the remote copy process and hence difference information is stored in the bit map in respect of information of Data2 and ensuing data.
- the secondary storage subsystem 104b stores in the bit map the difference information concerning other pieces of update information which have been updated by the secondary storage subsystem 104b after the Data2.
- the secondary storage subsystem 104b Available as the case where data is updated and the difference information is prepared in the secondary storage subsystem 104b are an instance in which for maintenance, for instance, the primary storage subsystem is placed in suspend condition and the secondary host 102b carries out business affairs by using the secondary storage subsystem and an instance in which after a fault takes place in the primary storage subsystem as in the present embodiment, business affairs are taken over to the secondary host 102b and the secondary host issues an input/output request to the secondary storage subsystem.
- information of the secondary storage subsystem 104b is settled up to Data2 and in write operation from the secondary host 102, difference information is held as update information for Data2 and ensuing data.
- the secondary storage subsystem 104c checks the control information area 113 of intermediate storage subsystem 104c to examine whether requests not acquired by the secondary storage subsystem exist therein (1303). In the presence of a request in the intermediate storage subsystem, the secondary storage subsystem 104b acquires update requests from the queue area 114 and reflects data on the volume of secondary storage subsystem 104b on the basis of the acquired update request (1304).
- the secondary storage subsystem 104b increments the head position of control information 107b by the amount of reflected update requests (1305) and again consults the control information area 113 of intermediate storage subsystem 104c to check the presence or absence of the next request (1306).
- methods can be considered including (1) one method in which after acquiring only a header portion of the update request and analyzing it, the secondary storage subsystem acquires data portion representing a main part of the update request and information and (2) the other method in which the secondary storage subsystem 104b acquires data from the update request by a constant size and analyzes it inside the secondary storage subsystem.
- the present invention can be implemented with either method.
- the secondary storage subsystem first acquires a header portion of the queue request and examines group ID, request ID and time stamp for remote copy.
- the secondary storage subsystem analyzes by using the group ID and queue ID whether there is any logical contradiction, checks by using the request ID whether the update request has a serial number next to that of the immediately preceding update request and examines the time stamp to check whether the time stamp has a larger value than that of the immediately preceding update request. If the result of the examination shows that there is no logical inconsistency, the secondary storage subsystem determines that the update request is present but in the case of the occurrence of logical inconsistency, nonexistence of the update request is determined.
- the secondary storage subsystem acquires, if possible, all update requests from the intermediate storage subsystem 104c and inspects each of the blocks and end information. If the inspection results are normal, the secondary storage subsystem reflects data on its volume (1307).
- the last written update request suspends in the midst of write operation. Accordingly, each block and the end information are inspected to assure consistency of the update requests.
- the secondary storage subsystem again increments the address (1305) so as to again check whether the next update request is present in the intermediate storage subsystem 104c (1306).
- the secondary storage subsystem repeats the process ranging from the step 1305 to the step 1307 until any update request does not exist in the intermediate storage subsystem.
- the secondary storage subsystem 104b consults the control information 113 to collectively acquire Data3 and Data4 from the queue area 114 of intermediate storage subsystem in the step 1304. Thereafter, the secondary storage subsystem analyzes, in the steps 1305 to 1307, the header and end information of each update request in respect of Data5 and Data6 and if no logical inconsistency is found, it acquires Data5 and Data6. Then, when trying to acquire an update request next to Data6, the secondary storage subsystem finds inconsistency of information in any of header, data and end information and ends the acquisition of the update request.
- FIG. 14 there is illustrated an example of a resynchronization process between the primary storage subsystem 104a and the secondary storage subsystem 104b.
- the manager uses the management console 117 to acquire the state of each unit and the connection state between the units so as to examine whether resynchronization is placed in permissible condition (1401).
- permissible condition (1401)
- an error process is executed (1414).
- the manager also checks logical consistency of the data processing system by using the management console.
- the manager uses the management console to confirm if volumes belonging to the group of remote copy are usable in respect of all storage subsystems 104 and comparison of time for detection of a fault and comparison of time for start of bit map are made among individual groups and among individual volumes. By comparing times for start of bit maps used for the resynchronization process, the manager can confirm the logical consistency of the data processing system (1403).
- the secondary storage subsystem 104b fails to be stored with information equivalent or comparable to that in the primary storage subsystem 104a owing to such a situation as the case where, for example in the step 1302 of Fig. 13, the secondary storage subsystem is determined not to be permitted to access the intermediate storage subsystem (1310). Accordingly, as described above, the primary storage subsystem 104a needs to prepare a bit map in respect of an update request whose acquisition and processing by means of the secondary storage subsystem 104b are not confirmed.
- the primary storage subsystem 104a must start acquisition of a bit map from Data1.
- data inconsistency takes place when the resynchronization process proceeds between the primary and secondary storage subsystems.
- inconsistency takes place such that Data3, Data4, Data5 and Data6 cannot be resynchronized.
- bit map preparing time difference information acquisition start time
- the manager When inconsistency of data is determined not to occur, the manager carries out the resynchronization process by first creating paths to bring the individual pairs for remote copy into resynchronization, on the basis of the status of the system acquired on the management console 117 (1404). At that time, it is not always necessary that the path be routed through the intermediate storage subsystem 104c and the primary storage subsystem 104a and the secondary storage subsystem 104b may be connected directly to each other.
- the manager prepares new configuration information 106 in the management console and the management console 117 transmits the new configuration information 106 to each storage subsystem 104 (1405).
- each storage subsystem 104 tries to connect to a designated unit through a designated path.
- the data contents is decided which one of the storage subsystems 104 the data contents should be synchronized (1407).
- business affairs are temporarily transferred to the secondary storage subsystem 104b because of maintenance or when business affairs are transferred from the primary storage subsystem 104a to the secondary storage subsystem 104b because of the occurrence of a fault, resynchronization must be made from the secondary storage subsystem 104b to the primary storage subsystem 104a.
- resynchronization of data from the primary storage subsystem 104a to the secondary storage subsystem 104b is needed.
- the data contents is usually synchronized with the system side to which business affairs are transferred or the system side which has successively executed business affairs.
- a bit map the secondary storage subsystem has is forwarded from the secondary storage subsystem 104b to the primary storage subsystem 104a (1408) and a bit map of the primary storage subsystem 104a and the bit map of the secondary storage subsystem 104b are consolidated in the primary storage subsystem 104a to create a bit map for resynchronization.
- the consolidation of the bit maps can be achieved by calculating the logical sum (OR) of individual bits indicated on both the bit maps.
- update requests which have not yet been reflected on the secondary storage subsystem are transmitted from the primary storage subsystem to the secondary storage subsystem, thereby executing the resynchronization process (1409).
- the storage subsystem 104 can resume a process for IO process request received from the host 102 even in the course of resynchronization by using a technique of "Copy on write", for instance (1410).
- the resynchronization process from the secondary storage subsystem 104b to the primary storage subsystem 104a can be executed similarly to the resynchronization process from the primary storage subsystem to the secondary storage subsystem, with the only exception that the direction of forwarding of the bit map (1411) and the direction of transmission of the update request (1412) are reversed.
- a remote copy queue 114 exists on an intermediate storage subsystem 104c but configuration information 106 and control information 107 are not stored in the intermediate storage subsystem 104c.
- the configuration information 106 is forwarded from a management console 117 to primary and secondary storage subsystems.
- the control information 107 is exchanged directly between the primary storage subsystem 104a and the secondary storage subsystem 104b without being routed through the intermediate storage subsystem 104c.
- an update request is transmitted from the primary storage subsystem 104a to the intermediate storage subsystem 104c in synchronism with write of data from a primary host 102a to the primary storage subsystem 104a.
- the control information 107 is transmitted from the primary storage subsystem 104a directly to the secondary storage subsystem 104b asynchronously with a write request issued form the host 102.
- the secondary storage subsystem 104b acquires the update request from the intermediate storage subsystem 104c and stores write data acquired from the update request in a disk of secondary storage subsystem, thereby realizing remote copy.
- a system having a plurality of intermediate storage subsystems 104 is exemplified as still another embodiment of the data processing system according to the invention.
- a plurality of intermediate sites 101 exist, including an intermediate site A and an intermediate site B.
- there are two kinds of carrying-out modes in which (1) an intermediate storage subsystem 104 to be used is allotted to each remote copy group and (2) in a remote copy process for a desired remote copy group, a plurality of intermediate storage subsystems 104c and 104d are used.
- mode (1) the process is similar to that in the first embodiment from the viewpoint of each remote copy group.
- each update request has an ID which is a sole serial number through out the remote copy groups and intermediate storage subsystems 104c and 104d.
- the secondary storage subsystem 104b aligns the update requests acquired from the plurality of intermediate storage subsystems 104c and 104d and thereafter, reflects data on volumes of a disk device the own unit has in accordance with the update requests.
- the present embodiment is characteristic of a fault coping process or a process during the occurrence of a fault.
- the plurality of intermediate storage subsystems 104c and 104d are provided in the present embodiment and therefore, even when any one of the storage subsystems 104 becomes faulty, the alternating path control can be accomplished by reconstructing the configuration information of the data processing system. In performing the alternating path control process, it is necessary that an update request stagnating on an intermediate storage subsystem 104 suffering from the occurrence of the fault be retransmitted from the primary storage subsystem to the secondary storage subsystem through the medium of a different indefectible intermediate storage subsystem 104.
- the update request to be retransmitted is one the primary storage subsystem has written to the intermediate storage subsystem in which the fault occurred, there is a possibility that an ID of the retransmitted update request is not contiguous to an ID of an update request the primary storage subsystem 104a transmitted immediately precedently to the intermediate storage subsystem. But, it does not matter if update requests acquired from the intermediate storage subsystem are aligned in order of request ID's on the secondary storage subsystem side and thereafter a process for storing data in the volumes is executed.
- a remote copy link 1701 for coupling a primary storage subsystem 104a and a secondary storage subsystem 104b is provided.
- the remote copy link 1701 is usable as (1) an alternating path for remote copy when an intermediate storage subsystem 104c becomes faulty or (2) as a path for control information communication.
- FIG. 18 there is illustrated yet still another embodiment of the data processing system according to the invention having a plurality of secondary storage subsystems 104.
- configuration information 106 and control information 107 are also needed for each of the plurality of secondary storage subsystems 104b and 104e.
- one to plurality control is needed for update requests transmitted from a primary storage subsystem 104a to the secondary storage subsystems 104b and 104e through an intermediate storage subsystem, and secondary control information 632 and secondary information 612 stored in the intermediate storage subsystem are needed by the number of secondary storage subsystems 104.
- the secondary storage subsystem 104 performs control similar to that in the first embodiment.
- the process as shown in Fig. 10 to be executed on the primary storage subsystem 104a is changed.
- the primary storage subsystem 104a acquires secondary control information 632 of the plurality of secondary storage subsystems 104b and 104e and in process 1003, compares pieces of secondary control information 632 of all secondary storage subsystems 104b and 104e with each other.
- the head position of a secondary storage subsystem 104 for which the update is determined to be the most retarded (oldest) as a result of comparison is used as a new head position to prepare new control information 107a.
- the primary storage subsystem proceeds with the process similarly to cause the difference information acquisition start time to meet the secondary storage subsystem 104 for which the update is the most retarded, in bit map creation of the step 1107. Through this, effective difference information can be held between the primary storage subsystem 104a and the secondary storage subsystems 104b and 104e.
- the update requests are transmitted from the secondary storage subsystems 104 to the primary storage subsystem 104a in one-to-one correspondence relation and hence control operation similar to that in the first embodiment is executed.
- the ordinary intermediate storage subsystem for practicing the I/O process is arranged in the intermediate site for connecting the primary and secondary storage subsystems and a process related to remote copy is carried out by the primary storage subsystem at primary site and the secondary storage subsystem at secondary site.
- the intermediate storage subsystem is an ordinary storage subsystem for executing a read or write process in accordance with an I/O request from the primary or secondary storage subsystem during the remote copy process.
- the primary and secondary storage subsystems execute the remote copy process through the intermediate storage subsystem.
- the primary and secondary storage subsystems exchange the data and messages therebetween.
- the primary storage subsystem When receiving a write request and write data from the host, the primary storage subsystem transmits an update request having the write data and control information contained in the write request, such as a storage position of the write data, to the intermediate storage subsystem and thereafter returns a response to the host (synchronous remote copy). Besides, the primary storage subsystem also transmits to the intermediate storage subsystem the queue control information including a pointer such as head position indicative of progress of the remote copy queue 114 and a size of the update request.
- the control information can be updated either synchronously or asynchronously with the write request by the host but preferably the update may be carried out asynchronously at intervals of constant times for the purpose of improving the performance.
- the primary and secondary storage subsystems may be connected directly to each other so as to exchange the control information directly between them without routing through the intermediate storage subsystem.
- the secondary storage subsystem at the secondary site reads the primary control information written to the intermediate storage subsystem and acquires update requests including write data from the queue on the intermediate storage subsystem on the basis of the primary control information. To add, the secondary storage subsystem acquires the update requests asynchronously with transmission of the update requests from the primary storage subsystem to the intermediate storage subsystem (asynchronous remote copy). Then, on the basis of position information contained in the acquired update requests, the secondary storage subsystem stores write data contained in the acquired update requests into the disk inside the secondary storage subsystem.
- exchange of the update request and control information is done by causing the primary storage subsystem and secondary storage subsystem to issue a command for read or write to the intermediate storage subsystem and causing the intermediate storage subsystem to execute a read or write process in accordance with the issued command.
- the intermediate storage subsystem does not require functions and programs concerning the remote copy, so that the n-site remote copy can be implemented at low costs.
- JBOD Just a Bunch of Service
- the intermediate storage subsystem requires the storage area for storing the update requests and control information but it need not keep holding copies of data the primary and secondary storage subsystems have and therefore the storage capacity necessary for the intermediate storage subsystem can be less than that of the primary or secondary storage subsystem. Accordingly, with less storage capacity, the remote copy process can be executed between n sites.
- the intermediate storage subsystem must, in general, execute a remote copy process between it and each of the primary and secondary storage subsystems and is imposed with a heavy process load but in the embodiments of the invention, the process to be executed by the intermediate storage subsystem is simplified and it suffices that the intermediate storage subsystem processes the I/O request issued from the primary or secondary storage subsystem and hence the load on the intermediate storage subsystem can be reduced. Furthermore, such a process as checking the progress of remote copy and status thereof is not necessary for the intermediate storage subsystem to execute and therefore, the monitor/management process required for the remote copy can be alleviated.
- the remote copy between n sites can be executed at low costs. Besides, according to the invention, the load and capacity necessary for the process of executing the remote copy between n sites can be reduced.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Quality & Reliability (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims (20)
- A system comprising:a first storage unit system (104a) connected to a computer (102a) and having a first disk device (210a) and a first controller (201a);a second storage unit system (104b) having a second disk device (210b) and a second controller (201b); anda third storage unit system (104c) connected to said first storage unit system and said second storage unit system and having a third storage area and a third controller (201c),
stores said write data in said first disk device (210a), and
returns a response to said write request to said computer (102a) after transmitting said journal, and
wherein said second controller (201b) receives first control information (107a) issued by said first controller (201a) and including a storage position of said journal used when said second storage unit system acquires said journal,
acquires said journal from said third storage unit system (104c) on the basis of said first control information (107a), and
stores said write data in said second disk device (210b) on the basis of the address information included in said journal. - A system according to claim 1,
wherein said first storage unit system (104a) stores said first control information (107a);
wherein said second controller (201b) acquires said journal and thereafter issues second control information (107b) indicative of the acquisition of said journal; and
wherein said first controller (201a) acquires said second control information (107b) and thereafter makes said first control information (107a) stored in said first storage unit system ready to be discarded. - A system according to claim 2,
wherein said first controller (201a) transmits said first control information (107a) to said third storage unit system (104c) and said second controller (201b) acquires said first control information (107a) stored in said third storage area (113) from said third storage unit system (104c); and
wherein said second controller (201b) transmits said second control information to said third storage unit system and said first controller (201a) acquires said second control information stored in said third storage area (113) from said third storage unit system (104c). - A system according to claim 3,
wherein said third storage unit system stores said first control information and said second control information in different logical volumes inside said third storage area, respectively, and makes such setting that a write request from either of said first and second storage unit systems (104a, 104b) is to be permitted in respect of each logical volume. - A system according to claim 3,
wherein in the event that a fault takes place in said first storage unit system, said second storage unit system consults the first control information (107a) stored in said third storage unit system to acquire a journal having write data not stored in said second disk device (201b) from said third storage unit system, and stores write data included in the acquired journal in said second disk device (210b) on the basis of address information included in said acquired journal. - A system according to claim 5,
wherein when receiving a write request from a computer (102b) connected to said second storage unit system after a fault takes place in said first storage unit system, said second storage unit system (104b) has difference information indicative of a storage position of write data written in accordance with said write request; and
wherein when said first storage unit system recovers from the fault, said second storage unit system transmits the data stored at the storage position indicated by said difference information to said first storage unit system through a communication path (116) connecting said first storage unit system and said second storage unit system. - A system according to claim 2 further comprising a communication path (1701) connected to said first storage unit system and said second storage unit system,
wherein said first controller (201a) transmits said first control information (107a) to said second storage unit system through said communication path (1701); and
wherein said second controller transmits said second control information (107b) to said first storage unit system through said communication path (1701). - A system according to claim 7,
wherein when a fault takes place in said third storage unit system (104c), said first controller (201a) transmits write data received from said computer (102a) to said second controller (201b) through said communication path (1701), and said second controller stores the write data received through said communication path (1701) in said second disk device (210b). - A system according to claim 3 further comprising a fourth storage unit system (104d) connected to said first storage unit system (104a) and said second storage unit system (104b),
wherein when a fault takes place in said third storage unit system, said first storage unit system and said second storage unit system transmit/receive therebetween a journal, first control information (107a) and second control information (107b) through said fourth storage unit system (104d). - A system according to claim 3 further comprising a fourth storage unit system (104d) connected to said first storage unit system and said second storage unit system,
wherein said first controller (201a) transmits a journal having time information to either of said third storage unit system and said fourth storage unit system; and
wherein said second controller (201b) acquires the journal from said third storage unit system or said fourth storage unit system and writes write data included in the acquired journal to said second disk device (210b) in order of times indicated by the time information assigned to said journal. - A remote copy method for use in a system having a first storage unit system (104a) connected to a computer (102a) and having a first disk device (210a), a second storage unit system (104b) having a second disk device (210b) and a third storage unit system (104c) connected to said first storage unit system and said second storage unit system, said method being executed among said first storage unit system, second storage unit system and third storage unit system and comprising:a write request receiving step of causing said first storage unit system to receive a write request and write data from said computer (102a);a journal writing step of causing said first storage unit system to write a journal having said write data and address information included in said write request to said third storage unit system;a first control information issuing step of causing said first storage unit system to issue first control information (107a) including a storage position of said journal necessary for said second storage unit system to read said journal;a first control information acquiring step of causing said second storage unit system to acquire said first control information;a journal reading step of causing said second storage unit system to read said journal on the basis of said first control information; anda write data writing step of causing said second storage unit system to store the write data included in said journal in a disk device (210b) owned by said second storage unit system in accordance with the address information included in said journal.
- A remote copy method according to claim 11,
wherein said first storage unit system stores said first control information,
said method further comprising:a second control information issuing step of causing said second storage unit system to issue, after said journal reading step, second control information (107b) indicating that said second storage unit system reads said journal; anda second control information receiving step of causing said first storage unit system to receive said second control information, - A remote control method according to claim 11,
wherein said first control information issuing step has a step of writing said first control information into said third storage unit system; and
wherein said first control information acquiring step has a step of reading said first control information from said third storage unit system. - A remote copy method according to claim 13,
wherein said third storage unit system stores said first control information and said second control information in different logical volumes inside a third storage area (113), respectively, and makes such setting that a write request from either of said first and second storage unit systems is to be permitted in respect of each logical volume. - A remote copy method according to claim 13 further comprising:a step of causing, in the event that a fault takes place in said first storage unit system, said second storage unit system to acquire from said third storage unit system a journal having write data not stored in said disk device (210b) owned by said second storage unit system by consulting said first control information stored in said third storage unit system; anda step of causing said second storage unit system to store, on the basis of address information included in the acquired journal, write data included in said acquired journal in said disk device (210b) said second storage unit system has.
- A remote copy method according to claim 15 further comprising:a step of causing said second storage unit system (104b) to receive a write request from a computer (102b) connected to said second storage unit system in the event that a fault takes place in said first storage unit system (104a);a step of causing said second storage unit system to hold difference information indicative of a storage position of write data written in accordance with said write request; anda step of causing, when aid first storage unit system recovers from the fault, said second storage unit system to transmit the data stored at the storage position indicated by said difference information to said first storage unit system through a communication path (116) connecting said first storage unit system and said second storage unit system.
- A remote copy method according to claim 11,
wherein said first storage unit system and said second storage unit system are connected to each other by a communication path (1701);
wherein said first control information issuing step has a step of causing said first storage unit system to transmit said first control information (107a) to said second storage unit system through said communication path (1701); and
wherein said first control information acquiring step has a step of causing said second storage unit system to receive said first control information from said first storage unit system through said communication path (1701). - A remote copy method according to claim 17 further comprising, when a fault takes place in said third storage unit system:a step of causing said first storage unit system to transmit write data received from said computer (102a) to said second storage unit system through said communication path (1701); anda step of causing said second storage unit system to store the write data received via said communication path (1701) in said disk device (210b) said second storage unit system has.
- A remote copy method according to claim 13,
wherein said system further comprises a fourth storage unit system (104d) connected to said first storage unit system (104a) and said second storage unit system (104b) ; and
wherein when a fault takes place in said third storage unit system (104c), said first and second storage unit systems transmit/receive a journal and first control information therebetween through said fourth storage unit system (104d). - A remote copy method according to claim 13,
wherein said system further comprises a fourth storage unit system (104d) connected to said first storage unit system and said second storage unit system;
wherein said journal writing step has a step of causing said first storage unit system to write a journal having time information to either of said third and fourth storage unit systems; and
wherein said second controller (201b) acquires the journal from said third and fourth storage unit systems and writes write data contained in the acquired journal to said second disk device in order of times indicated by said time information assigned to said journal.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2003205617A JP4419460B2 (en) | 2003-08-04 | 2003-08-04 | Remote copy system |
JP2003205617 | 2003-08-04 |
Publications (2)
Publication Number | Publication Date |
---|---|
EP1505504A2 true EP1505504A2 (en) | 2005-02-09 |
EP1505504A3 EP1505504A3 (en) | 2009-04-29 |
Family
ID=33549891
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP03021756A Withdrawn EP1505504A3 (en) | 2003-08-04 | 2003-09-25 | Remote copy system |
Country Status (4)
Country | Link |
---|---|
US (1) | US20050033828A1 (en) |
EP (1) | EP1505504A3 (en) |
JP (1) | JP4419460B2 (en) |
CN (1) | CN1285036C (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1783595A2 (en) * | 2005-10-05 | 2007-05-09 | Hitachi, Ltd. | Information processing system, control method for information processing system, and storage system |
WO2012139559A1 (en) * | 2011-04-13 | 2012-10-18 | Continental Automotive Gmbh | Method for storing a security-relevant data unit |
Families Citing this family (54)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7133982B2 (en) * | 2003-06-18 | 2006-11-07 | International Business Machines Corporation | Method, system, and article of manufacture for consistent copying of storage volumes |
US7475134B2 (en) * | 2003-10-14 | 2009-01-06 | International Business Machines Corporation | Remote activity monitoring |
JP2006221487A (en) * | 2005-02-14 | 2006-08-24 | Hitachi Ltd | Remote copy system |
US8032726B2 (en) * | 2003-12-03 | 2011-10-04 | Hitachi, Ltd | Remote copy system |
US7437389B2 (en) * | 2004-03-10 | 2008-10-14 | Hitachi, Ltd. | Remote copy system |
US7724599B2 (en) * | 2003-12-03 | 2010-05-25 | Hitachi, Ltd. | Remote copy system |
US7085788B2 (en) | 2003-12-03 | 2006-08-01 | Hitachi, Ltd. | Remote copy system configured to receive both a write request including a write time and a write request not including a write time. |
JP4477370B2 (en) * | 2004-01-30 | 2010-06-09 | 株式会社日立製作所 | Data processing system |
JP2005258850A (en) * | 2004-03-12 | 2005-09-22 | Hitachi Ltd | Computer system |
US7366727B2 (en) * | 2004-03-12 | 2008-04-29 | Sap Aktiengesellschaft | Management of inbound conflicts when merging data of distributed systems |
JP4382602B2 (en) * | 2004-04-23 | 2009-12-16 | 株式会社日立製作所 | Remote copy system |
JP4563725B2 (en) * | 2004-05-17 | 2010-10-13 | 株式会社日立製作所 | Computer system |
JP4477950B2 (en) * | 2004-07-07 | 2010-06-09 | 株式会社日立製作所 | Remote copy system and storage device system |
US7330861B2 (en) | 2004-09-10 | 2008-02-12 | Hitachi, Ltd. | Remote copying system and method of controlling remote copying |
JP4915775B2 (en) * | 2006-03-28 | 2012-04-11 | 株式会社日立製作所 | Storage system and storage system remote copy control method |
JP4355674B2 (en) * | 2005-03-17 | 2009-11-04 | 富士通株式会社 | Remote copy method and storage system |
JP4566808B2 (en) * | 2005-04-21 | 2010-10-20 | 株式会社日立製作所 | Information processing system, storage system, and information processing method |
JP2006323663A (en) * | 2005-05-19 | 2006-11-30 | Nec Corp | Information processing system, replication method, and difference information holding device and program |
JP4752334B2 (en) * | 2005-05-26 | 2011-08-17 | 日本電気株式会社 | Information processing system, replication auxiliary device, replication control method, and program |
JP4963808B2 (en) * | 2005-08-05 | 2012-06-27 | 株式会社日立製作所 | Storage control system |
JP2007047892A (en) | 2005-08-08 | 2007-02-22 | Hitachi Ltd | Computer system and status management method for computer system |
JP4756992B2 (en) * | 2005-10-25 | 2011-08-24 | 株式会社日立製作所 | Storage system and storage control method |
JP4942371B2 (en) * | 2006-03-23 | 2012-05-30 | 株式会社日立製作所 | Storage system and data management method |
US20070234118A1 (en) * | 2006-03-30 | 2007-10-04 | Sardella Steven D | Managing communications paths |
US9383938B2 (en) * | 2006-04-18 | 2016-07-05 | International Business Machines Corporation | Method, system, and apparatus for re-conveying input/output operations utilizing a sequential-access data storage device secondary communication port |
JP4842720B2 (en) * | 2006-06-29 | 2011-12-21 | 株式会社日立製作所 | Storage system and data replication method |
US7546485B2 (en) * | 2006-08-15 | 2009-06-09 | Hewlett-Packard Development Company, L.P. | Method and system for efficient journal-based resynchronization |
JP5244332B2 (en) | 2006-10-30 | 2013-07-24 | 株式会社日立製作所 | Information system, data transfer method, and data protection method |
JP4935901B2 (en) | 2007-05-01 | 2012-05-23 | 富士通株式会社 | Storage system, storage device, and remote copy method |
JP2009003499A (en) * | 2007-06-19 | 2009-01-08 | Hitachi Ltd | File sharing system and method for generating single logical directory configuration by using file sharing system |
JP4824638B2 (en) * | 2007-06-27 | 2011-11-30 | 株式会社日立製作所 | Multi-host management server, program thereof, and path information management method in storage system |
US8341459B2 (en) * | 2007-08-01 | 2012-12-25 | Brocade Communications Systems, Inc. | Data migration without interrupting host access and with data lock for write access requests such that held write access requests do not expire |
JP4959630B2 (en) * | 2008-05-23 | 2012-06-27 | 株式会社東芝 | Data storage |
JP5147570B2 (en) | 2008-07-02 | 2013-02-20 | 株式会社日立製作所 | Storage system and remote copy recovery method |
JP5422147B2 (en) * | 2008-07-08 | 2014-02-19 | 株式会社日立製作所 | Remote copy system and remote copy method |
US9805082B2 (en) * | 2008-09-10 | 2017-10-31 | Sap Se | Recording information about an item |
US8195956B2 (en) * | 2009-08-17 | 2012-06-05 | Brocade Communications Systems, Inc. | Re-keying data in place |
JP2013012137A (en) * | 2011-06-30 | 2013-01-17 | Nec System Technologies Ltd | Disk mirroring system and its management method |
JP5782925B2 (en) * | 2011-08-31 | 2015-09-24 | 富士通株式会社 | Information processing apparatus, program, and control method |
US9003081B2 (en) * | 2012-05-18 | 2015-04-07 | Hitachi, Ltd. | Computer system and control method thereof |
US9923762B1 (en) * | 2013-08-13 | 2018-03-20 | Ca, Inc. | Upgrading an engine when a scenario is running |
CN103843309A (en) * | 2013-11-22 | 2014-06-04 | 华为技术有限公司 | Data synchronizing method and data synchronizing system |
JP6203087B2 (en) * | 2014-03-11 | 2017-09-27 | 三菱電機株式会社 | Air conditioner, air conditioner system, and rewrite control program |
JP6354507B2 (en) | 2014-10-01 | 2018-07-11 | 富士通株式会社 | Storage control device and storage system |
US9836368B2 (en) * | 2015-10-22 | 2017-12-05 | Netapp, Inc. | Implementing automatic switchover |
CN105430063A (en) * | 2015-11-05 | 2016-03-23 | 浪潮(北京)电子信息产业有限公司 | Method for remote copying between multi-control shared memory systems |
US11061735B2 (en) * | 2019-01-02 | 2021-07-13 | Alibaba Group Holding Limited | System and method for offloading computation to storage nodes in distributed system |
EP4081909A4 (en) * | 2019-12-27 | 2023-09-13 | Hitachi Vantara LLC | Hybrid cloud asynchronous data synchronization |
US20230289322A1 (en) * | 2020-07-22 | 2023-09-14 | Jared Ringenberg | Data Analysis and Data Forensics System and Method |
JP2022047912A (en) * | 2020-09-14 | 2022-03-25 | 株式会社東芝 | Magnetic disk device |
US11797236B2 (en) | 2021-10-27 | 2023-10-24 | EMC IP Holding Company LLC | Near continuous data protection without using snapshots |
US11914866B2 (en) * | 2021-10-27 | 2024-02-27 | EMC IP Holding Company LLC | Optimized one pass delta streaming filter |
US11822432B2 (en) | 2021-10-27 | 2023-11-21 | EMC IP Holding Company LLC | Near continuous data protection without using snapshots |
US11899537B2 (en) | 2021-10-27 | 2024-02-13 | EMC IP Holding Company LLC | Near continuous data protection without using snapshots |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030014432A1 (en) * | 2001-07-13 | 2003-01-16 | Sun Microsystems, Inc. | Storage network data replicator |
EP1283469A2 (en) * | 2001-08-08 | 2003-02-12 | Hitachi, Ltd. | Data storage system with remote copy control method |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6209002B1 (en) * | 1999-02-17 | 2001-03-27 | Emc Corporation | Method and apparatus for cascading data through redundant data storage units |
US6225076B1 (en) * | 1999-09-15 | 2001-05-01 | The Rockefeller University | Crystal of bacterial core RNA polymerase and methods of use thereof |
US6766359B1 (en) * | 1999-12-29 | 2004-07-20 | Emc Corporation | Method and apparatus for utilizing multiple paths in a file transfer utility employing an intermediate data storage system |
US6598174B1 (en) * | 2000-04-26 | 2003-07-22 | Dell Products L.P. | Method and apparatus for storage unit replacement in non-redundant array |
US6457109B1 (en) * | 2000-08-18 | 2002-09-24 | Storage Technology Corporation | Method and apparatus for copying data from one storage system to another storage system |
US6804690B1 (en) * | 2000-12-27 | 2004-10-12 | Emc Corporation | Method for physical backup in data logical order |
JP3714184B2 (en) * | 2001-03-29 | 2005-11-09 | 富士通株式会社 | Copying method between data areas of storage device and storage system |
JP2004259079A (en) * | 2003-02-27 | 2004-09-16 | Hitachi Ltd | Data processing system |
US7065589B2 (en) * | 2003-06-23 | 2006-06-20 | Hitachi, Ltd. | Three data center remote copy system with journaling |
-
2003
- 2003-08-04 JP JP2003205617A patent/JP4419460B2/en not_active Expired - Fee Related
- 2003-09-25 US US10/669,325 patent/US20050033828A1/en not_active Abandoned
- 2003-09-25 EP EP03021756A patent/EP1505504A3/en not_active Withdrawn
- 2003-09-26 CN CNB031598625A patent/CN1285036C/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030014432A1 (en) * | 2001-07-13 | 2003-01-16 | Sun Microsystems, Inc. | Storage network data replicator |
EP1283469A2 (en) * | 2001-08-08 | 2003-02-12 | Hitachi, Ltd. | Data storage system with remote copy control method |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1783595A2 (en) * | 2005-10-05 | 2007-05-09 | Hitachi, Ltd. | Information processing system, control method for information processing system, and storage system |
EP1783595A3 (en) * | 2005-10-05 | 2009-06-24 | Hitachi, Ltd. | Information processing system, control method for information processing system, and storage system |
WO2012139559A1 (en) * | 2011-04-13 | 2012-10-18 | Continental Automotive Gmbh | Method for storing a security-relevant data unit |
Also Published As
Publication number | Publication date |
---|---|
CN1581105A (en) | 2005-02-16 |
US20050033828A1 (en) | 2005-02-10 |
JP2005055948A (en) | 2005-03-03 |
CN1285036C (en) | 2006-11-15 |
EP1505504A3 (en) | 2009-04-29 |
JP4419460B2 (en) | 2010-02-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1505504A2 (en) | Remote copy system | |
US9753663B1 (en) | Triangular asynchronous replication | |
US7197616B2 (en) | Virtual ordered writes | |
US7167963B2 (en) | Storage system with multiple remote site copying capability | |
US6912629B1 (en) | System and method for restoring data from secondary volume to primary volume in a data storage system | |
US8924665B2 (en) | Virtual ordered writes transfer log | |
US7386668B2 (en) | Handling data writes copied from a remote data storage device | |
US7228456B2 (en) | Data recovery for virtual ordered writes for multiple storage devices | |
US6360306B1 (en) | Relocation of suspended data to a remote site in a distributed storage system | |
EP1632843A2 (en) | Data processing system and copy processing method thereof | |
US20040193816A1 (en) | Reading virtual ordered writes at a remote storage device | |
US20020026603A1 (en) | Method and apparatus for maintaining data coherency | |
US8707003B1 (en) | Rate control for asynchronous data replication |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL LT LV MK |
|
17P | Request for examination filed |
Effective date: 20060331 |
|
PUAL | Search report despatched |
Free format text: ORIGINAL CODE: 0009013 |
|
AK | Designated contracting states |
Kind code of ref document: A3 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL LT LV MK |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06F 11/20 20060101AFI20090320BHEP |
|
AKX | Designation fees paid |
Designated state(s): DE FR GB IE |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20091030 |