Redpaper Alex Osuna Roland Tretau IBM System Storage N series MetroCluster Introduction The IBM® System Storage™ N series MetroCluster extends Clustered Failover capabilities from a primary to a remote site. It replicates data from the primary site to the remote site to ensure that the data there is completely up-to-date and available. If Site A goes down, MetroCluster allows you to rapidly resume operations at a remote site minutes after a disaster: Stretch MetroCluster provides a disaster recovery option at distances up to 500 meters between each N series system (Figure 1 on page 2). – Available on N5000 & N7000 series. Fabric MetroCluster provides a disaster recovery option at distances up to 100 km using a Fibre Channel switched network. – Available on N5000 series In this IBM Redpaper publication, we discuss the following: MetroCluster Features & Benefits MetroCluster Functionality Failure Scenarios © Copyright IBM Corp. 2006. All rights reserved. ibm.com/redbooks 1 Figure 1 shows a MetroCluster configuration. Site A Site B Stretch Fabric X Vol-X Ym Mirrored Vol-Y Xm Mirrored Vol-X Y Vol-Y Figure 1 MetroCluster configurations MetroCluster Benefits The benefits of using MetroCluster are: MetroCluster is designed to be a simple-to-administer solution that extends fail-over capability from within a data center to a remote site. MetroCluster is also designed to provide replication of data from the primary site to a remote site, which helps to keep the data at the remote site current. The combination of failover and data replication aids in the recovery from disaster — helping prevent loss of data — in less time than otherwise possible. Extends Clustered Failover capabilities from a primary to a remote site. Replicates data from the primary site to the remote site to ensure that data there is completely up-to-date and available. If Site A goes down, MetroCluster allows you to rapidly resume operations at a remote site minutes after a disaster. N series Business Continuity Solutions The environment or business need determines the necessary configuration for you. 2 IBM System Storage N series MetroCluster Protection levels MetroCluster provides three protection levels: Within the data center Campus distances WAN distances In the following section, we further discuss the protection levels. Within the Data Center For protection of power faults, node failure, and network connector failure, a Cluster Failover may be the right protection for you, as shown in CFO Figure 2. Campus distances For protection from building, power, node, volume, switch, and network connector failures, the Stretch MetroCluster might be the right configuration for you. By creating a scenario where failover occurs to another building or installation on the campus, availability is increased even one step further. WAN distances For those of you whose business requires the highest level of availability, the MetroCluster Fabric configuration delivers the needed options. By making available different synchronous versions of SnapMirror®, the Stretch MetroCluster configuration protects against location, building, power, node, volume, switch, and network connector failures. Figure 2 shows the protection levels. IBM System Storage WAN Distances Campus distances Async SnapMirror Primary Data Center Within DataCenter Clustered Failover (CFO) • High system protection MetroCluster (Stretch) • Cost effective zero RPO protection • Most cost effective with RPO from 10 min. to 1 day MetroCluster (Fabric) • Cost effective zero RPO protection Sync SnapMirror • Most robust zero RPO protection 5 1 N Series Metro Cluster © 2006 IBM Corporation Figure 2 Protection levels IBM System Storage N series MetroCluster 3 What is MetroCluster MetroCluster software provides an enterprise solution for high availability over wide area networks (WAN). MetroCluster deployments of IBM N series storage systems are used for: Business continuance Disaster recovery Achieving recovery point and recovery time objectives MetroCluster technology is an important component of enterprise data protection strategies. If a disaster occurs at a source site, businesses can continue to run and access data from a clustered node in a remote site. The primary goal of MetroCluster is to provide mission-critical applications with redundant storage services in case of site-specific disasters. It is designed to tolerate site-specific disasters with minimal interruption to mission-critical applications and zero data loss by synchronously mirroring data between two sites. A MetroCluster system is made up of the following components and requires the following licenses: Multiple storage controllers and high availability (HA) configuration Provides automatic failover capability between sites in case of hardware failures. SyncMirror® Provides an up-to-date copy of data at the remote site. The data is ready for access after failover without administrator intervention. Cluster remote Provides a mechanism for the administrator to declare site disaster and initiate a site failover using a single command, for ease of use. FC switches Provide storage system connectivity between sites that are greater than 500* meters apart. MetroCluster allows the active/active configuration to be spread across data centers up to 100 kilometers apart (see Figure 3 on page 5). In the event of an outage at one data center, the second data center can assume all affected storage operations, that are lost, with the original data center. SyncMirror is required as part of MetroCluster to ensure that an identical copy of the data exists in the second data center, should the original data center be lost: MetroCluster and SyncMirror extend active/active clustering across data centers that are up to 100 kilometers apart. Either dark-fiber or Dense Wave Division Multiplexing (DWDM) between the switches is required. MetroCluster and SyncMirror provide the highest level of storage resiliency across a local region. Highest levels of regional storage resiliency ensure continuous data availability in a particular geography. 4 IBM System Storage N series MetroCluster Figure 3 shows how MetroCluster allows the active/active configuration to be spread across data centers that are up to 100 kilometers apart. IBM System Storage MetroCluster is a unique, cost-effective synchronous replication solution for combined high availability and disaster recovery within a campus or metro area Major Data Center Nearby Office LAN/SAN Stretch MetroCluster provides Campus DR protection - Can stretch up to 500m N Series Fabric MetroCluster provides Metropolitan DR protection Disks - Can stretch up to 100km with FC switches 5 1 Configurations N Series Metro Cluster © 2006 IBM Corporation Figure 3 MetroCluster The benefits of MetroCluster: Synchronous Mirroring with SyncMirror Synchronous Mirroring for MetroCluster mirrors the Write Anywhere File Layout (WAFL®) volumes (aggregates). Both copies or plexes are updated synchronously on writes, which insures consistency. The design of IBM N series and MetroCluster provides data availability even in the event of multiple disk failures. Read performance is optimized by performing application reads from both plexes (Figure 4 on page 6). Performance is directly affected by the distance between plexes. One of the additional performance benefits of Synchronous mirroring is the low write overhead and latencies. Mirroring is done closer to the hardware, and there is Zero impact on the secondary controller. IBM System Storage N series MetroCluster 5 Figure 4 illustrates a MetroCluster with mirrored plexes. IBM System Storage FC FC x Ple0 Plex1 6 1 RAID Groups N Series Metro Cluster RAID Groups © 2006 IBM Corporation Figure 4 MetroCluster with mirrored plexes MetroCluster: SyncMirror configuration Use the sysconfig –r command (Figure 5) to see the volume/plex/raidgroup relationship in Data ONTAP®. Î M irrored volum e Volume spiel (online, normal, mirrored) (zoned checksums) Plex /spiel/plex0 (online, normal, active) RAID group /spiel/plex0/rg0 RAID Disk --------parity data Device -----------3.2 3.3 HA SHELF BAY --------------3 0 2 3 0 3 CHAN ---FC:A FC:A Used (MB/blks) -------------8579/17570816 8579/17570816 Phys (MB/blks) -------------8683/17783112 8683/17783112 CHAN ---FC:A FC:A Used (MB/blks) -------------16979/34774016 8579/17570816 Phys (MB/blks) -------------17560/35964296 17560/35964296 Plex /spiel/plex2 (online, normal, active) RAID group /spiel/plex2/rg0 RAID Disk --------parity data Device -----------6.8 6.9 HA SHELF BAY --------------6 1 0 6 1 1 7 1 N Series Metro Cluster Figure 5 Volume relationships 6 IBM System Storage N series MetroCluster © 2006 IBM Corporation Use aggr mirror to start mirroring the plexes, as we show in Example 1. Example 1 aggr mirror syntax aggr mirror <atv-name> [-n] [-f] [-v <victim-aggr>] [-d <disk_name1> ... <disk_n ameN>] - add a mirror to aggregate or traditional volume <atv-name> MetroCluster: Cluster_Remote License The Cluster_Remote License provides features that enable the administrator to declare a site disaster and initiate a site failover using a single command: Enables the cf forcetakeover -d. – Initiates a takeover of the local partner even in the absence of a quorum of partner mailbox disks. Provides the ability for an administrator to declare a site-specific disaster and have one node takeover its partner’s identity without quorum of disks. Root volumes of both filers MUST be synchronously mirrored. Only synchronously mirrored aggregates are available during a site disaster. Requires administrator intervention as a safety precaution against a ‘split brain’ scenario (cf forcetakeover –d). Important: Site-specific disasters are not the same as a Normal Cluster Failover! MetroCluster components MetroCluster is an integrated solution made up of the following components: Clustered Failover Provides the high-availability failover capability between sites. The administrator controls the failover decision depending on the impact of the disaster. SyncMirror Provides an up-to-date copy of data at the remote site. You can access the data after failover, without administrator intervention. Cluster _Remote Provides a mechanism for an administrator to declare a site disaster and initiate a site failover using a single command, for ease of use. FC Switches Provides filer connectivity between sites that are greater than 500 meters apart, which enables sites to be located at a safe distance away from each other. IBM System Storage N series MetroCluster 7 MetroCluster configurations Table 1 lists the MetroCluster configurations. Table 1 MetroCluster configurations Configuration Supported platform Switch Mode operation Number of dedicated switches Drives supported capacity Shelves per loop Maximum number of shelves Short Distance <500 Meters N5000 N7000 series N/A N/A SATA or FC 6 Same as N5XXX A20 and N7XXX A20 Long Distance 500 Meters to 100KM N5000 series Fabric 2 per site (4 per MetroClust er pair) FC 2 24 (12 each side) Table 2 Switch distances Switch Up to 500m <=10km <=10km <=50km <=80km IBM 2005-B16 Brocade 200E Standard SFPs Vendor: Finisar Vendor #: FTLF-1319P1xCL Vendor: Finisar Vendor #: FTLF-1419P1BCL Vendor: Finisar Vendor #: FTLF-1419P1BCL Vendor: Finisar Vendor #: FTLF-1519P1BCL IBM 2005-H16 Brocade 3850 IBM 2005-H08 Brocade 3250 IBM 2109-F16 Brocade 3800 IBM 3534-F08 Brocade 3200 Table 3 shows the supported N series Gateway configurations. 8 IBM System Storage N series MetroCluster Table 3 Supported Gateway Configurations Gateway Model Data ONTAP Versions Switches Storage N5500 N5200 7.2.3 Brocade 48000, 7500, 5000, 4900, 4100, 200E with switch FW 5.2.1b HDS TagmaStore FW 50-09-06, 50-08-06, 50-07-64 HDS Lightning 9980V/9970V FW 21-14-33 HDS Lightning 9960V/9910 FW 01-19-99 IBM DS4700, DS4800 FW 6.23.05.00 N5500 N5200 N/A (Stretch MetroCluster) N5500 N5200 Cisco MDS 9506, 9509 with switch FW 3.0(2) HDS TagmaStore FW 50-09-06, 50-08-06, 50-07-64 HDS Lightning 9980V/9970V FW 21-14-33 HDS Lightning 9960V/9910 FW 01-19-99 N7800 N7600 N5600 N5300 McData 6140 FW v9.02.01 N7800 N7600 N5600 N5300 Brocade 48000, 7500, 5000, 4900, 4100, 200E with switch FW 5.2.1b HDS TagmaStore FW 50-09-06, 50-08-06, 50-07-64 HDS Lightning 9980V/9970V FW 21-14-33 HDS Lightning 9960V/9910 FW 01-19-99 IBM DS4700, DS4800 FW 6.23.05.00 N7800 N7600 N5600 N5300 N/A (Stretch MetroCluster) Stretch MetroCluster and Fabric MetroCluster The IBM N series supports two configurations: Stretch and Fabric. We cover both configurations in this section. Stretch Stretch MetroCluster provides a disaster recovery option at distances up to 500 meters between each N series system. It only requires the MetroCluster license and the SyncMirror license. To connect the N series nodes over a distance greater than 30m, you need to convert the IB-copper to fiber using feature 1042 - Copper-Fiber Converter. This adaptor has multi fiber Push-On (MPO) connectors and needs a MTP plug (NVRAM5). What is MTP/MPO? Optical links specify the LC duplex connector for 1X links, the MTP/MPO for 4X links, and dual MTP/MPO for 12X links. All of these optical connectors are accepted as ad hoc industry standards for other optical data, communication protocols as well. IBM System Storage N series MetroCluster 9 Stretch is available on N5000 & N7000 series. Fabric Fabric MetroCluster provides a disaster recovery option at distances up to 100km using a Fibre Channel switched network. Available on N5000 and N7000 series Fabric MetroCluster configuration details First of all you will need to make sure you have the right Storage Initiator HBA’s see Table 4. Platform N5200,N5500 N5300,N5600 N7000 Min Data ONTAP Version Feature Code 1006 Yes No 7.1h2 Feature Code 1014 No Yes 7.2.3 Feature Code 1029 No Yes 7.2.3 Table 4 Fabric MetroCluster Storage Initiator HBA’s Onboard ports 0a,0b,0c and 0d can be used The heartbeat and NVRAM uses a separate Dual-port MetroCluster HBA. This HBA is dedicated and not used for the IO traffic of SyncMirror see Table 5. Table 5 Fabric MetroCluster Interconnect HBA’s Platform N5200,N5500 N5300,N5600 N7600,N7800 Min Data ONTAP Feature Code 1018 Yes No 7.1.H12 Feature Code 1032 No Yes 7.2.3 Additional Requirements for Fabric MetroCluster support Must use Clustered models No-Charge SyncMirror License on each storage controller EXN2000 or EXN4000 storage only (EXN1000 Fabric MetroCluster HBAs In this section we discuss 1042 - Copper-Fiber Converter (1042). To connect the filer heads over a distance greater than 30 m, for a stretched cluster you need to convert the IB-copper to fiber using feature 1042 - Copper-Fiber Converter. This adaptor has MPO connectors and needs a MTP plug (NVRAM5). 10 IBM System Storage N series MetroCluster Table 6 lists the Fabric MetroCluster Storage Initiator HBAs. Table 6 Fabric MetroCluster Storage Initiator HBAs Platform N5200, N5500 N5300,N5600 N7600,N7800 Min Data ONTAP Feature code 1006a Yesb No No 7.1h2 a. Fabric MetroCluster only 1006 supported, Stretch MetroCluster all Storage Initiator HBAs supported. b. Can use onboard ports 0a,0b,0c,0d. Table 7 shows the Fabric MetroCluster Interconnect HBAs. Table 7 Fabric MetroCluster Interconnect HBAs Platform N5200,N5500 N5300,N5600 N7600,N7800 Min Data ONTAP Feature Code 1018 Yes No No 7.1.H12 Stretch MetroCluster: Campus Distance (<500m) In a stretch MetroCluster, the controllers and expansion shelves are attached to Fibre Channel switches, and the switches have GBICs to communicate across the WAN to one another. SyncMirror is built into MetroCluster in that every write is written to two separate expansion units to two separate aggregate groups. The advantage of MetroCluster is that it can take a high availability solution and stretch it outside of the frame. Think about taking a disk subsystem, such as the IBM DS4800, separating the controllers miles apart, and maintaining two separate disk groups to ensure failover, instead of building two completely separate DS4800 systems and synchronously mirroring between the two. The benefits of Stretch MetroCluster, Campus Distance are: Stretch MetroCluster provides a disaster recovery option at distances of up to 500 meters between each IBM N series system. See Figure 6 on page 12. Stretch MetroCluster only requires the MetroCluster license and the SyncMirror license. Stretch MetroCluster is available on N5000 and N7000 series. IBM System Storage N series MetroCluster 11 Figure 6 shows the MetroCluster at less than 500 meters. Figure 6 MetroCluster <500M You can use SATA disks either in mirrored configuration or non-mirrored configuration (with one of the nodes only) on a Stretch MetroCluster. Fabric MetroCluster: Metro Distance (>500m) The benefits of MetroCluster, Metro Distance are: Fabric MetroCluster provides a disaster recovery option at distances up to 100km using a Fibre Channel switched network (Figure 7) Fabric MetroCluster uses switches for longer distance disaster recovery solutions. 12 IBM System Storage N series MetroCluster Figure 7 shows an example of the Fabric MetroCluster. Figure 7 Fabric MetroCluster Using DWDM switches Dense Wave Division Multiplexing is a method of multiplexing multiple channels of fiber optic-based protocols, such as ESCON®, Fibre Channel, FICON®, and Gbit Ethernet, onto physical cable by assigning different wavelengths of light (that is colors) to each channel and then fanning it back out at the receiving end. The major players in the enterprise class DWDM marketplace are: Nortel Networks, Cisco (ONS 15540), and Lucent. Dense Wave Division Multiplexors are data link Layer 2 tools. Thus, the typical DWDM machine does not perform any switching, routing, or protocol conversion. IBM System Storage N series MetroCluster 13 Figure 8 depicts a Fabric MetroCluster installation using DWDM. Figure 8 DWDM environment Fabric attach MetroCluster Switch Matrix Table 8 is shows the required Firmware and Data ONTAP levels for San switches. Table 8 MetroCluster Switch Matrix Data ONTAP level Switch Model IBM 2105-H08/Brocade 3250 7.1h2 7.2 or above Minimum Firmware level 5.2.0a IBM 2105-H16/Brocade 3850 IBM 3534-F08/Brocade 3200 3.3.1b IBM 2109-F16/Brocade 3800 IBM 2005-B16/Brocade 200E 5.1.0 Cable selection Use Table 9 on page 15 to determine if the desired cable length is within the supported maximum specification. We provide this table for guidance to determine the distances; however, the best way to determine the best cable type is by testing the cable length in the environment (because there are many variables such as types of cables, panels, and so forth), for example, a client needs to run an FC cable over approximately 260 m (850 ft.) and wants to run it at 2 Gbps. How do we determine the maximum distance that can be allowed 14 IBM System Storage N series MetroCluster for cabling between nodes? Table 9 is a cable selection chart than can help you determine if the cable length you want is within the supported maximum specification. Table 9 Cable selection chart Cable type Fiber core type Mode Wave length Maximum distance (m) Attenuation (Db/Km) Maximum channel attenuation Splice loss Connector pair loss OM2 50/ 125 um Multi 850 550 3.00 3.25 0.3 0.75 OM3 50/ 125 um Multi 850 550 3.00 3.50 0.3 0.75 OS1 9/125 um Single 1310 2000 0.40 7.80 0.3 0.75 OM2 50/ 125 um Multi 850 550a 3.00 2.62 0.3 0.75 OM3 50/ 125 um Multi 850 550 3.00 3.25 0.3 0.75 OS1 9/125 um Single 1310 2000 0.40 7.80 0.3 0.75 OM2 50/ 125 um Multi 850 150 3.00 2.06 0.3 0.75 OM3 50/ 125 um Multi 850 150 3.00 3.00 0.3 0.75 OS1 9/125 um Single 1310 500 0.40 7.80 0.3 0.75 1Gbps 2Gbps 4Gbps IB 1X 250MB/sec OM2 50/ 125 um Multi 850 250 3.50 2.38 0.3 0.75 OM3 50/ 125 um Multi 850 500 3.50 3.25 0.3 0.75 a. According to the maximum channel attenuation (2.62 dB) and an attenuation of 3.00 dB/km, the maximum distance for this cable type is 406 m, so be careful with longer distances (up to 550 m) with OM2 cable at 2 Gbps. Table 9 summarizes data that is related to optical cabling for data communications, which is available in documents that are published by various standards organizations. We focus on data that is relevant to fiber deployments, which are supported on IBM N series systems. To determine whether the desired cable run length is within the supported maximum specification: 1. Determine the needed transfer rate based on the type of shelf you use. 2. Find out what fiber type is installed for the system. 3. Determine the number of connectors in the path between the nodes. Refer to Table 9 and locate the maximum supported operating distance, compare this value to the distance for the actual intended application, and verify that the actual does not exceed the supported maximum. IBM System Storage N series MetroCluster 15 Consider the 260 m distance for the above client. Let us assume: Desired transfer rate: 2 Gbps Fiber core type: 50/125 OM2 Multimode cabling Number of connector pairs: two Table 9 on page 15 shows that 260 m is within the operating maximum distance for all Fibre Channel transfer rates. This also assumes that the fiber connection is point-to-point with only source and destination connections and no patch panels or splices. MetroCluster best practices The following list contains MetroCluster best practices: Read the Data ONTAP 7.2 Active/Active Configuration Guide. Read Chapter 3: “Installing a MetroCluster” in the Data ONTAP 7.2 Active/Active Configuration Guide. Ensure correct cabling!!! Ensure correct switch configuration, and apply 16port and fabric licenses. Test the installation and fail-over scenarios. Train the staff. N series resiliency Figure 9 gives you an idea of what events the N series can protect against with MetroCluster. Event Does th e Event T rigger a Failover? Single Disk F ailure No Yes Double Disk Failure (sam e RAID group) No Yes, with no CFO necessary Yes, with no CFO necessary Triple Disk Failure (sam e RAID group) No Single HBA (initiator) failure, LO OP A No Yes, with no CFO necessary Single HBA (initiator) failure, No Yes, with no CFO necessary Single HBA initiator failure, (Loop A + B) No Yes, with no CFO necessary LRC F ailure (Loop A) No Yes, with no CFO necessary LRC F ailure (Loop B) No Yes ESH Failure (Loop A) No Yes, with no CFO necessary LO OP B ESH Failure (Loop B) No Yes Shelf (backplane) Failure No Yes, with no CFO necessary Shelf, Single Power F ailure No Yes Shelf, Dual Power Failure No Yes, with no CFO necessary Head, Single Power Failure No Yes Head, Dual Power F ailure Yes Yes, if CFO is successful Cluster Interconnect Failure (1 port) No Yes Cluster Interconnect Failure (both ports) No Yes Ethernet Interface Failure (prim ary, no VIF ) No No Ethernet Interface Failure (prim ary, VIF ) No Yes Ethernet Interface Failure (secondary, VIF) No Yes Ethernet Interface Failure (VIF, all ports) No No Tape Interface Failure No Yes Heat +/- operating tem peratures Yes Yes, if partner is within operating tem p... Fan Failures (disk shelves or filer head) No Yes Reboot/Panic No* M aybe (depends on cause of panic) 22 Figure 9 N series resiliency 16 Is data still available o n the affected volum e after th e event? IBM System Storage N series MetroCluster Failure scenarios In this section, we give examples of some possible failures and the associated recovery that is available when you use MetroCluster. Host failure In this scenario, Host 1 is lost; however, access to data continues uninterrupted. Host 2 continues accessing data, as shown in Figure 10. Figure 10 Host failure Node failure In this scenario, environmental factors destroyed IBM N series N1. IBM N series takes over access to its disks, as shown in Figure 11 on page 18. The fabric switches provide the connectivity for the IBM N series and the hosts, to continue data access uninterrupted. IBM System Storage N series MetroCluster 17 Figure 11 is an example of Node failure. Host 1 Data Center #1 Host 2 Data Center #2 Recovery - Automatic N #2 can access the disks in both DCs N1 DC1 Primary DC 2 Mirror N2 DC 2 Primary DC1 Mirror Figure 11 Node failure Node and expansion unit failure This scenario is likely happen when a catastrophic action occurs to a rack with a IBM N series and its expansion units. Both the IBM N series and the expansion units become unavailable, as shown in Figure 12 on page 19. In order to continue access, you must perform the cluster failure with the cfo -d command. Data access is restored because DC1 Mirror was synchronized with DC1 Primary. All hosts through connectivity provided by the fabric switches can once again have access to required data. 18 IBM System Storage N series MetroCluster Figure 12 depicts a node and expansion unit failure. Host 1 Data Center #1 Data Center #2 Host 2 Dual Failure Recovery – One Step failover with cfo –d command N1 DC1 Primary DC 2 Mirror N2 DC 2 Primary DC1 Mirror Figure 12 Node and expansion unit failure MetroCluster interconnect failure In this scenario, the fabric switch interconnects failed (Figure 13 on page 20). Although this is not a critical failure, resolution must occur promptly in case of a more critical failure. During this period, data access is uninterrupted to all hosts. Mirroring and failover are disabled, which reduces data protection. Upon resolution of interconnect failure, re-syncing of mirrors occurs. IBM System Storage N series MetroCluster 19 Figure 13 illustrates the MetroCluster interconnect failure, where the fabric switch interconnects failed. Host 1 Data Center #1 Data Center #2 Host 2 Recovery - No failover; Mirroring disabled - Both filer heads will continue to run serving it’s LUNs/volumes N1 DC1 Primary DC 2 Mirror N2 - Re-syncing happens automatically after interconnect is reestablished DC 2 Primary DC1 Mirror Figure 13 Interconnect failure MetroCluster site failure In this scenario, a site disaster occurred, and all switches, storage systems, and hosts are lost (Figure 14 on page 21). To continue data access a cluster failover must be initiated with the cfo -d command. Both primaries now exist at Data Center 2, and hosting of Host1 is also done at Data Center 2. Note: If the site failure is staggered in nature and the interconnect fails before the rest of the site is destroyed there is a chance of data loss. This occurs because processing has continued after the interconnect has failed. Typically site failures occur pervasively and at the same time. 20 IBM System Storage N series MetroCluster Figure 14 depicts the MetroCluster site failure, where a site disaster occurs, and all switches, storage systems, and hosts are lost. Figure 14 MetroCluster site failure MetroCluster site recovery After the hosts, switches, and storage systems are recovered at Data Center 1, a recovery can be performed. A cf giveback command is performed to resume normal operations, as shown in Figure 15 on page 22. Mirrors are re-synchronized and primaries and mirrors are reversed to previous status. IBM System Storage N series MetroCluster 21 Figure 15 illustrates a MetroCluster site recovery. IBM System Storage MetroCluster – Site Recovery Data Center Center #1 Host 1 Host 2 Host 1 Emulation Data Center #2 Site Recovery One Step return to normal CF giveback N2 N1 DC1 Primary reverts back to DC1 after rebuild DC 2 Primary DC1 Primary DC1 Break DC1 Mirror Mirror DC 2 Mirror N Series Metro Cluster © 2006 IBM Corporation Figure 15 MetroCluster site recovery The team that wrote this Redpaper This Redpaper was produced by a team of specialists from around the world working at the International Technical Support Organization (ITSO), Poughkeepsie Center. Alex Osuna is a Project Leader at the International Technical Support Organization, San Jose Center. He has worked in the IT industry for 26 years, including 25 years with IBM. He has extensive experience in software and hardware storage in service, support, planning, early ship programs, performance analysis, education, and published flashes. Alex has also provided presales and postsales support. Roland Tretau is an IBM Certified IT Specialist and Consulting IT Specialist who provides technical leadership for IBM System Storage N series NAS solutions throughout Europe. He has a solid background in project management, consulting, operating systems, storage solutions, enterprise search technologies, and data management. Roland led software development, large-scale IT solution and architectures designs in customers’ mission critical environments. He holds a Master's degree in Electrical Engineering with an emphasis in telecommunications. He is a Red Hat Certified Engineer (RHCE) and a Microsoft® Certified Systems Engineer (MCSE). He also holds a Masters Certificate in Project Management from The George Washington University School of Business and Public Management. 22 IBM System Storage N series MetroCluster Notices This information was developed for products and services offered in the U.S.A. IBM may not offer the products, services, or features discussed in this document in other countries. Consult your local IBM representative for information on the products and services currently available in your area. Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and verify the operation of any non-IBM product, program, or service. IBM may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to: IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A. The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. Any references in this information to non-IBM Web sites are provided for convenience only and do not in any manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the materials for this IBM product and use of those Web sites is at your own risk. IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. To illustrate them as completely as possible, the examples include the names of individuals, companies, brands, and products. All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. COPYRIGHT LICENSE: This information contains sample application programs in source language, which illustrate programming techniques on various operating platforms. You may copy, modify, and distribute these sample programs in any form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of these programs. Send us your comments in one of the following ways: Use the online Contact us review redbooks form found at: ibm.com/redbooks Send your comments in an email to: redbook@us.ibm.com Mail your comments to: © Copyright International Business Machines Corporation 2006. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. 23 IBM Corporation, International Technical Support Organization Dept. HYTD Mail Station P099 2455 South Road Poughkeepsie, NY 12601-5400 U.S.A. ® Trademarks The following terms are trademarks of the International Business Machines Corporation in the United States, other countries, or both: Redbooks (logo) ESCON® ® FICON® IBM® Redpaper System Storage™ The following terms are trademarks of other companies: WAFL, SyncMirror, SnapMirror, Data ONTAP, and the Network Appliance logo are trademarks or registered trademarks of Network Appliance, Inc. in the U.S. and other countries. Microsoft, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Other company, product, or service names may be trademarks or service marks of others. 24 IBM System Storage N series MetroCluster