Upgrade or Expand The MetroCluster Configuration
Upgrade or Expand The MetroCluster Configuration
configuration
ONTAP MetroCluster
NetApp
May 18, 2022
• Upgrade procedures apply only to the controller modules. The controllers are replaced with a new
controller model.
◦ In switchover and switchback procedures, the MetroCluster switchover operation is used to provide
nondisruptive service to clients while the controller modules on the partner cluster are upgraded.
◦ In an ARL-based controller upgrade procedure, the aggregate relocation operations are used to
nondisruptively move data from the old configuration to the new, upgraded configuration.
• Refresh procedures apply to the controllers and the storage shelves.
In the refresh procedures, new controllers and shelves are added to the MetroCluster configuration,
creating a second DR group, and then data is nondisruptively migrated to the new nodes.
1
• Scope: Platform FC 9.8 Upgrading controllers in a
(controller modules) MetroCluster FC
only configuration using
switchover and
• Method:
switchback
Switchover/switchbac
k
2
Aggregate relocation MetroCluster type First ONTAP version Procedure
procedure support
Other components in the configuration, such as storage shelves or switches, cannot be upgraded at the
same time.
Do not use this procedure on four- or eight-node configurations running ONTAP versions prior to 9.8.
If the original or new platforms are FAS8020 or AFF8020 systems using ports 1c and 1d in
FC-VI mode, contact technical support.
3
• All controllers in the configuration should be upgraded during the same maintenance period.
Operating the MetroCluster configuration with different controller types is not supported outside of this
maintenance activity.
Upgrading a FAS8020 or AFF8020 is not supported when using 1c and 1d ports in FC-VI mode.
• Mapping of storage, FC and Ethernet connections between original nodes and new nodes in advance is
recommended.
• If the new platform has fewer slots than the original system, or if it has fewer or different types of ports, you
might need to add an adapter to the new system.
• site_A
◦ Before upgrade:
▪ node_A_1-old
▪ node_A_2-old
◦ After upgrade:
▪ node_A_1-new
▪ node_A_2-new
• site_B
◦ Before upgrade:
▪ node_B_1-old
▪ node_B_2-old
◦ After upgrade:
▪ node_B_1-new
▪ node_B_2-new
4
Preparing for the upgrade
Before making any changes to the existing MetroCluster configuration, you must check the health of the
configuration, prepare the new platforms, and perform other miscellaneous tasks.
You must verify the health and connectivity of the MetroCluster configuration prior to performing the upgrade.
Steps
1. Verify the operation of the MetroCluster configuration in ONTAP:
a. Check whether the nodes are multipathed:
node run -node node-name sysconfig -a
You should issue this command for each node in the MetroCluster configuration.
You should issue this command on each node in the MetroCluster configuration.
f. Verify that the time zone and time are set correctly on both sites:
You should issue this command on each cluster. You can use the cluster date commands to
configure the time and time zone.
3. Confirm the operational mode of the MetroCluster configuration and perform a MetroCluster check.
5
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
b. After running Config Advisor, review the tool’s output and follow the recommendations in the output to
address any issues discovered.
You must plan the mapping of the LIFs on physical ports on the old nodes to the physical ports on the new
nodes.
The following table shows examples of configuration changes related to the port requirements of the new
nodes.
e0c, e0d e0a,e0b,e0c,e0d e0c and e0d are matching ports. You do not have to change the
configuration, but after upgrade you can spread your cluster LIFs
across the available cluster ports.
Steps
1. Determine what physical ports are available on the new controllers and what LIFs can be hosted on the
ports.
6
The controller’s port usage depends on the platform module and which switches you will use in the
MetroCluster IP configuration. You can gather the port usage of the new platforms from the NetApp
Hardware Universe.
2. Plan your port usage and, if desired, fill in the following tables for reference for each of the new nodes.
You will refer to the table as you carry out the upgrade procedure.
node_A_1-old node_A_1-new
LIF Ports IPspaces Broadcast Ports IPspaces Broadcast
domains domains
Cluster 1
Cluster 2
Cluster 3
Cluster 4
Node
management
Cluster
management
Data 1
Data 2
Data 3
Data 4
SAN
Intercluster
port
Before upgrading, you must gather information for each of the nodes, and, if necessary, adjust the network
broadcast domains, remove any VLANs and interface groups, and gather encryption information.
7
Steps
1. Label the cables for the existing controllers, to allow easy identification of cables when setting up the new
controllers.
2. Gather the system IDs of the nodes in the MetroCluster configuration:
During the replacement procedure you will replace these system IDs with the system IDs of the new
controller modules.
In this example for a four-node MetroCluster FC configuration, the following old system IDs are retrieved:
◦ node_A_1-old: 4068741258
◦ node_A_2-old: 4068741260
◦ node_B_1-old: 4068741254
◦ node_B_2-old: 4068741256
In this example for a two-node MetroCluster FC configuration, the following old system IDs are
retrieved:
◦ node_A_1: 4068741258
◦ node_B_1: 4068741254
8
metrocluster node show -fields node-systemid,dr-partner-systemid
You should gather the output of the following commands for each node:
6. If the MetroCluster nodes are using encryption for volumes or aggregates, copy information about the keys
and passphrases.
For additional information, see Backing up onboard key management information manually.
9
You will need the passphrase later in the upgrade procedure.
Removing the existing configuration from the Tiebreaker or other monitoring software
If the existing configuration is monitored with the MetroCluster Tiebreaker configuration or other third-party
applications (for example, ClusterLion) that can initiate a switchover, you must remove the MetroCluster
configuration from the Tiebreaker or other software prior to transition.
Steps
1. Remove the existing MetroCluster configuration from the Tiebreaker software.
2. Remove the existing MetroCluster configuration from any third-party application that can initiate switchover.
Before performing the maintenance, you should issue an AutoSupport message to notify NetApp technical
support that maintenance is underway. Informing technical support that maintenance is underway prevents
them from opening a case on the assumption that a disruption has occurred.
Steps
1. To prevent automatic support case generation, send an Autosupport message to indicate maintenance is
underway.
a. Issue the following command:
10
About this task
This task must be performed on site_A.
After completing this task, cluster_A is active and serving data for both sites. cluster_B is inactive, and ready to
begin the upgrade process, as shown in the following illustration.
Steps
1. Switch over the MetroCluster configuration to site_A so that site_B’s nodes can be upgraded:
a. Issue the following command on cluster_A:
c. After the operation is complete, confirm that the nodes are in switchover state:
metrocluster show
b. Confirm the heal operation is complete by running the metrocluster operation show command
on the healthy cluster:
11
cluster_A::> metrocluster operation show
Operation: heal-aggregates
State: successful
Start Time: 7/29/2020 20:54:41
End Time: 7/29/2020 20:54:42
Errors: -
b. Confirm the heal operation is complete by running the metrocluster operation show command
on the healthy cluster:
Steps
1. Boot the old nodes and then log in to the nodes:
boot_ontap
2. Assign the home port of all data LIFs on the old controller to a common port that is the same on both the
old and new controller modules.
a. Display the LIFs:
All data LIFS including SAN and NAS will be admin up and operationally down since those are up at
switchover site (cluster_A).
b. Review the output to find a common physical network port that is the same on both the old and new
12
controllers that is not used as a cluster port.
For example, e0d is a physical port on old controllers and is also present on new controllers. e0d is not
used as a cluster port or otherwise on the new controllers.
For port usage for platform models, see the NetApp Hardware Universe
c. Modify all data LIFS to use the common port as the home port:
For example:
3. Modify broadcast domains to remove vlan and physical ports that need to be deleted:
4. Remove any VLAN ports using cluster ports as member ports and ifgrps using cluster ports as member
ports.
a. Delete VLAN ports:
For example:
For example:
network port ifgrp remove-port -node node1 -ifgrp a1a -port e0d
d. Modify interface group ports to use other physical ports as member as needed.:
13
ifgrp add-port -node node-name -ifgrp interface-group-name -port port-id
Steps
1. Connect to the serial console of the old controllers (node_B_1-old and node_B_2-old) at site_B and verify it
is displaying the LOADER prompt.
2. Disconnect the storage and network connections on node_B_1-old and node_B_2-old and label the cables
so they can be reconnected to the new nodes.
3. Disconnect the power cables from node_B_1-old and node_B_2-old.
4. Remove the node_B_1-old and node_B_2-old controllers from the rack.
Steps
1. Plan out the positioning of the new controller modules and storage shelves as needed.
The rack space depends on the platform model of the controller modules, the switch types, and the number
of storage shelves in your configuration.
4. If the new controller modules did not come with FC-VI cards of their own and if FC-VI cards from old
controllers are compatible on new controllers, swap FC-VI cards and install those in correct slots.
See the NetApp Hardware Universe for slot info for FC-VI cards.
5. Cable the controllers' power, serial console and management connections as described in the MetroCluster
Installation and Configuration Guides.
14
Do not connect any other cables that were disconnected from old controllers at this time.
6. Power up the new nodes and press Ctrl-C when prompted to display the LOADER prompt.
After you install the new nodes, you need to netboot to ensure the new nodes are running the same version of
ONTAP as the original nodes. The term netboot means you are booting from an ONTAP image stored on a
remote server. When preparing for netboot, you must put a copy of the ONTAP 9 boot image onto a web server
that the system can access.
Steps
1. Access the NetApp Support Site to download the files used for performing the netboot of the system.
2. Download the appropriate ONTAP software from the software download section of the NetApp Support Site
and store the ontap-version_image.tgz file on a web-accessible directory.
3. Go to the web-accessible directory and verify that the files you need are available.
4. At the LOADER prompt, configure the netboot connection for a management LIF:
◦ If IP addressing is DHCP, configure the automatic connection:
netboot http://web_server_ip/path_to_web-accessible_directory/netboot/kernel
15
◦ If the platform is any other system, use the following command:
netboot http://web_server_ip/path_to_web-accessible_directory/ontap-
version_image.tgz
6. From the boot menu, select option (7) Install new software first to download and install the new software
image to the boot device.
7. If you are prompted to continue the procedure, enter y, and when prompted for the package, enter the URL
of the image file: http://web_server_ip/path_to_web-accessible_directory/ontap-
version_image.tgz
8. Be sure to enter n to skip the backup recovery when you see a prompt similar to the following:
The node must be rebooted to start using the newly installed software.
Do you want to reboot now? {y|n}
Before using a new controller module in the MetroCluster configuration, you must clear
the existing configuration.
Steps
1. If necessary, halt the node to display the LOADER prompt:
halt
set-defaults
saveenv
16
4. At the LOADER prompt, launch the boot menu:
boot_ontap menu
wipeconfig
6. At the boot menu, select option 5 to boot the system into Maintenance mode.
Depending on the presence and configuration of HBA cards in the controller module, you need to configure
them correctly for your site’s usage.
Steps
1. In Maintenance mode configure the settings for any HBAs in the system:
halt
After you run the command, wait until the node stops at the LOADER prompt.
3. Boot the node back into Maintenance mode to enable the configuration changes to take effect:
boot_ontap maint
17
If you have this type of HBA… Use this command…
CNA ucadmin show
FC fcadmin show
You must verify the HA state of the controllers and chassis, and, if necessary, update the state to match your
system configuration.
Steps
1. In Maintenance mode, display the HA state of the controller module and chassis:
ha-config show
2. If the displayed system state of the controller is not correct, set the HA state for the controller module and
chassis:
Reassign the root aggregate disks to the new controller module, using the sysids gathered earlier
The old system IDs were identified in Gathering information before the upgrade.
The examples in this procedure use controllers with the following system IDs:
18
Node Old system ID New system ID
node_B_1 4068741254 1574774970
Steps
1. Cable all other connections to the new controller modules (FC-VI, storage, cluster interconnect, etc.).
2. Halt the system and boot to Maintenance mode from the LOADER prompt:
boot_ontap maint
disk show -a
The command output shows the system ID of the new controller module (1574774970). However, the root
aggregate disks are still owned by the old system ID (4068741254). This example does not show drives
owned by other nodes in the MetroCluster configuration.
4. Reassign the root aggregate disks on the drive shelves to the new controller:
19
*> disk reassign -s 4068741254 -d 1574774970
Partner node must not be in Takeover mode during disk reassignment from
maintenance mode.
Serious problems could result!!
Do not proceed with reassignment if the partner is in takeover mode.
Abort reassignment (y/n)? n
After the node becomes operational, you must perform a takeover and
giveback of the HA partner node to ensure disk reassignment is
successful.
Do you want to continue (y/n)? Jul 14 19:23:49
[localhost:config.bridge.extra.port:error]: Both FC ports of FC-to-SAS
bridge rtp-fc02-41-rr18:9.126L0 S/N [FB7500N107692] are attached to this
controller.
y
Disk ownership will be updated on all disks previously belonging to
Filer with sysid 4068741254.
Do you want to continue (y/n)? y
disk show
20
6. Display the aggregate status:
aggr status
You must reboot the controllers from the boot menu to update the controller flash image. Additional steps are
required if encryption is configured.
Steps
1. Halt the node:
halt
boot_ontap menu
4. If root encryption is used, depending on the ONTAP version you are using, select the boot menu option or
issue the boot menu command for your key management configuration.
◦ Beginning with ONTAP 9.8, select the boot menu option.
21
Onboard key management Option “10”
Respond “y” to the system id change prompts. Wait for the second reboot messages:
printenv partner-sysid
8. If root encryption is used, depending on the ONTAP version you are using, select the boot menu option or
issue the boot menu command again for your key management configuration.
◦ Beginning with ONTAP 9.8, select the boot menu option.
22
Onboard key management Option “10”
Depending on the key manager setting, perform the recovery procedure by selecting option “10” or
option “11”, followed by option “6” at the first boot menu prompt. To boot the nodes completely, you
might need to repeat the recovery procedure continued by option “1” (normal boot).
You might need to issue the recover_xxxxxxxx_keymanager command at the boot menu prompt
multiple times until the nodes completely boot.
boot_ontap
c. Add the physical port that will host the intercluster LIFs to the corresponding Broadcast domain.
d. Modify intercluster LIFs to use the new physical port as home port.
e. After the intercluster LIFs are up, check the cluster peer status and re-establish cluster peering as
23
needed.
VLAN and interface group membership might be different than that of the old node.
Creating a VLAN
12. If encryption is used, restore the keys using the correct command for your key management configuration.
Verify that LIFs are hosted on appropriate node/ports prior to switchback. The following steps need to be
performed
Steps
1. Verify that LIFs are hosted on the appropriate node and ports prior to switchback.
a. Change to the advanced privilege level:
When entering the network interface modify command within the vserver config
24
override command, you cannot use the tab autocomplete feature. You can create the network
interface modify using autocomplete and then enclose it in the vserver config override
command.
Steps
1. Issue the metrocluster node show command on site_B and check the output.
a. Verify that the new nodes are represented correctly.
b. Verify that the new nodes are in "Waiting for switchback state."
2. Switchback the cluster:
metrocluster switchback
metrocluster show
25
The switchback operation is still in progress when the output displays waiting-for-switchback:
If a switchback takes a long time to finish, you can check on the status of in-progress baselines by using
the metrocluster config-replication resync-status show command. This command is at the
advanced privilege level.
Steps
1. Verify the operation of the MetroCluster configuration:
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
26
metrocluster check show
Step
1. Repeat the steps to upgrade the nodes on cluster_A, beginning with Preparing for the upgrade.
As you perform the tasks, all example references to the clusters and nodes are reversed. For example,
when the example is given to switchover from cluster_A, you will switchover from cluster_B.
Step
1. To resume automatic support case generation, send an AutoSupport message to indicate that the
maintenance is complete.
a. Issue the following command:
1. Use the steps in Adding MetroCluster configurations in MetroCluster Tiebreaker Installation and
Configuration.
You cannot upgrade other components in the configuration, such as storage shelves or switches, at the
same time.
27
• You can use this procedure with ONTAP 9.10.1 and later.
◦ Four and eight-node configurations are supported in ONTAP 9.10.1 and later.
• All controllers in the configuration should be upgraded during the same maintenance period.
The following table shows the supported model matrix for the controller upgrade.
• During the upgrade procedure, you are required to change the MetroCluster fabric, including the RCF and
physical changes of cabling. You can perform the RCF and cabling changes before performing the
controller upgrade.
• This upgrade procedure does not require you do not change the storage, FC, and Ethernet connections
between the original nodes and the new nodes.
• During the upgrade procedure, you should not add or remove other cards from the AFF A700 system. For
more information, see the NetApp Hardware Universe
• site_A
◦ Before upgrade:
▪ node_A_1-A700
▪ node_A_2-A700
◦ After upgrade:
▪ node_A_1-A900
▪ node_A_2-A900
• site_B
◦ Before upgrade:
▪ node_B_1-A700
▪ node_B_2-A700
◦ After upgrade:
▪ node_B_1-A900
▪ node_B_2-A900
28
Clear slot 7 on the AFF A700 controller
The MetroCluster configuration on an AFF A900 requires 8 FC-VI ports across FC-VI cards in slots 5 and 7.
Before starting the upgrade, if there are cards in slot 7 on the AFF A700, you must move them to other slots for
all the nodes of the cluster.
Before you update the RCF files and cabling for the AFF A900 fabric MetroCluster configuration, you must
verify the health and connectivity of the configuration.
Steps
1. Verify the operation of the MetroCluster configuration in ONTAP:
a. Check whether the nodes are multipathed:
node run -node node-name sysconfig -a
You should issue this command for each node in the MetroCluster configuration.
You should issue this command on each node in the MetroCluster configuration.
f. Verify that the time zone and time are set correctly on both sites:
You should issue this command on each cluster. You can use the cluster date commands to
configure the time and time zone.
29
3. Confirm the operational mode of the MetroCluster configuration and perform a MetroCluster check.
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
b. After running Config Advisor, review the tool’s output and follow the recommendations in the output to
address any issues discovered.
The AFF A900 fabric MetroCluster requires two four-port FC-VI adapters per node compared to a single four-
port FC-VI adapter required by an AFF A700. Before you start the controller upgrade to the AFF A900
controller, you must modify the fabric switch RCF files to support the AFF A900 connection topology.
1. From the MetroCluster RCF file download page, download the correct RCF file for a AFF A900 fabric
MetroCluster and the switch model that is in use on the AFF A700 configuration.
2. Update the RCF file on the fabric A switches, switch A1, and switch B1 by following the steps in
Configuring the FC switches.
The RCF file update to support the AFF A900 fabric MetroCluster configuration does not
affect the port and connections used for the AFF A700 fabric MetroCluster configuration.
3. After updating the RCF files on the fabric A switches, all storage and FC-VI connections should come
online. Check the FC-VI connections:
a. Verify that the local and remote site disks are listed in the sysconfig output.
4. You must verify that MetroCluster is in a healthy state after the RCF file update for fabric A switches.
a. Check metro cluster connections: metrocluster interconnect mirror show
b. Run metrocluster check: metrocluster check run
c. See the MetroCluster run results when the run completes: metrocluster check show
30
5. Update the fabric B switches (switches 2 and 4) by repeating Step 2 to Step 5.
Verify the health of the MetroCluster configuration after the RCF file update
You must verify the health and connectivity of the MetroCluster configuration before performing the upgrade.
Steps
1. Verify the operation of the MetroCluster configuration in ONTAP:
a. Check whether the nodes are multipathed:
node run -node node-name sysconfig -a
You should issue this command for each node in the MetroCluster configuration.
You should issue this command on each node in the MetroCluster configuration.
f. Verify that the time zone and time are set correctly on both sites:
You should issue this command on each cluster. You can use the cluster date commands to
configure the time and time zone.
3. Confirm the operational mode of the MetroCluster configuration and perform a MetroCluster check.
a. Confirm the MetroCluster configuration and that the operational mode is normal:
31
metrocluster show
b. After running Config Advisor, review the tool’s output and follow the recommendations in the output to
address any issues discovered.
Map ports from the AFF A700 nodes to the AFF A900 nodes
During the controller upgrade process, you must only change the connections that are mentioned in this
procedure.
If the AFF A700 controllers have a card in slot 7, you should move it to another slot before starting the
controller upgrade procedure. You must have slot 7 available for the addition of the second FC-VI adapter that
is required for the functioning of fabric MetroCluster on the AFF A900.
Before upgrading, you must gather information for each of the nodes, and, if necessary, adjust the network
broadcast domains, remove any VLANs and interface groups, and gather encryption information.
Steps
1. Gather the MetroCluster configuration node system IDs:
During the replacement procedure you will replace these system IDs with the system IDs of the controller
modules.
In this example for a four-node MetroCluster FC configuration, the following old system IDs are retrieved:
◦ node_A_1-A700: 537037649
◦ node_A_2-A700: 537407030
◦ node_B_1-A700: 0537407114
32
◦ node_B_2-A700: 537035354
You should gather the output of the following commands for each node:
33
security key-manager backup show
5. If the MetroCluster nodes are using encryption for volumes or aggregates, copy information about the keys
and passphrases.
For additional information, see Backing up onboard key management information manually.
Remove the existing configuration from the Tiebreaker or other monitoring software
If the existing configuration is monitored with the MetroCluster Tiebreaker configuration or other third-party
applications (for example, ClusterLion) that can initiate a switchover, you must remove the MetroCluster
configuration from the Tiebreaker or other software prior to transition.
Steps
1. Remove the existing MetroCluster configuration from the Tiebreaker software.
2. Remove the existing MetroCluster configuration from any third-party application that can initiate switchover.
Before performing the maintenance, you should issue an AutoSupport message to notify NetApp technical
support that maintenance is underway. Informing technical support that maintenance is underway prevents
them from opening a case on the assumption that a disruption has occurred.
Steps
1. To prevent automatic support case generation, send an Autosupport message to indicate maintenance is
underway.
a. Issue the following command:
34
AutoSupport message indicating the end of the maintenance period:
After completing this task, site_A is active and serving data for both sites. Site_B is inactive, and ready to begin
the upgrade process, as shown in the following illustration.
Steps
1. Switch over the MetroCluster configuration to site_A so that site_B’s nodes can be upgraded:
a. Issue the following command on site_A:
35
c. After the operation is complete, confirm that the nodes are in switchover state:
metrocluster show
b. Confirm the heal operation is complete by running the metrocluster operation show command
on the healthy cluster:
b. Confirm the heal operation is complete by running the metrocluster operation show command
on the healthy cluster:
36
Steps
1. Connect to the serial console of the old controllers (node_B_1-700 and node_B_2-700) at site_B and verify
it is displaying the LOADER prompt.
2. Gather the bootarg values from both nodes at site_B: printenv
3. Power off the chassis at site_B.
Remove the controller module and NVS from both nodes at site_B
Use the following procedure to remove the AFF A700 controller module.
Steps
1. Detach the console cable, if any, and the management cable from the controller module before removing
the controller module.
2. Unlock and remove the controller module from the chassis.
a. Slide the orange button on the cam handle downward until it unlocks.
Cam handle
b. Rotate the cam handle so that it completely disengages the controller module from the chassis, and
then slide the controller module out of the chassis. Make sure that you support the bottom of the
controller module as you slide it out of the chassis.
Use the following procedure to remove the AFF A700 NVS module.
37
The AFF A700 NVS module is in slot 6 and is double the height compared to the other modules
in the system.
If there are any add-on modules used as coredump devices on the AFF A700 non-volatile
storage module, do not transfer those to the AFF A900 NVS. Do not transfer any parts from the
AFF A700 controller module and NVS to the AFF A900.
Use the following procedure to install the AFF A900 NVS in slot 6 of both nodes at site_B
Steps
1. Align the NVS with the edges of the chassis opening in slot 6.
2. Gently slide the NVS into the slot until the lettered and numbered I/O cam latch begins to engage with the
38
I/O cam pin, and then push the I/O cam latch all the way up to lock the NVS in place.
Use the following procedure to install the AFF A900 controller module.
Steps
1. Align the end of the controller module with the opening in the chassis, and then gently push the controller
module halfway into the system.
2. Firmly push the controller module into the chassis until it meets the midplane and is fully seated. The
locking latch rises when the controller module is fully seated.
Do not use excessive force when sliding the controller module into the chassis to avoid
damaging the connectors.
39
Cam handle release button
Cam handle
After swapping the AFF A900 controller module and NVS, you need to netboot the AFF A900 nodes and install
the same ONTAP version and patch level that is running on the cluster. The term netboot means you are
booting from an ONTAP image stored on a remote server. When preparing for netboot, you must add a copy
of the ONTAP 9 boot image onto a web server that the system can access.
It is not possible to check the ONTAP version installed on the boot media of an AFF A900 controller module
unless it is installed in a chassis and powered ON. The ONTAP version on the AFF A900 boot media must be
same as the ONTAP version running on the AFF A700 system that is being upgraded and both the primary and
backup boot images should match. You can configure the images by performing a netboot followed by the
wipeconfig command from the boot menu. If the controller module was previously used in another cluster,
the wipeconfig command clears any residual configuration on the boot media.
40
Before you start
• Verify that you can access a HTTP server with the system.
• You need to download the necessary system files for your system and the correct version of ONTAP from
the NetApp Support site. About this task You must netboot the new controllers if the version of ONTAP
installed is not the same as the version installed on the original controllers. After you install each new
controller, you boot the system from the ONTAP 9 image stored on the web server. You can then download
the correct files to the boot media device for subsequent system boots.
Steps
1. Access NetApp Support to download the files required to perform a system netboot used for performing the
netboot of the system.
2. Download the appropriate ONTAP software from the software download section of the NetApp Support
Site and store the <ontap_version>_image.tgz file on a web-accessible directory.
3. Change to the web-accessible directory and verify that the files you need are available. Your directory
listing should contain <ontap_version>_image.tgz.
4. Configure the netboot connection by choosing one of the following actions. Note: You should use the
management port and IP as the netboot connection. Do not use a data LIF IP or a data outage might
occur while the upgrade is being performed.
41
6. Wait for node 1 that is running on the AFF A900 controller module to boot and display the boot menu
options as shown below:
7. From the boot menu, select option (7) Install new software first. This menu option downloads
and installs the new ONTAP image to the boot device.
Disregard the following message: This procedure is not supported for Non-
Disruptive Upgrade on an HA pair. This note applies to nondisruptive ONTAP
software upgrades, and not controller upgrades. Always use netboot to update the new node
to the desired image. If you use another method to install the image on the new controller,
the wrong incorrect image might install. This issue applies to all ONTAP releases.
8. If you are prompted to continue the procedure, enter y, and when prompted for the package, enter the
URL: http://<web_server_ip/path_to_web-accessible_directory>/<ontap_version>_image.tgz
9. Complete the following substeps to reboot the controller module:
a. Enter n to skip the backup recovery when you see the following prompt: Do you want to restore
the backup configuration now? {y|n}
b. Enter y to reboot when you see the following prompt: The node must be rebooted to start
using the newly installed software. Do you want to reboot now? {y|n}
The controller module reboots but stops at the boot menu because the boot device was reformatted,
and the configuration data needs to be restored.
10. At the prompt, run the wipeconfig command to clear any previous configuration on the boot media:
a. When you see the message below, answer yes: This will delete critical system
configuration, including cluster membership. Warning: do not run this option
on a HA node that has been taken over. Are you sure you want to continue?:
b. The node reboots to finish the wipeconfig and then stops at the boot menu.
11. Select option 5 to go to maintenance mode from the boot menu. Answer yes to the prompts until the node
stops at maintenance mode and the command prompt *>.
42
Restore the HBA configuration
Depending on the presence and configuration of HBA cards in the controller module, you need to configure
them correctly for your site’s usage.
Steps
1. In Maintenance mode configure the settings for any HBAs in the system:
If you have this type of HBA and desired mode… Use this command…
CNA FC ucadmin modify -m fc -t initiator
adapter-name
You must verify the HA state of the controllers and chassis, and, if necessary, update the state to match your
system configuration.
Steps
1. In Maintenance mode, display the HA state of the controller module and chassis:
ha-config show
2. If the displayed system state of the controller or chassis is not correct, set the HA state:
3. Halt the node: halt The node should stop at the LOADER> prompt.
4. On each node, check the system date, time, and time zone: Show date
5. If necessary, set the date in UTC or Greenwich Mean Time (GMT): set date <mm/dd/yyyy>
6. Check the time by using the following command at the boot environment prompt: show time
7. If necessary, set the time in UTC or GMT: set time <hh:mm:ss>
8. Save the settings: saveenv
43
9. Gather environment variables: printenv
10. Boot the node back into Maintenance mode to enable the configuration changes to take effect:
boot_ontap maint
11. Verify the changes you made are effective and ucadmin shows FC initiator ports online.
FC fcadmin show
You must verify the HA state of the controllers and chassis, and, if necessary, update the state to match your
system configuration.
Steps
1. In Maintenance mode, display the HA state of the controller module and chassis:
ha-config show
2. If the displayed system state of the controller is not correct, set the HA state for the controller module and
chassis:
44
Four or eight nodes ha-config modify controller mcc
Reassign the root aggregate disks to the new controller module, using the sysids gathered earlier
The old system IDs were identified in Gathering information before the upgrade.
The examples in this procedure use controllers with the following system IDs:
Steps
1. Cable all other connections to the new controller modules (FC-VI, storage, cluster interconnect, etc.).
2. Halt the system and boot to Maintenance mode from the LOADER prompt:
boot_ontap maint
disk show -a
The example output shows the system ID of the new controller module (1574774970). However, the root
aggregate disks are still owned by the old system ID (4068741254). This example does not show drives
owned by other nodes in the MetroCluster configuration.
45
*> disk show -a
Local System ID: 1574774970
4. Reassign the root aggregate disks on the drive shelves to the new controller:
46
*> disk reassign -s 4068741254 -d 1574774970
Partner node must not be in Takeover mode during disk reassignment from
maintenance mode.
Serious problems could result!!
Do not proceed with reassignment if the partner is in takeover mode.
Abort reassignment (y/n)? n
After the node becomes operational, you must perform a takeover and
giveback of the HA partner node to ensure disk reassignment is
successful.
Do you want to continue (y/n)? Jul 14 19:23:49
[localhost:config.bridge.extra.port:error]: Both FC ports of FC-to-SAS
bridge rtp-fc02-41-rr18:9.126L0 S/N [FB7500N107692] are attached to this
controller.
y
Disk ownership will be updated on all disks previously belonging to
Filer with sysid 4068741254.
Do you want to continue (y/n)? y
47
*> aggr status
Aggr State Status Options
aggr0_node_b_1-root online raid_dp, aggr root, nosnap=on,
mirrored
mirror_resync_priority=high(fixed)
fast zeroed
64-bit
You must reboot the controllers from the boot menu to update the controller flash image. Additional steps are
required if encryption is configured.
Steps
1. Halt the node: halt
2. If external key manager is configured, set the related bootargs:
External key management Option 11 and follow the prompts to provide the
required inputs to recover or restore the key-
manager configuration
48
Respond y to the system id change prompts. Wait for the second reboot messages:
8. If root encryption is used, issue the boot menu command again for your key management configuration.
External key management Option 11 and follow the prompts to provide the
required inputs to recover or restore the key-
manager configuration
You might need to issue the recover_xxxxxxxx_keymanager command at the boot menu prompt
multiple times until the nodes completely boot.
If either node is in takeover mode, perform a giveback using the storage failover giveback
command.
c. Add the physical port that will host the intercluster LIFs to the corresponding Broadcast domain.
d. Modify intercluster LIFs to use the new physical port as home port.
e. After the intercluster LIFs are up, check the cluster peer status and re-establish cluster peering as
needed.
49
f. Recreate VLANs and interface groups as needed.
VLAN and interface group membership might be different than that of the old node.
Creating a VLAN
12. If encryption is used, restore the keys using the correct command for your key management configuration.
Verify that LIFs are hosted on appropriate node/ports prior to switchback. The following steps need to be
performed
Steps
1. Verify that LIFs are hosted on the appropriate node and ports prior to switchback.
a. Change to the advanced privilege level:
When entering the network interface modify command within the vserver config
override command, you cannot use the tab autocomplete feature. You can create the network
interface modify using autocomplete and then enclose it in the vserver config override
command.
50
set -privilege admin
2. Revert the interfaces to their home node:
Steps
1. Issue the metrocluster node show command on site_B and check the output.
a. Verify that the new nodes are represented correctly.
b. Verify that the new nodes are in "Waiting for switchback state."
2. Switchback the cluster:
metrocluster switchback
51
metrocluster show
The switchback operation is still in progress when the output displays waiting-for-switchback:
If a switchback takes a long time to finish, you can check on the status of in-progress baselines by using
the metrocluster config-replication resync-status show command. This command is at the
advanced privilege level.
Steps
1. Verify the operation of the MetroCluster configuration:
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
52
c. Display the results of the MetroCluster check:
Step
1. Repeat the steps to upgrade the nodes on site_A, beginning with Prepare for the upgrade.
As you perform the tasks, all example references to the sites and nodes are reversed. For example, when
the example is given to switchover from site_A, you will switchover from Site_B.
Step
1. To resume automatic support case generation, send an Autosupport message to indicate that the
maintenance is complete.
a. Issue the following command:
1. Use the steps in: Adding MetroCluster configurations in the MetroCluster Tiebreaker Installation and
Configuration section.
53
Other components in the configuration, such as storage shelves or switches, cannot be upgraded at the
same time.
The following table shows the supported model matrix for the controller upgrade.
FAS8200 FAS8300
• You can use this procedure to upgrade controllers in a four-node MetroCluster FC configuration using NSO
based automated switchover and switchback. If you want to perform a controller upgrade using aggregate
relocation (ARL), refer to Use "system controller replace" commands to upgrade controller hardware
running ONTAP 9.8 or later. It is recommended to use the NSO based automated procedure.
• If your MetroCluster sites are physically at two different locations, you should use the automated NSO
controller upgrade procedure to upgrade the controllers at both sites in sequence.
• This automated NSO based controller upgrade procedure gives you the capability to initiate controller
replacement to a MetroCluster disaster recovery (DR) site. You can only initiate a controller replacement at
one site at a time.
• To initiate a controller replacement at site A, you need to run the controller replacement start command
from site B. The operation guides you to replace controllers of both the nodes at site A only. To replace the
controllers at site B, you need to run the controller replacement start command from site A. A message
displays identifying the site at which the controllers are being replaced.
• site_A
◦ Before upgrade:
▪ node_A_1-old
▪ node_A_2-old
◦ After upgrade:
▪ node_A_1-new
▪ node_A_2-new
• site_B
◦ Before upgrade:
▪ node_B_1-old
▪ node_B_2-old
◦ After upgrade:
54
▪ node_B_1-new
▪ node_B_2-new
At any stage during the upgrade, you can run the system controller replace show or system
controller replace show-details command from site A to check the status. If the commands return a
blank output, wait for a few minutes and rerun the command.
Steps
1. Start the automated controller replacement procedure from site A to replace the controllers at site B:
The automated operation executes the prechecks. If no issues are found, the operation pauses so you can
manually collect the configuration related information.
The current source system and all compatible target systems are displayed. If you have
replaced the source controller with a controller that has a different ONTAP version or a non-
compatible platform, the automation operation halts and reports an error after the new nodes
are booted up. To bring the cluster back to a healthy state, you need to follow the manual
recovery procedure.
The system controller replace start command might report the following precheck error:
Check if this error occurred because you have unmirrored aggregates or due to another aggregate issue.
Verify that all mirrored aggregates are healthy and not degraded or mirror-degraded. If this error is due to
unmirrored aggregates only, you can override this error by selecting the -skip-metrocluster-check
true option on the system controller replace start command. If remote storage is accessible,
the unmirrored aggregates come online after switchover. If the remote storage link fails, the unmirrored
aggregates fail to come online.
2. Manually collect the configuration information by logging in at site B and following the commands listed in
the console message under the system controller replace show or system controller
replace show-details command.
Before upgrading, if the root volume is encrypted, you must gather the backup key and other information to
boot the new controllers with the old encrypted root volumes.
55
About this task
This task is performed on the existing MetroCluster FC configuration.
Steps
1. Label the cables for the existing controllers, so you can easily identify the cables when setting up the new
controllers.
2. Display the commands to capture the backup key and other information:
Run the commands listed under the show command from the partner cluster.
During the replacement procedure you will replace these system IDs with the system IDs of the new
controller modules.
In this example for a four-node MetroCluster FC configuration, the following old system IDs are retrieved:
◦ node_A_1-old: 4068741258
◦ node_A_2-old: 4068741260
◦ node_B_1-old: 4068741254
◦ node_B_2-old: 4068741256
In this example for a two-node MetroCluster FC configuration, the following old system IDs are retrieved:
◦ node_A_1: 4068741258
◦ node_B_1: 4068741254
56
metrocluster node show -fields node-systemid,dr-partner-systemid
You should gather the output of the following commands for each node:
7. If the MetroCluster nodes are using encryption for volumes or aggregates, copy information about the keys
and passphrases.
For additional information, see Backing up onboard key management information manually.
57
You will need the passphrase later in the upgrade procedure.
8. After you finish collecting the configuration information, resume the operation:
Removing the existing configuration from the Tiebreaker or other monitoring software
If the existing configuration is monitored with the MetroCluster Tiebreaker configuration or other third-party
applications (for example, ClusterLion) that can initiate a switchover, you must remove the MetroCluster
configuration from the Tiebreaker or other software prior to replacing the old controller.
Steps
1. Remove the existing MetroCluster configuration from the Tiebreaker software.
2. Remove the existing MetroCluster configuration from any third-party application that can initiate switchover.
To ensure that the networking resumes cleanly on the new controllers, you must move LIFs to a common port
and then remove the networking configuration of the old controllers.
Steps
1. Boot the old nodes and then log in to the nodes:
58
boot_ontap
2. Assign the home port of all data LIFs on the old controller to a common port that is the same on both the
old and new controller modules.
a. Display the LIFs:
All data LIFS including SAN and NAS will be admin “up” and operationally “down” since those are up at
switchover site (cluster_A).
b. Review the output to find a common physical network port that is the same on both the old and new
controllers that is not used as a cluster port.
For example, “e0d” is a physical port on old controllers and is also present on new controllers. “e0d” is
not used as a cluster port or otherwise on the new controllers.
For port usage for platform models, see the NetApp Hardware Universe
c. Modify all data LIFS to use the common port as the home port:
For example:
3. Modify broadcast domains to remove VLAN and physical ports that need to be deleted:
4. Remove any VLAN ports using cluster ports as member ports and interface groups using cluster ports as
member ports.
a. Delete VLAN ports:
For example:
59
For example:
network port ifgrp remove-port -node node1 -ifgrp a1a -port e0d
d. Modify interface group ports to use other physical ports as member as needed.:
Steps
1. Plan out the positioning of the new controller modules and storage shelves as needed.
The rack space depends on the platform model of the controller modules, the switch types, and the number
of storage shelves in your configuration.
4. If the new controller modules did not come with FC-VI cards of their own and if FC-VI cards from old
controllers are compatible on new controllers, swap FC-VI cards and install those in correct slots.
See the NetApp Hardware Universe for slot info for FC-VI cards.
5. Cable the controllers' power, serial console and management connections as described in the MetroCluster
Installation and Configuration Guides.
Do not connect any other cables that were disconnected from old controllers at this time.
6. Power up the new nodes and press Ctrl-C when prompted to display the LOADER prompt.
After you install the new nodes, you need to netboot to ensure the new nodes are running the same version of
ONTAP as the original nodes. The term netboot means you are booting from an ONTAP image stored on a
60
remote server. When preparing for netboot, you must put a copy of the ONTAP 9 boot image onto a web server
that the system can access.
Steps
1. Access the NetApp Support Site to download the files used for performing the netboot of the system.
2. Download the appropriate ONTAP software from the software download section of the NetApp Support Site
and store the ontap-version_image.tgz file on a web-accessible directory.
3. Go to the web-accessible directory and verify that the files you need are available.
4. At the LOADER prompt, configure the netboot connection for a management LIF:
◦ If IP addressing is DHCP, configure the automatic connection:
netboot http://web_server_ip/path_to_web-accessible_directory/netboot/kernel
netboot http://web_server_ip/path_to_web-accessible_directory/ontap-
version_image.tgz
6. From the boot menu, select option (7) Install new software first to download and install the new software
image to the boot device.
61
Disregard the following message: "This procedure is not supported for
Non-Disruptive Upgrade on an HA pair". It applies to nondisruptive
upgrades of software, not to upgrades of controllers.
7. If you are prompted to continue the procedure, enter y, and when prompted for the package, enter the URL
of the image file: http://web_server_ip/path_to_web-accessible_directory/ontap-
version_image.tgz
8. Be sure to enter n to skip the backup recovery when you see a prompt similar to the following:
The node must be rebooted to start using the newly installed software.
Do you want to reboot now? {y|n}
Before using a new controller module in the MetroCluster configuration, you must clear
the existing configuration.
Steps
1. If necessary, halt the node to display the LOADER prompt:
halt
set-defaults
saveenv
boot_ontap menu
wipeconfig
62
Respond yes to the confirmation prompt.
6. At the boot menu, select option 5 to boot the system into Maintenance mode.
Depending on the presence and configuration of HBA cards in the controller module, you need to configure
them correctly for your site’s usage.
Steps
1. In Maintenance mode configure the settings for any HBAs in the system:
If you have this type of HBA and desired mode… Use this command…
CNA FC ucadmin modify -m fc -t initiator
adapter-name
halt
After you run the command, wait until the node stops at the LOADER prompt.
3. Boot the node back into Maintenance mode to enable the configuration changes to take effect:
boot_ontap maint
FC fcadmin show
63
Reassigning root aggregate disks
Reassign the root aggregate disks to the new controller module, using the sysids gathered earlier
The old system IDs were identified in Gathering information before the upgrade.
The examples in this procedure use controllers with the following system IDs:
Steps
1. Cable all other connections to the new controller modules (FC-VI, storage, cluster interconnect, etc.).
2. Halt the system and boot to Maintenance mode from the LOADER prompt:
boot_ontap maint
disk show -a
The command output shows the system ID of the new controller module (1574774970). However, the root
aggregate disks are still owned by the old system ID (4068741254). This example does not show drives
owned by other nodes in the MetroCluster configuration.
64
*> disk show -a
Local System ID: 1574774970
4. Reassign the root aggregate disks on the drive shelves to the new controller:
65
*> disk reassign -s 4068741254 -d 1574774970
Partner node must not be in Takeover mode during disk reassignment from
maintenance mode.
Serious problems could result!!
Do not proceed with reassignment if the partner is in takeover mode.
Abort reassignment (y/n)? n
After the node becomes operational, you must perform a takeover and
giveback of the HA partner node to ensure disk reassignment is
successful.
Do you want to continue (y/n)? Jul 14 19:23:49
[localhost:config.bridge.extra.port:error]: Both FC ports of FC-to-SAS
bridge rtp-fc02-41-rr18:9.126L0 S/N [FB7500N107692] are attached to this
controller.
y
Disk ownership will be updated on all disks previously belonging to
Filer with sysid 4068741254.
Do you want to continue (y/n)? y
disk show
66
6. Display the aggregate status:
aggr status
You must reboot the controllers from the boot menu to update the controller flash image. Additional steps are
required if encryption is configured.
You can reconfigure VLANs and interface groups. If required, manually modify the ports for the cluster LIFs
and broadcast domain details before resuming the operation by using the system controller replace
resume command.
Steps
1. Halt the node:
halt
boot_ontap menu
4. If root encryption is used, select the boot menu option for your key management configuration.
67
Onboard key management Option “10”
Respond “y” to the system id change prompts. Wait for the second reboot messages:
printenv partner-sysid
8. If root encryption is used, select the boot menu option again for your key management configuration.
Depending on the key manager setting, perform the recovery procedure by selecting option “10” or option
“11”, followed by option “6” at the first boot menu prompt. To boot the nodes completely, you might need to
repeat the recovery procedure continued by option “1” (normal boot).
68
boot_ontap
If either node is in takeover mode, perform a giveback using the storage failover giveback
command.
c. Add the physical port that will host the intercluster LIFs to the corresponding Broadcast domain.
d. Modify intercluster LIFs to use the new physical port as home port.
e. After the intercluster LIFs are up, check the cluster peer status and re-establish cluster peering as
needed.
VLAN and interface group membership might be different than that of the old node.
Creating a VLAN
12. If encryption is used, restore the keys using the correct command for your key management configuration.
13. Before you resume the operation, verify that the MetroCluster is configured correctly. Check the node
status:
69
metrocluster node show
Verify that the new nodes (site_B) are in Waiting for switchback state from site_A.
Steps
1. Verify the network reachability by following the console message.
2. After you complete the verification, resume the operation:
3. The automation operation performs switchback at site A and the post upgrade checks. When the operation
pauses, manually check the SAN LIF status and verify the network configuration by following the console
message.
4. After you complete the verification, resume the operation:
If the post upgrade checks did not report any errors, the upgrade is complete.
6. After you complete the controller upgrade, log in at site B and verify that the replaced controllers are
configured correctly.
If the MetroCluster configuration was previously configured for monitoring by the Tiebreaker software, you can
restore the Tiebreaker connection.
70
as part of this procedure.
About this task
• The platforms must be running ONTAP 9.8 or later.
• This procedure applies to controller modules in a MetroCluster IP configuration.
• The supported upgrade path depends on the original platform model.
• FAS8200 • FAS9000
• FAS8300
• FAS8700
AFF A320 platform models are not supported for upgrade when using BES-53248 IP switches.
• All controllers in the configuration should be upgraded during the same maintenance period.
Operating the MetroCluster configuration with different controller types is not supported outside of this
maintenance activity.
• The new platform must be a different model than the original platform.
• The IP switches must be running a supported firmware version.
• If the new platform has fewer slots than the original system, or if it has fewer or different types of ports, you
might need to add an adapter to the new system.
• You will reuse the IP addresses, netmasks, and gateways of the original platforms on the new platforms.
• The following example names are used in this procedure:
◦ site_A
▪ Before upgrade:
▪ node_A_1-old
▪ node_A_2-old
▪ After upgrade:
▪ node_A_1-new
▪ node_A_2-new
◦ site_B
▪ Before upgrade:
▪ node_B_1-old
▪ node_B_2-old
71
▪ After upgrade:
▪ node_B_1-new
▪ node_B_2-new
Depending on the old platform models, or if switch configuration is not on the minimum version, or if you want
to change VLAN IDs used by the back-end MetroCluster connections, you must update the switch RCF files
before you begin the platform upgrade procedure.
72
About this task
You must update the RCF file in the following scenarios:
• For certain platform models, the switches must be using a supported VLAN ID for the back-end
MetroCluster IP connections. If the old or new platform models are in the following table, and not using a
supported VLAN ID, you must update the switch RCF files.
The local cluster connections can use any VLAN, they do not need to be in the given range.
• The switch configuration was not configured with minimum supported RCF version:
The switches at site_A will be upgraded when the controllers on site_A are upgraded.
Steps
1. Prepare the IP switches for the application of the new RCF files.
Follow the steps in the section for your switch vendor from the MetroCluster IP installation and
configuration.
73
Mapping ports from the old nodes to the new nodes
You must verify that the physical ports on node_A_1-old map correctly to the physical ports on node_A_1-new,
which will allow node_A_1-new to communicate with other nodes in the cluster and with the network after the
upgrade.
The following table shows examples of configuration changes related to the port requirements of the new
nodes.
Steps
1. Determine what physical ports are available on the new controllers and what LIFs can be hosted on the
ports.
The controller’s port usage depends on the platform module and which switches you will use in the
MetroCluster IP configuration. You can gather the port usage of the new platforms from the NetApp
Hardware Universe.
2. Plan your port usage and fill in the following tables for reference for each of the new nodes.
You will refer to the table as you carry out the upgrade procedure.
node_A_1-old node_A_1-new
LIF Ports IPspaces Broadcast Ports IPspaces Broadcast
domains domains
Cluster 1
Cluster 2
Cluster 3
74
Cluster 4
Node
management
Cluster
management
Data 1
Data 2
Data 3
Data 4
SAN
Intercluster
port
After you install the new nodes, you need to netboot to ensure the new nodes are running the same version of
ONTAP as the original nodes. The term netboot means you are booting from an ONTAP image stored on a
remote server. When preparing for netboot, you must put a copy of the ONTAP 9 boot image onto a web server
that the system can access.
Steps
1. Netboot the new controllers:
a. Access the NetApp Support Site to download the files used for performing the netboot of the system.
b. Download the appropriate ONTAP software from the software download section of the NetApp Support
Site and store the ontap-version_image.tgz file on a web-accessible directory.
c. Change to the web-accessible directory and verify that the files you need are available.
75
FAS/AFF8000 series systems Extract the contents of the ontap-
version_image.tgz file to the target directory:
netboot/kernel
_ontap-version_image.tgz
d. At the LOADER prompt, configure the netboot connection for a management LIF:
76
All other systems netboot
http://_web_server_ip/path_to_web-
accessible_directory/ontap-
version_image.tgz
f. From the boot menu, select option (7) Install new software first to download and install the new
software image to the boot device.
g. If you are prompted to continue the procedure, enter y, and when prompted for the package, enter the
URL of the image file:
http://web_server_ip/path_to_web-accessible_directory/ontap-
version_image.tgz
h. Enter the user name and password if applicable, or press Enter to continue.
i. Be sure to enter n to skip the backup recovery when you see a prompt similar to the following:
Before using a new controller module in the MetroCluster configuration, you must clear
the existing configuration.
Steps
1. If necessary, halt the node to display the LOADER prompt:
halt
set-defaults
saveenv
77
4. At the LOADER prompt, launch the boot menu:
boot_ontap menu
wipeconfig
6. At the boot menu, select option 5 to boot the system into Maintenance mode.
You must verify the health and connectivity of the MetroCluster configuration prior to performing the upgrade.
Steps
1. Verify the operation of the MetroCluster configuration in ONTAP:
a. Check whether the nodes are multipathed:
node run -node node-name sysconfig -a
You should issue this command for each node in the MetroCluster configuration.
You should issue this command on each node in the MetroCluster configuration.
f. Verify that the time zone and time is set correctly on both sites:
78
You should issue this command on each cluster. You can use the cluster date commands to
configure the time and time zone.
2. Confirm the operational mode of the MetroCluster configuration and perform a MetroCluster check.
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
b. Confirm that all expected nodes are shown:
metrocluster node show
c. Issue the following command:
b. After running Config Advisor, review the tool’s output and follow the recommendations in the output to
address any issues discovered.
Before upgrading, you must gather information for each of the nodes, and, if necessary, adjust the network
broadcast domains, remove any VLANs and interface groups, and gather encryption information.
Steps
1. Record the physical cabling for each node, labelling cables as needed to allow correct cabling of the new
nodes.
2. Gather interconnect, port and LIF information for each node.
You should gather the output of the following commands for each node:
79
◦ storage aggregate show
◦ system node run -node node-name sysconfig -a
◦ vserver fcp initiator show
◦ storage disk show
◦ metrocluster configuration-settings interface show
3. Gather the UUIDs for the site_B (the site whose platforms are currently being upgraded):
These values must be configured accurately on the new site_B controller modules to ensure a successful
upgrade. Copy the values to a file so that you can copy them into the proper commands later in the
upgrade process.
The following example shows the command output with the UUIDs:
It is recommended that you record the UUIDs into a table similar to the following.
node_B_1 f37b240b-9ac1-11e7-9b42-00a098c9e55d
node_B_2 bf8e3f8f-9ac4-11e7-bd4e-00a098ca379f
cluster_A ee7db9d5-9a82-11e7-b68b-00a098908039
80
node_A_1 f03cb63c-9a7e-11e7-b68b-00a098908039
node_A_2 aa9a7a7a-9a81-11e7-a4e9-00a098908c35
4. If the MetroCluster nodes are in a SAN configuration, collect the relevant information.
6. If the MetroCluster nodes are using encryption for volumes or aggregates, copy information about the keys
and passphrases.
For additional information, see Backing up onboard key management information manually.
81
::> metrocluster node show -fields node-systemid,ha-partner-systemid,dr-
partner-systemid,dr-auxiliary-systemid
Before the upgrading the platforms, you must remove monitoring if the MetroCluster configuration is monitored
with the Tiebreaker or Mediator utility.
Steps
1. Collect the output for the following command:
2. Remove the existing MetroCluster configuration from Tiebreaker, Mediator, or other software that can
initiate switchover.
metrocluster configuration-settings
mediator remove
82
Sending a custom AutoSupport message prior to maintenance
Before performing the maintenance, you should issue an AutoSupport message to notify NetApp technical
support that maintenance is underway. Informing technical support that maintenance is underway prevents
them from opening a case on the assumption that a disruption has occurred.
Steps
1. Log in to the cluster.
2. Invoke an AutoSupport message indicating the start of the maintenance:
The maintenance-window-in-hours parameter specifies the length of the maintenance window, with a
maximum of 72 hours. If the maintenance is completed before the time has elapsed, you can invoke an
AutoSupport message indicating the end of the maintenance period:
After completing this task, cluster_A is active and serving data for both sites. cluster_B is inactive, and ready to
begin the upgrade process.
Steps
83
1. Switch over the MetroCluster configuration to site_A so that site_B’s nodes can be upgraded:
a. Issue the following command on cluster_A:
c. After the operation is complete, confirm that the nodes are in switchover state:
metrocluster show
Automatic healing of aggregates after negotiated switchover is disabled during controller upgrade.
Steps
1. Boot the old nodes and log in to the nodes:
boot_ontap
2. Assign the home port of all data LIFs on the old controller to a common port that is the same on both the
old and new controller modules.
a. Display the LIFs:
All data LIFS including SAN and NAS will be admin up and operationally down since those are up at
switchover site (cluster_A).
b. Review the output to find a common physical network port that is the same on both the old and new
controllers that is not used as a cluster port.
For example, e0d is a physical port on old controllers and is also present on new controllers. e0d is not
used as a cluster port or otherwise on the new controllers.
For port usage for platform models, see the NetApp Hardware Universe
84
c. Modify all data LIFS to use the common port as the home port:
network interface modify -vserver svm-name -lif data-lif -home-port port-id
For example:
3. Remove any VLAN ports using cluster ports as member ports and ifgrps using cluster ports as member
ports.
a. Delete VLAN ports:
network port vlan delete -node node-name -vlan-name portid-vlandid
For example:
For example:
network port ifgrp remove-port -node node1 -ifgrp a1a -port e0d
d. Modify interface group ports to use other physical ports as member as needed.:
5. Connect to the serial console of the old controllers (node_B_1-old and node_B_2-old) at site_B and verify it
is displaying the LOADER prompt.
6. Gather the bootarg values:
printenv
7. Disconnect the storage and network connections on node_B_1-old and node_B_2-old and label the cables
so they can be reconnected to the new nodes.
85
8. Disconnect the power cables from node_B_1-old and node_B_2-old.
9. Remove the node_B_1-old and node_B_2-old controllers from the rack.
The switches at site_A will be upgraded when the controllers on site_A are upgraded.
Steps
1. Prepare the IP switches for the application of the new RCF files.
Follow the steps in the section for your switch vendor from the MetroCluster IP installation and
configuration.
Steps
1. Plan out the positioning of the new controller modules and storage shelves as needed.
The rack space depends on the platform model of the controller modules, the switch types, and the number
of storage shelves in your configuration.
4. Cable the controllers to the IP switches as described in MetroCluster IP installation and configuration.
86
◦ Cabling the IP switches
5. Power up the new nodes and boot them to Maintenance mode.
Depending on the presence and configuration of HBA cards in the controller module, you need to configure
them correctly for your site’s usage.
Steps
1. In Maintenance mode configure the settings for any HBAs in the system:
ucadmin show
If you have this type of HBA and desired mode… Use this command…
CNA FC ucadmin modify -m fc -t initiator
adapter-name
halt
After you run the command, wait until the node stops at the LOADER prompt.
3. Boot the node back into Maintenance mode to enable the configuration changes to take effect:
boot_ontap maint
FC fcadmin show
87
Setting the HA state on the new controllers and chassis
You must verify the HA state of the controllers and chassis, and, if necessary, update the state to match your
system configuration.
Steps
1. In Maintenance mode, display the HA state of the controller module and chassis:
ha-config show
2. If the displayed system state of the controller or chassis is not correct, set the HA state:
Certain MetroCluster IP bootarg values must be configured on the new controller modules. The values must
match those configured on the old controller modules.
Steps
1. If the nodes being upgraded are AFF A400, FAS8300, or FAS8700 models, set the following bootargs at
the LOADER prompt:
If the interfaces are using the default VLANs, the vlan-id is not necessary.
The following commands set the values for node_B_1-new using VLAN 120 for the first network and VLAN
130 for the second network:
setenv bootarg.mcc.port_a_ip_config
172.17.26.10/23,0,172.17.26.11,172.17.26.13,172.17.26.12,120
setenv bootarg.mcc.port_b_ip_config
172.17.27.10/23,0,172.17.27.11,172.17.27.13,172.17.27.12,130
The following commands set the values for node_B_2-new using VLAN 120 for the first network and VLAN
130 for the second network:
88
setenv bootarg.mcc.port_a_ip_config
172.17.26.11/23,0,172.17.26.10,172.17.26.12,172.17.26.13,120
setenv bootarg.mcc.port_b_ip_config
172.17.27.11/23,0,172.17.27.10,172.17.27.12,172.17.27.13,130
The following example shows the commands for node_B_1-new when the default VLAN is used:
setenv bootarg.mcc.port_a_ip_config
172.17.26.10/23,0,172.17.26.11,172.17.26.13,172.17.26.12
setenv bootarg.mcc.port_b_ip_config
172.17.27.10/23,0,172.17.27.11,172.17.27.13,172.17.27.12
The following example shows the commands for node_B_2-new when the default VLAN is used:
setenv bootarg.mcc.port_a_ip_config
172.17.26.11/23,0,172.17.26.10,172.17.26.12,172.17.26.13
setenv bootarg.mcc.port_b_ip_config
172.17.27.11/23,0,172.17.27.10,172.17.27.12,172.17.27.13
2. If the nodes being upgraded are not systems listed in the previous step, at the LOADER prompt for each of
the surviving nodes, set the following bootargs with local_IP/mask:
setenv bootarg.mcc.port_a_ip_config
172.17.26.10/23,0,172.17.26.11,172.17.26.13,172.17.26.12
setenv bootarg.mcc.port_b_ip_config
172.17.27.10/23,0,172.17.27.11,172.17.27.13,172.17.27.12
setenv bootarg.mcc.port_a_ip_config
172.17.26.11/23,0,172.17.26.10,172.17.26.12,172.17.26.13
setenv bootarg.mcc.port_b_ip_config
172.17.27.11/23,0,172.17.27.10,172.17.27.12,172.17.27.13
89
3. At the new nodes' LOADER prompt, set the UUIDs:
The following example shows the commands for setting the UUIDs on node_B_1-new:
The following example shows the commands for setting the UUIDs on node_B_2-new:
4. If the original systems were configured for ADP, at each of the replacement nodes' LOADER prompt,
enable ADP:
90
setenv bootarg.mcc.dr_partner dr-partner-sys-id
The following example shows the commands for setting the values on node_B_1-new:
The following example shows the commands for setting the values on node_B_2-new:
6. If using encryption with external key manager, set the required bootargs:
setenv bootarg.kmip.init.ipaddr
setenv bootarg.kmip.kmip.init.netmask
setenv bootarg.kmip.kmip.init.gateway
setenv bootarg.kmip.kmip.init.interface
Reassign the root aggregate disks to the new controller module, using the sysids gathered earlier.
Steps
1. Boot the system to Maintenance mode:
boot_ontap maint
disk show -a
The command output shows the system ID of the new controller module (1574774970). However, the root
aggregate disks are still owned by the old system ID (537403322). This example does not show drives
owned by other nodes in the MetroCluster configuration.
91
*> disk show -a
Local System ID: 1574774970
DISK OWNER POOL SERIAL NUMBER HOME
DR HOME
------------ --------- ----- -------------
------------- -------------
prod3-rk18:9.126L44 node_B_1-old(537403322) Pool1 PZHYN0MD
node_B_1-old(537403322) node_B_1-old(537403322)
prod4-rk18:9.126L49 node_B_1-old(537403322) Pool1 PPG3J5HA
node_B_1-old(537403322) node_B_1-old(537403322)
prod4-rk18:8.126L21 node_B_1-old(537403322) Pool1 PZHTDSZD
node_B_1-old(537403322) node_B_1-old(537403322)
prod2-rk18:8.126L2 node_B_1-old(537403322) Pool0 S0M1J2CF
node_B_1-old(537403322) node_B_1-old(537403322)
prod2-rk18:8.126L3 node_B_1-old(537403322) Pool0 S0M0CQM5
node_B_1-old(537403322) node_B_1-old(537403322)
prod1-rk18:9.126L27 node_B_1-old(537403322) Pool0 S0M1PSDW
node_B_1-old(537403322) node_B_1-old(537403322)
.
.
.
3. Reassign the root aggregate disks on the drive shelves to the new controllers.
4. Reassign the root aggregate disks on the drive shelves to the new controllers:
92
*> disk reassign -s 537403322 -d 1574774970
Partner node must not be in Takeover mode during disk reassignment from
maintenance mode.
Serious problems could result!!
Do not proceed with reassignment if the partner is in takeover mode.
Abort reassignment (y/n)? n
After the node becomes operational, you must perform a takeover and
giveback of the HA partner node to ensure disk reassignment is
successful.
Do you want to continue (y/n)? y
Disk ownership will be updated on all disks previously belonging to
Filer with sysid 537403322.
Do you want to continue (y/n)? y
5. Verify that the disks of the root aggregate are properly reassigned old-remove:
disk show
93
*> disk show
Local System ID: 537097247
You must boot the new controllers, taking care to ensure that the bootarg variables are correct and, if needed,
perform the encryption recovery steps.
Steps
1. Halt the new nodes:
halt
94
setenv bootarg.kmip.init.interface interface-id
printenv partner-sysid
boot_ontap menu
5. If root encryption is used, select the boot menu option for your key management configuration.
6. From the boot menu, select “(6) Update flash from backup config”.
Respond “y” to the system id change prompts. Wait for the second reboot messages:
7. On LOADER, double-check the bootarg values and update the values as needed.
printenv partner-sysid
95
9. If root encryption is used, select the boot menu option again for your key management configuration.
Depending on the key manager setting, perform the recovery procedure by selecting option “10” or option
“11”, followed by option 6 at the first boot menu prompt. To boot the nodes completely, you might need to
repeat the recovery procedure continued by option “1” (normal boot).
If either node is in takeover mode, perform a giveback using the storage failover giveback
command.
11. If encryption is used, restore the keys using the correct command for your key management configuration.
96
VLAN and interface group membership might be different than that of the old node.
Creating a VLAN
Verify that LIFs are hosted on appropriate nodes and ports as mapped out at the beginning of the upgrade
procedure.
Steps
1. Verify that LIFs are hosted on the appropriate node and ports prior to switchback.
a. Change to the advanced privilege level:
When entering the network interface modify command within the vserver config override
command, you cannot use the tab autocomplete feature. You can create the network interface
modify using autocomplete and then enclose it in the vserver config override command.
97
Steps
1. Issue the metrocluster node show command on site_B and check the output.
a. Verify that the new nodes are represented correctly.
b. Verify that the new nodes are in "Waiting for switchback state."
2. Perform the healing and switchback by running the required commands from any node in the active cluster
(the cluster that is not undergoing upgrade).
a. Heal the data aggregates:
metrocluster heal aggregates
b. Heal the root aggregates:
metrocluster switchback
metrocluster show
The switchback operation is still in progress when the output displays waiting-for-switchback:
98
cluster_B::> metrocluster show
Cluster Entry Name State
------------------------- ------------------- -----------
Local: cluster_B Configuration state configured
Mode switchover
AUSO Failure Domain -
Remote: cluster_A Configuration state configured
Mode waiting-for-switchback
AUSO Failure Domain -
If a switchback takes a long time to finish, you can check on the status of in-progress baselines by using
the metrocluster config-replication resync-status show command. This command is at the
advanced privilege level.
Steps
1. Verify the operation of the MetroCluster configuration:
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
b. Perform a MetroCluster check:
metrocluster check run
c. Display the results of the MetroCluster check:
99
a. Check the MetroCluster IP connections:
Steps
1. Repeat the steps to upgrade the nodes on cluster_A, beginning with Preparing for the upgrade.
As you perform the tasks, all example references to the clusters and nodes are reversed. For example,
when the example is given to switchover from cluster_A, you will switchover from cluster_B.
Steps
1. Restore monitoring if necessary, using the procedure for your configuration.
Steps
1. To resume automatic support case generation, send an Autosupport message to indicate that the
100
maintenance is complete.
a. Issue the following command:
system node autosupport invoke -node * -type all -message MAINT=end
b. Repeat the command on the partner cluster.
• All controllers in the configuration should be upgraded during the same maintenance period.
Operating the MetroCluster configuration with an AFF A700 and an AFF A900 controller is not supported
outside of this maintenance activity.
101
▪ node_B_1-A900
▪ node_B_2-A900
The MetroCluster configuration on an AFF A900 uses one of each of the ports on the DR cards located in slots
5 and 7. Before starting the upgrade, if there are cards in slot 7 on the AFF A700, you must move them to other
slots for all the nodes of the cluster.
102
Update the MetroCluster switch RCF files before upgrading controllers
The MetroCluster IP configuration on an AFF A700 does not use VLANs. The MetroCluster configuration on an
AFF A900 does use VLANs. As a result, you need to change the RCF file when upgrading from an AFF A700
to an AFF A900.
• If the switch is not configured with the minimum supported RCF file version, you must update the RCF file.
For the correct RCF file version for your switch model, refer to the RcfFileGenerator Tool. The following
steps are for the RCF file application.
Steps
1. Prepare the IP switches for the application of the new RCF files.
Follow the steps in the section for your switch vendor from the MetroCluster IP Installation and
Configuration content.
When upgrading from an AFF A700 to an AFF A900, you do not change the data network ports, FCP SAN
adapter ports, and SAS and NVMe storage ports. Data LIFs stay where they are during and after the upgrade.
Therefore, you are not required to map the network ports from the old nodes to the new nodes.
You must verify the health and connectivity of the MetroCluster configuration prior to performing the upgrade.
Steps
1. Verify the operation of the MetroCluster configuration in ONTAP:
a. Check whether the nodes are multipathed:
node run -node node-name sysconfig -a
You should issue this command for each node in the MetroCluster configuration.
103
You should issue this command on each node in the MetroCluster configuration.
f. Verify that the time zone and time is set correctly on both sites:
You should issue this command on each cluster. You can use the cluster date command to
configure the time and time zone.
2. Confirm the operational mode of the MetroCluster configuration and perform a MetroCluster check.
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
b. Confirm that all expected nodes are shown:
metrocluster node show
c. Issue the following command:
b. After running Config Advisor, review the tool’s output and follow the recommendations in the output to
address any issues discovered.
Before upgrading, you must gather information for each of the nodes, and, if necessary, adjust the network
broadcast domains, remove any VLANs and interface groups, and gather encryption information.
104
Steps
1. Record the physical cabling for each node, labelling cables as needed to allow correct cabling of the new
nodes.
2. Gather the output of the following commands for each node:
◦ metrocluster interconnect show
◦ metrocluster configuration-settings connection show
◦ network interface show -role cluster,node-mgmt
◦ network port show -node node_name -type physical
◦ network port vlan show -node node-name
◦ network port ifgrp show -node node_name -instance
◦ network port broadcast-domain show
◦ network port reachability show -detail
◦ network ipspace show
◦ volume show
◦ storage aggregate show
◦ system node run -node node-name sysconfig -a
◦ vserver fcp initiator show
◦ storage disk show
◦ metrocluster configuration-settings interface show
3. Gather the UUIDs for the site_B (the site whose platforms are currently being upgraded): metrocluster
node show -fields node-cluster-uuid, node-uuid
These values must be configured accurately on the new site_B controller modules to ensure a successful
upgrade. Copy the values to a file so that you can copy them into the proper commands later in the
upgrade process.
The following example shows the command output with the UUIDs:
105
cluster_B::> metrocluster node show -fields node-cluster-uuid, node-uuid
(metrocluster node show)
dr-group-id cluster node node-uuid
node-cluster-uuid
----------- --------- -------- ------------------------------------
------------------------------
1 cluster_A node_A_1-A700 f03cb63c-9a7e-11e7-b68b-00a098908039
ee7db9d5-9a82-11e7-b68b-00a098908039
1 cluster_A node_A_2-A700 aa9a7a7a-9a81-11e7-a4e9-00a098908c35
ee7db9d5-9a82-11e7-b68b-00a098908039
1 cluster_B node_B_1-A700 f37b240b-9ac1-11e7-9b42-00a098c9e55d
07958819-9ac6-11e7-9b42-00a098c9e55d
1 cluster_B node_B_2-A700 bf8e3f8f-9ac4-11e7-bd4e-00a098ca379f
07958819-9ac6-11e7-9b42-00a098c9e55d
4 entries were displayed.
cluster_B::*
It is recommended that you record the UUIDs into a table similar to the following.
node_B_1-A700 f37b240b-9ac1-11e7-9b42-00a098c9e55d
node_B_2-A700 bf8e3f8f-9ac4-11e7-bd4e-00a098ca379f
cluster_A ee7db9d5-9a82-11e7-b68b-00a098908039
node_A_1-A700 f03cb63c-9a7e-11e7-b68b-00a098908039
node_A_2-A700 aa9a7a7a-9a81-11e7-a4e9-00a098908c35
4. If the MetroCluster nodes are in a SAN configuration, collect the relevant information.
106
and passphrases. For additional information, see Backing up onboard key management information
manually.
a. If Onboard Key Manager is configured: security key-manager onboard show-backup
You will need the passphrase later in the upgrade procedure.
b. If enterprise key management (KMIP) is configured, issue the following commands:
7. Gather the system IDs of the existing nodes: metrocluster node show -fields node-
systemid,ha-partner-systemid,dr-partner-systemid,dr-auxiliary-systemid
Before the upgrading the platforms, you must remove monitoring if the MetroCluster configuration is monitored
with the Tiebreaker or Mediator utility.
Steps
1. Collect the output for the following command:
2. Remove the existing MetroCluster configuration from Tiebreaker, Mediator, or other software that can
initiate switchover.
107
Tiebreaker Removing MetroCluster Configurations in the
MetroCluster Tiebreaker Installation and
Configuration content
metrocluster configuration-settings
mediator remove
Before performing the maintenance, you should issue an AutoSupport message to notify technical support that
maintenance is underway. Informing technical support that maintenance is underway prevents them from
opening a case on the assumption that a disruption has occurred.
Steps
1. Log in to the cluster.
2. Invoke an AutoSupport message indicating the start of the maintenance:
The maintenance-window-in-hours parameter specifies the length of the maintenance window, with a
maximum of 72 hours. If the maintenance is completed before the time has elapsed, you can invoke an
AutoSupport message indicating the end of the maintenance period:
After completing this task, site_A is active and serving data for both sites. site_B is inactive, and ready to begin
the upgrade process.
108
Steps
1. Switch over the MetroCluster configuration to site_A so that site_B’s nodes can be upgraded:
a. Issue the following command on site_A:
c. After the operation is complete, confirm that the nodes are in switchover state:
metrocluster show
Automatic healing of aggregates after negotiated switchover is disabled during controller upgrade.
Nodes at site_B are halted and stopped at the LOADER prompt.
109
Steps
1. Gather the bootarg values from both nodes at site_B: printenv
2. Power off the chassis at site_B.
Use the following procedure to remove the AFF A700 controller module
Steps
1. Detach the console cable, if any, and the management cable from the controller module before removing
the controller module.
2. Unlock and remove the controller module from the chassis.
a. Slide the orange button on the cam handle downward until it unlocks.
Cam handle
b. Rotate the cam handle so that it completely disengages the controller module from the chassis, and
then slide the controller module out of the chassis. Make sure that you support the bottom of the
controller module as you slide it out of the chassis.
Use the following procedure to remove the AFF A700 NVS module.
Note: The AFF A700 NVS module NVS module is in slot 6 and is double the height compared to other modules
in the system.
Steps
110
1. Unlock and remove the NVS from slot 6.
a. Depress the lettered and numbered 'cam' button. The cam button moves away from the chassis.
b. Rotate the cam latch down until it is in a horizontal position. The NVS disengages from the chassis and
moves a few inches.
c. Remove the NVS from the chassis by pulling on the pull tabs on the sides of the module face.
2. If you are using add-on modules used as coredump devices on the AFF A700 NVS, do not transfer them to
the AFF A900 NVS. Do not transfer any parts from the AFF A700 controller module and NVS to the AFF
A900.
Use the following procedure to install the AFF A900 NVS in slot 6 of both nodes at site_B.
Steps
1. Align the NVS with the edges of the chassis opening in slot 6.
2. Gently slide the NVS into the slot until the lettered and numbered I/O cam latch begins to engage with the
I/O cam pin, and then push the I/O cam latch all the way up to lock the NVS in place.
111
Lettered and numbered I/O cam latch
Use the following procedure to install the AFF A900 controller module.
Steps
1. Align the end of the controller module with the opening in the chassis, and then gently push the controller
module halfway into the system.
2. Firmly push the controller module into the chassis until it meets the midplane and is fully seated. The
locking latch rises when the controller module is fully seated. Attention: To avoid damaging the connectors,
do not use excessive force when sliding the controller module into the chassis.
3. Cable the management and console ports to the controller module.
112
Cam handle release button
Cam handle
Slot 7 on all nodes of the cluster should be empty as mentioned in Map ports from the
old nodes to the new nodes section.
After swapping the AFF A900 controller module and NVS, you need to netboot the AFF A900 nodes and install
the same ONTAP version and patch level that is running on the cluster. The term netboot means you are
booting from an ONTAP image stored on a remote server. When preparing for netboot, you must add a copy of
the ONTAP 9 boot image onto a web server that the system can access. It is not possible to check the version
of ONTAP installed on the boot media of an AFF A900 controller module unless it is installed in a chassis and
powered ON. The ONTAP version on the AFF A900 boot media must be the same as the ONTAP version
running on the AFF A700 system that is being upgraded and both the primary and backup boot images should
match. You can configure the images by performing a netboot followed by the wipeconfig command from the
boot menu. If the controller module was previously used in another cluster, the wipeconfig command clears
any residual configuration on the boot media.
Steps
1. Access the NetApp Support Site to download the files used for performing the netboot of the system.
2. Download the appropriate ONTAP software from the software download section of the NetApp Support
Site and store the ontap-version_image.tgz file on a web-accessible directory.
113
3. Change to the web-accessible directory and verify that the files you need are available.
4. Your directory listing should contain <ontap_version>\_image.tgz.
5. Configure the netboot connection by choosing one of the following actions.
You should use the management port and IP as the netboot connection. Do not use a data
LIF IP or a data outage might occur while the upgrade is being performed.
7. Wait for the node_B_1 now running on the AFF A900 controller module to boot and display the boot menu
options as shown below:
114
Please choose one of the following:
8. From the boot menu, select option (7) Install new software first. This menu option downloads
and installs the new ONTAP image to the boot device. NOTE: Disregard the following message: This
procedure is not supported for Non-Disruptive Upgrade on an HA pair. This note
applies to nondisruptive ONTAP software upgrades, and not controller upgrades.
Always use netboot to update the new node to the desired image. If you use another method to install the
image on the new controller, the incorrect image might install. This issue applies to all ONTAP releases.
9. If you are prompted to continue the procedure, enter y, and when prompted for the package, enter the
URL: http://<web_server_ip/path_to_web-accessible_directory>/
<ontap_version>\_image.tgz
10. Complete the following substeps to reboot the controller module:
a. Enter n to skip the backup recovery when you see the following prompt: Do you want to restore
the backup configuration now? {y|n}
b. Enter y to reboot when you see the following prompt: `The node must be
rebooted to start using the newly installed software. Do you want to reboot
now? {y|n} The controller module reboots but stops at the boot menu because the boot device was
reformatted, and the configuration data needs to be restored.
11. At the prompt, run the wipeconfig command to clear any previous configuration on the boot media:
a. When you see the following message, answer yes: This will delete critical system
configuration, including cluster membership. Warning: do not run this option
on a HA node that has been taken over. Are you sure you want to continue?:
b. The node reboots to finish the wipeconfig and then stops at the boot menu.
12. Select option 5 to go to maintenance mode from the boot menu. Answer yes to the prompts until the node
stops at maintenance mode and the command prompt \*>.
13. Repeat these steps to netboot node_B_2.
Depending on the presence and configuration of HBA cards in the controller module, you need to configure
115
them correctly for your site’s usage.
Steps
1. In Maintenance mode configure the settings for any HBAs in the system:
ucadmin show
If you have this type of HBA and desired mode… Use this command…
CNA FC ucadmin modify -m fc -t initiator
adapter-name
halt
After you run the command, wait until the node stops at the LOADER prompt.
3. Boot the node back into Maintenance mode to enable the configuration changes to take effect:
boot_ontap maint
FC fcadmin show
You must verify the HA state of the controllers and chassis, and, if necessary, update the state to match your
system configuration.
Steps
1. In Maintenance mode, display the HA state of the controller module and chassis:
ha-config show
116
The HA state for all components should be mccip.
2. If the displayed system state of the controller or chassis is not correct, set the HA state:
4. On each node, check the system date, time, and time zone: show date
5. If necessary, set the date in UTC or GMT: set date <mm/dd/yyyy>
6. Check the time by using the following command at the boot environment prompt: show time
7. If necessary, set the time in UTC or GMT: set time <hh:mm:ss>
8. Save the settings: saveenv
9. Gather environment variables: printenv
The switches at site_A will be upgraded when the controllers on site_A are upgraded.
Steps
1. Prepare the IP switches for the application of the new RCF files.
Follow the steps in the section for your switch vendor from the MetroCluster IP Installation and
Configuration section.
Follow the steps in the section for your switch vendor from the MetroCluster IP installation and
configuration.
117
Configure the new controllers
New controllers should be ready and cabled at this point.
Certain MetroCluster IP bootarg values must be configured on the new controller modules. The values must
match those configured on the old controller modules.
Steps
1. At the LOADER> prompt, set the following bootargs on the new nodes at site_B:
The following example sets the values for node_B_1-A900 using VLAN 120 for the first network and VLAN
130 for the second network:
setenv bootarg.mcc.port_a_ip_config
172.17.26.10/23,0,172.17.26.11,172.17.26.13,172.17.26.12,120
setenv bootarg.mcc.port_b_ip_config
172.17.27.10/23,0,172.17.27.11,172.17.27.13,172.17.27.12,130
The following example sets the values for node_B_2-A900 using VLAN 120 for the first network and VLAN
130 for the second network:
setenv bootarg.mcc.port_a_ip_config
172.17.26.11/23,0,172.17.26.10,172.17.26.12,172.17.26.13,120
setenv bootarg.mcc.port_b_ip_config
172.17.27.11/23,0,172.17.27.10,172.17.27.12,172.17.27.13,130
118
a. Set the UUIDs on node_B_1-A900.
The following example shows the commands for setting the UUIDs on node_B_1-A900:
The following example shows the commands for setting the UUIDs on node_B_2-A900:
3. If the original systems were configured for ADP, at each of the replacement nodes' LOADER prompt,
enable ADP:
The following example shows the commands for setting the values on node_B_1-A900:
119
setenv bootarg.mcc.local_config_id 537403322
setenv bootarg.mcc.dr_partner 537403324
The following example shows the commands for setting the values on node_B_2-A900:
5. If using encryption with external key manager, set the required bootargs:
setenv bootarg.kmip.init.ipaddr
setenv bootarg.kmip.kmip.init.netmask
setenv bootarg.kmip.kmip.init.gateway
setenv bootarg.kmip.kmip.init.interface
Reassign the root aggregate disks to the new controller module, using the sysids gathered earlier.
Steps
1. Boot the system to Maintenance mode:
boot_ontap maint
disk show -a
The command output shows the system ID of the new controller module (1574774970). However, the root
aggregate disks are still owned by the old system ID (537403322). This example does not show drives
owned by other nodes in the MetroCluster configuration.
120
*> disk show -a
Local System ID: 1574774970
DISK OWNER POOL SERIAL NUMBER HOME
DR HOME
------------ --------- ----- -------------
------------- -------------
prod3-rk18:9.126L44 node_B_1-A700(537403322) Pool1 PZHYN0MD
node_B_1-A700(537403322) node_B_1-A700(537403322)
prod4-rk18:9.126L49 node_B_1-A700(537403322) Pool1 PPG3J5HA
node_B_1-A700(537403322) node_B_1-700(537403322)
prod4-rk18:8.126L21 node_B_1-A700(537403322) Pool1 PZHTDSZD
node_B_1-A700(537403322) node_B_1-A700(537403322)
prod2-rk18:8.126L2 node_B_1-A700(537403322) Pool0 S0M1J2CF
node_B_1-(537403322) node_B_1-A700(537403322)
prod2-rk18:8.126L3 node_B_1-A700(537403322) Pool0 S0M0CQM5
node_B_1-A700(537403322) node_B_1-A700(537403322)
prod1-rk18:9.126L27 node_B_1-A700(537403322) Pool0 S0M1PSDW
node_B_1-A700(537403322) node_B_1-A700(537403322)
.
.
.
3. Reassign the root aggregate disks on the drive shelves to the new controllers.
4. Reassign the root aggregate disks on the drive shelves to the new controllers:
121
*> disk reassign -s 537403322 -d 1574774970
Partner node must not be in Takeover mode during disk reassignment from
maintenance mode.
Serious problems could result!!
Do not proceed with reassignment if the partner is in takeover mode.
Abort reassignment (y/n)? n
After the node becomes operational, you must perform a takeover and
giveback of the HA partner node to ensure disk reassignment is
successful.
Do you want to continue (y/n)? y
Disk ownership will be updated on all disks previously belonging to
Filer with sysid 537403322.
Do you want to continue (y/n)? y
5. Verify that the disks of the root aggregate are correctly reassigned old-remove:
disk show
122
*> disk show
Local System ID: 537097247
You must boot the new controllers, taking care to ensure that the bootarg variables are correct and, if needed,
perform the encryption recovery steps.
Steps
1. Halt the new nodes:
halt
123
setenv bootarg.kmip.init.interface interface-id
printenv partner-sysid
boot_ontap menu
5. If root encryption is used, select the boot menu option for your key management configuration.
External key management Option 11 and follow the prompts to provide the
required inputs to recover or restore the key-
manager configuration
6. From the boot menu, select (6) Update flash from backup config.
Respond y to the system id change prompts. Wait for the second reboot messages:
On each node, check the bootargs set in Setting the MetroCluster IP bootarg variables and
correct any incorrect values. Only move to the next step after you have checked the bootarg
values.
printenv partner-sysid
124
9. If root encryption is used, select the boot menu option for your key management configuration.
External key management Option 11 and follow the prompts to provide the
required inputs to recover or restore the key-
manager configuration
You need to perform the recovery procedure by selecting Option 10 or option 11 depending on the key
manager setting and Option 6 at the boot menu prompt. To boot the nodes completely, you might need to
perform the recovery procedure continued by option 1 (normal boot).
10. Wait for the new nodes, node_B_1-A900 and node_B_2-A900 to boot up.
If either node is in takeover mode, perform a giveback using the storage failover giveback
command.
11. If encryption is used, restore the keys using the correct command for your key management configuration.
VLAN and interface group membership might be different than that of the old node.
Creating a VLAN
125
Combining physical ports to create interface groups
Verify that LIFs are hosted on appropriate nodes and ports as mapped out at the beginning of the upgrade
procedure.
Steps
1. Verify that LIFs are hosted on the appropriate node and ports prior to switchback.
a. Change to the advanced privilege level:
When entering the network interface modify command within the vserver config override
command, you cannot use the tab autocomplete feature. You can create the network interface
modify using autocomplete and then enclose it in the vserver config override command.
126
Steps
1. Issue the metrocluster node show command from site_B and check the output.
a. Verify that the new nodes are represented correctly.
b. Verify that the new nodes are in "Waiting for switchback state."
2. Perform the healing and switchback by running the required commands from any node in the active cluster
(the cluster that is not undergoing upgrade).
a. Heal the data aggregates:
metrocluster heal aggregates
b. Heal the root aggregates:
metrocluster switchback
metrocluster show
The switchback operation is still in progress when the output displays waiting-for-switchback:
127
cluster_B::> metrocluster show
Cluster Entry Name State
------------------------- ------------------- -----------
Local: cluster_B Configuration state configured
Mode switchover
AUSO Failure Domain -
Remote: cluster_A Configuration state configured
Mode waiting-for-switchback
AUSO Failure Domain -
If a switchback takes a long time to finish, you can check on the status of in-progress baselines by using
the metrocluster config-replication resync-status show command. This command is at the
advanced privilege level.
Steps
1. Verify the operation of the MetroCluster configuration:
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
b. Perform a MetroCluster check:
metrocluster check run
c. Display the results of the MetroCluster check:
128
a. Check the MetroCluster IP connections:
Steps
1. Repeat the steps to upgrade the nodes on site_A, beginning with Prepare for the upgrade.
As you perform the tasks, all example references to the sites and nodes are reversed. For example, when
the example is given to switchover from site_A, you will switchover from site_B.
Steps
1. Restore monitoring if necessary, using the procedure for your configuration.
129
Steps
1. To resume automatic support case generation, send an Autosupport message to indicate that the
maintenance is complete.
a. Issue the following command:
system node autosupport invoke -node * -type all -message MAINT=end
b. Repeat the command on the partner cluster.
Steps
1. Gather information from the old nodes.
At this stage, the four-node configuration appears as shown in the following image:
2. Perform all of the steps in the four-node expansion procedure for your MetroCluster type.
When the expansion procedure is complete, the configuration appears as shown in the following image:
130
3. Move the CRS volumes.
4. Move the data from the old nodes to new nodes using the following procedures from Other platform
procedures: Controller Hardware Upgrade Express.
a. Perform all the steps in Creating an aggregate and moving volumes to the new nodes.
b. Perform all the steps in Moving non-SAN data LIFs and cluster management LIFs to the new nodes.
c. Perform all the steps in Deleting SAN LIFs from the original nodes.
5. Follow the steps in the procedure for removing the old DR group.
After you have removed the old DR group (DR group one), the configuration appears as shown in the
following image:
131
Refreshing a four-node MetroCluster IP configuration
(ONTAP 9.8 and later)
Beginning with ONTAP 9.8, you can upgrade the controllers and storage in a four-node
MetroCluster IP configuration by expanding the configuration to become a temporary
eight-node configuration and then removing the old disaster recovery (DR) group.
About this task
• This procedure is supported on systems running ONTAP 9.8 and later.
• If you are upgrading the IP switches, they should be upgraded before to performing this refresh procedure.
• References to "old nodes" mean the nodes that you intend to replace.
• This procedure is not supported on AFF A320 systems configured with Broadcom BES-53248 switches.
• This procedure is not supported on FAS2750 or AFF A220 systems.
Steps
1. Gather information from the old nodes.
At this stage, the four-node configuration appears as shown in the following image:
132
2. To prevent automatic support case generation, send an AutoSupport message to indicate the upgrade is
underway.
a. Issue the following command:
system node autosupport invoke -node * -type all -message "MAINT=10h
Upgrading old-model to new-model"
The following example specifies a 10 hour maintenance window. You might want to allow additional
time depending on your plan.
If the maintenance is completed before the time has elapsed, you can invoke an AutoSupport message
indicating the end of the maintenance period:
metrocluster configuration-settings
mediator remove
When the expansion procedure is complete, the configuration appears as shown in the following image:
133
5. Move the CRS volumes.
6. Move the data from the old nodes to new nodes using the following procedures in Other platform
procedures: Controller Hardware Upgrade Express Guide
a. Perform all the steps in Creating an aggregate and moving volumes to the new nodes.
b. Perform all the steps in Moving non-SAN data LIFs and cluster management LIFs to the new nodes.
7. Follow the steps in the procedure for removing the old DR group.
After you have removed the old DR group (DR group one), the configuration appears as shown in the
following image:
134
8. Confirm the operational mode of the MetroCluster configuration and perform a MetroCluster check.
a. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
10. To resume automatic support case generation, send an Autosupport message to indicate that the
maintenance is complete.
a. Issue the following command:
135
b. Repeat the command on the partner cluster.
• If the platforms in your two-node configuration are not supported in ONTAP 9.2 and you plan to upgrade to
platforms supported in ONTAP 9.2 and expand to a four-node cluster, you must upgrade the platforms in
the two-node configuration before expanding the MetroCluster FC configuration.
• The existing MetroCluster FC configuration must be healthy.
• The equipment you are adding must be supported and meet all of the requirements described in the
following procedures:
• You must have available FC switch ports to accommodate the new controllers and any new bridges.
• You need the admin password and access to an FTP or SCP server.
After completing this procedure, the MetroCluster FC configuration consists of two HA pairs, one at each
site:
136
• Both sites must be expanded equally.
• This procedure can take over an hour per site, with additional time for tasks such as initializing the disks
and netbooting the new nodes.
The time to initialize the disks depends on the size of the disks.
137
relationships between them, that the controllers are in normal mode, and that the
aggregates are mirrored.
Steps
1. Display the details of the nodes in the MetroCluster configuration from any node in the configuration:
The following output shows that this MetroCluster configuration has a single DR group and one node in
each cluster.
metrocluster show
The following output shows that the existing nodes in the MetroCluster configuration are in normal mode:
Configuration: two-node-fabric
3. Check the state of the aggregates on each node in the MetroCluster configuration:
138
storage aggregate show
The following output shows that the aggregates on cluster_A are online and mirrored:
cluster_A::>
Steps
1. Log in to the cluster at Site_A.
2. Invoke an AutoSupport message indicating the start of the maintenance:
The maintenance-window-in-hours parameter specifies the length of the maintenance window and
can be a maximum of 72 hours. If you complete the maintenance before the time has elapsed, you can
issue the following command to indicate that the maintenance period has ended:
139
Zoning for the new controller ports when adding a controller module in a fabric-
attached MetroCluster configuration
The FC switch zoning must accommodate the new controller connections. If you used the
NetApp-supplied reference configuration files (RCFs) to configure your switches, the
zoning is preconfigured and you do not need to make any changes.
If you manually configured your FC switches, you must ensure that the zoning is correct for the initiator
connections from the new controller modules. See the sections on zoning in Fabric-attached MetroCluster
installation and configuration.
You must add a new controller module to each site, creating an HA pair in each site. This
is a multistep process involving both hardware and software changes that must be
performed in the proper order at each site.
About this task
• The new controller module must be received from NetApp as part of the upgrade kit.
You should verify that PCIe cards in the new controller module are compatible and supported by the new
controller module.
• Your system must have an empty slot available for the new controller module when upgrading to a single-
chassis HA pair (an HA pair in which both controller modules reside in the same chassis).
This configuration is not supported on all systems. Platforms with single chassis
configurations that are supported in ONTAP 9 are AFF A300, FAS8200, FAS8300, AFF
A400, AFF80xx, FAS8020, FAS8060, FAS8080, and FAS9000.
• You must have rack space and cables for the new controller module when upgrading to a dual-chassis HA
pair (an HA pair in which the controller modules reside in separate chassis).
• You must connect each controller module to the management network through its e0a port or, if your
system has one, you can connect to the e0M port as the management port.
• These tasks must be repeated at each site.
• The preexisting controller modules are referred to as the existing controller modules.
• The controller modules that are being added are referred to as the new controller modules; the examples in
this procedure have the console prompt new_ctlr>.
• This task uses the following workflow:
140
Preparing for the upgrade
Before upgrading to an HA pair, you must verify that your system meets all requirements
and that you have all of the necessary information.
Steps
1. You need to identify unassigned disks or spare disks that you can assign to the new controller module.
2. Based on the results of the previous step, perform either of the following:
141
Not enough spare disks Contact technical support for more information.
available for the new
controller module on a
system without root-data
partitioning
a. Determine where the aggregates for the existing node are located:
d. Repeat the previous step for as many disks as you need for the new node.
3. Verify that you have cables ready for the following connections:
◦ Cluster connections
If you are creating a two-node switchless cluster, you require two cables to connect the controller
modules. Otherwise, you require a minimum of four cables, two for each controller module connection
to the cluster-network switch. Other systems (like the 80xx series) have defaults of either four or six
cluster connections.
6. Gather all of the IP addresses and other network parameters for the new controller module.
Before using a new controller module in the MetroCluster configuration, you must clear
the existing configuration.
Steps
1. If necessary, halt the node to display the LOADER prompt:
halt
142
set-defaults
saveenv
boot_ontap menu
wipeconfig
6. At the boot menu, select option 5 to boot the system into Maintenance mode.
Before installing a new controller module, you must configure cluster ports on the existing
controller module so that the cluster ports can provide cluster communication with the
new controller module.
About this task
If you are creating a two-node switchless cluster (with no cluster network switches), you must enable the
switchless cluster networking mode.
For detailed information about port, LIF, and network configuration in ONTAP, see Network Management.
Steps
1. Determine which ports should be used as the node’s cluster ports.
For a list of the default port roles for your platform, see the Hardware Universe
The Installation and Setup Instructions for your platform on the NetApp Support Site contains information
about the ports for cluster network connections.
In the following example, ports “e0a”, “e0b”, “e0c”, and “e0d” must be changed to cluster ports:
143
cluster_A::> network port show
Node: controller_A_1
Speed(Mbps) Health
Port IPspace Broadcast Domain Link MTU Admin/Oper Status
--------- ------------ ---------------- ---- ---- -----------
--------
e0M Default mgmt_bd_1500 up 1500 auto/1000 healthy
e0a Default Default up 1500 auto/10000 healthy
e0b Default Default up 1500 auto/10000 healthy
e0c Default Default up 1500 auto/10000 healthy
e0d Default Default up 1500 auto/10000 healthy
e0i Default Default down 1500 auto/10 -
e0j Default Default down 1500 auto/10 -
e0k Default Default down 1500 auto/10 -
e0l Default Default down 1500 auto/10 -
e2a Default Default up 1500 auto/10000 healthy
e2b Default Default up 1500 auto/10000 healthy
e4a Default Default up 1500 auto/10000 healthy
e4b Default Default up 1500 auto/10000 healthy
13 entries were displayed.
3. For any data LIF that is using a cluster port as the home-port or current-port, modify the LIF to use a data
port as its home-port:
The following example changes the home port of a data LIF to a data port:
4. For each LIF that you modified, revert the LIF to its new home port:
The following example reverts the LIF “datalif1” to its new home port “e1b”:
5. Remove any VLAN ports using cluster ports as member ports and ifgrps using cluster ports as member
ports.
a. Delete VLAN ports:
network port vlan delete -node node-name -vlan-name portid-vlandid
144
For example:
For example:
network port ifgrp remove-port -node node1 -ifgrp a1a -port e0d
d. Modify interface group ports to use other physical ports as member as needed.:
ifgrp add-port -node node-name -ifgrp interface-group-name -port port-id
6. Verify that the port roles have changed:
The following example shows that ports “e0a”, “e0b”, “e0c”, and “e0d” are now cluster ports:
Node: controller_A_1
Speed(Mbps) Health
Port IPspace Broadcast Domain Link MTU Admin/Oper Status
--------- ------------ ---------------- ---- ---- -----------
--------
e0M Default mgmt_bd_1500 up 1500 auto/1000 healthy
e0a Cluster Cluster up 9000 auto/10000 healthy
e0b Cluster Cluster up 9000 auto/10000 healthy
e0c Cluster Cluster up 9000 auto/10000 healthy
e0d Cluster Cluster up 9000 auto/10000 healthy
e0i Default Default down 1500 auto/10 -
e0j Default Default down 1500 auto/10 -
e0k Default Default down 1500 auto/10 -
e0l Default Default down 1500 auto/10 -
e2a Default Default up 1500 auto/10000 healthy
e2b Default Default up 1500 auto/10000 healthy
e4a Default Default up 1500 auto/10000 healthy
e4b Default Default up 1500 auto/10000 healthy
13 entries were displayed.
145
7. Add the ports to the cluster broadcast domain:
For example:
8. If your system is part of a switched cluster, create cluster LIFs on the cluster ports: network interface
create
The following example creates a cluster LIF on one of the node’s cluster ports. The -auto parameter
configures the LIF to use a link-local IP address.
9. If you are creating a two-node switchless cluster, enable the switchless cluster networking mode:
a. Change to the advanced privilege level from either node:
You can respond y when prompted whether you want to continue into advanced mode. The advanced
mode prompt appears (*>).
Cluster interface creation for the existing node in a two-node switchless cluster system is
completed after cluster setup is completed through a netboot on the new controller module.
When you are ready to prepare the netboot server, you must download the correct
ONTAP netboot image from the NetApp Support Site to the netboot server and note the
IP address.
About this task
• You must be able to access an HTTP server from the system before and after adding the new controller
module.
146
• You must have access to the NetApp Support Site to download the necessary system files for your platform
and your version of ONTAP.
• Both controller modules in the HA pair must run the same version of ONTAP.
Steps
1. Download the appropriate ONTAP software from the software download section of the NetApp Support Site
and store the <ontap_version>_image.tgz file on a web-accessible directory.
2. Change to the web-accessible directory and verify that the files you need are available.
For… Then…
FAS2200, FAS2500, FAS3200, FAS6200, Extract the contents of the
FAS/AFF8000 series systems <ontap_version>_image.tgz file to the target
directory:
netboot/kernel
All other systems Your directory listing should contain the following
file:
<ontap_version>_image.tgz
You must use the storage failover modify command to set the mode on the existing
controller module. The mode value is enabled later, after you reboot the controller
module.
147
Steps
1. Set the mode to HA:
You must perform a clean shutdown of the existing controller module to verify that all of
the data has been written to disk. You must also disconnect the power supplies.
About this task
You must perform a clean system shutdown before replacing the system components to avoid
losing unwritten data in the NVRAM or NVMEM.
Steps
1. Halt the node from the existing controller module prompt:
If you are prompted to continue the halt procedure, enter y when prompted, and then wait until the system
stops at the LOADER prompt.
In an 80xx system, the NVRAM LED is located on the controller module to the right of the network ports,
marked with a battery symbol.
This LED blinks if there is unwritten data in the NVRAM. If this LED is flashing amber after you enter the
halt command, you need to reboot your system and try halting it again.
You must physically install the new controller module in the chassis, and then cable it.
Steps
1. If you have an I/O expansion module (IOXM) in your system and are creating a single-chassis HA pair, you
must uncable and remove the IOXM.
148
You can then use the empty bay for the new controller module. However, the new configuration will not
have the extra I/O provided by the IOXM.
2. Physically install the new controller module and, if necessary, install additional fans:
In a separate chassis from its HA partner to create a Install the new system in the rack or system cabinet.
dual-chassis HA pair when the existing
configuration is in a controller-IOX module
configuration.
• FAS8200
• 80xx
a. Identify the ports on the controller module for the cluster connections.
b. If you are configuring a switched cluster, identify the ports that you will use on the cluster network
switches.
See the Clustered Data ONTAP Switch Setup Guide for Cisco Switches, NetApp 10G Cluster-Mode
Switch Installation Guide or NetApp 1G Cluster-Mode Switch Installation Guide, depending on what
switches you are using.
149
A two-node switchless cluster Directly connect the cluster ports on the existing
controller module to the corresponding cluster ports
on the new controller module.
Cabling the new controller module’s FC-VI and HBA ports to the FC switches
The new controller module’s FC-VI ports and HBAs (host bus adapters) must be cabled
to the site FC switches.
Steps
1. Cable the FC-VI ports and HBA ports, using the table for your configuration and switch model.
◦ Port assignments for FC switches when using ONTAP 9.1 and later
◦ Port assignments for FC switches when using ONTAP 9.0
◦ Port assignments for systems using two initiator ports
You must cable the new controller module to the cluster peering network so that it has
connectivity with the cluster on the partner site.
About this task
At least two ports on each controller module should be used for cluster peering.
The recommended minimum bandwidth for the ports and network connectivity is 1 GbE.
Steps
1. Identify and cable at least two ports for cluster peering and verify they have network connectivity with the
partner cluster.
You power up the existing controller module and the new controller module to display the
LOADER prompt.
Steps
Power up the controller modules and interrupt the boot process, using the steps for your configuration:
150
In the same chassis a. Verify that the new controller module is not fully inserted into the bay.
The existing controller module should be fully inserted into the bay because it
was never removed from the chassis, but the new controller module should
not be.
b. Connect the power and turn on the power supplies so that the existing
controller module receives power.
c. Interrupt the boot process on the existing controller module by pressing Ctrl-
C.
d. Push the new controller module firmly into the bay.
When fully seated, the new controller module receives power and
automatically boots.
In separate chassis a. Turn on the power supplies on the existing controller module.
b. Interrupt the boot process by pressing Ctrl-C.
c. Repeat these steps for the new controller module
Each controller module should display the LOADER prompt (LOADER>, LOADER-A>, or LOADER-B>).
If there is no LOADER prompt, record the error message and contact technical support. If the
system displays the boot menu, reboot and attempt to interrupt the boot process again.
Changing the ha-config setting on the existing and new controller modules
When you expand a MetroCluster configuration, you must update the ha-config setting of
the existing controller module and the new controller module. You must also determine
the system ID of the new controller module.
About this task
This task is performed in Maintenance mode on both the existing and new controller modules.
Steps
1. Change the ha-config setting of the existing controller module:
a. Display the ha-config setting of the existing controller module and chassis:
ha-config show
The ha-config setting is “mcc-2n” for all components because the controller module was in a two-node
MetroCluster configuration.
151
b. Change the ha-config setting of the existing controller module to “mcc”:
ha-config modify controller mcc
c. Change the ha-config setting of the existing chassis to “mcc”:
sysconfig
Note the system ID. You need it when you set the partner ID on the new controller module.
halt
2. Change the ha-config setting and retrieve the system ID of the new controller module:
a. If the new controller module is not already in Maintenance mode, boot it to Maintenance mode:
boot_ontap maint
sysconfig
Note the system ID. You need it when you set the partner ID and assign disks to the new controller
module.
halt
You must set the partner system ID on both controller modules so that they can form an
HA pair.
About this task
This task is performed with both controller modules at the LOADER prompt.
Steps
1. On the existing controller module, set the partner system ID to that of the new controller module:
152
2. On the new controller module, set the partner system ID to that of the existing controller module:
boot_ontap
Before you complete the configuration of the new controller module through netboot, you
must assign disks to it.
About this task
You must have made sure that there are enough spares, unassigned disks, or assigned disks that are not part
of an existing aggregate.
Steps
1. Assign the root disk to the new controller module:
If your platform model uses the Advanced Drive Partitioning (ADP) feature, you must include the -root true
parameter:
2. Assign the remaining required disks to the new controller module by entering the following command for
each disk:
Ensure that you have assigned all disks that you intend to assign to the new node.
153
Netbooting and setting up ONTAP on the new controller module
You must perform a specific sequence of steps to netboot and install the ONTAP
operating system on the new controller module when adding controller modules to an
existing MetroCluster configuration.
About this task
• This task starts at the LOADER prompt of the new controller module.
• This task includes initializing disks.
The amount of time you need to initialize the disks depends on the size of the disks.
• The system automatically assigns two disks to the new controller module.
Steps
1. At the LOADER prompt, configure the IP address of the new controller module based on DHCP availability:
dns_domain is the Domain Name System (DNS) domain name. If you use
this optional parameter, you do not need a fully qualified domain name in the
netboot server URL; you need only the server’s host name.
154
FAS2200, FAS2500, netboot http://web_server_ip/path_to_web-
FAS3200, FAS6200, accessible_directory/netboot/kernel
FAS/AFF8000 series
systems
3. Select the Install new software first option from the displayed menu.
This menu option downloads and installs the new ONTAP image to the boot device.
◦ You should enter “y” when prompted with the message that this procedure is not supported for
nondisruptive upgrade on an HA pair.
◦ You should enter “y” when warned that this process replaces the existing ONTAP software with new
software.
◦ You should enter the path as follows when prompted for the URL of the image.tgz file:
http://path_to_the_web-accessible_directory/image.tgz
4. Enter “y” when prompted regarding nondisruptive upgrade or replacement of the software.
5. Enter the path to the image.tgz file when prompted for the URL of the package.
6. Enter “n” to skip the backup recovery when prompted to restore the backup configuration.
****************************************************************
* Restore Backup Configuration *
* This procedure only applies to storage controllers that *
* are configured as an HA pair. *
* *
* Choose Yes to restore the "varfs" backup configuration *
* from the SSH server. Refer to the Boot Device Replacement *
* guide for more details. *
* Choose No to skip the backup recovery and return to the *
* boot menu. *
****************************************************************
155
7. Enter “y” when prompted to reboot now.
The node must be rebooted to start using the newly installed software.
Do you want to
reboot now? {y|n} `y`
8. If necessary, select the option to Clean configuration and initialize all disks after the node has booted.
Because you are configuring a new controller module and the new controller module’s disks are empty, you
can respond “y” when the system warns you that this will erase all disks.
The amount of time needed to initialize disks depends on the size of your disks and
configuration.
9. After the disks are initialized and the Cluster Setup wizard starts, set up the node:
10. Log in to the node, and enter the cluster setup and then enter “join” when prompted to join the cluster.
The Setup ONTAP for your version of ONTAP contains additional details.
12. If the system is in a two-node switchless cluster configuration, create the cluster interfaces on the existing
node using the network interface create command to create cluster LIFs on the cluster ports.
The following is an example command for creating a cluster LIF on one of the node’s cluster ports. The
-auto parameter configures the LIF to use a link-local IP address.
13. After setup is complete, verify that the node is healthy and eligible to participate in the cluster:
cluster show
The following example shows a cluster after the second node (cluster1-02) has been joined to it:
156
cluster_A::> cluster show
Node Health Eligibility
--------------------- ------- ------------
node_A_1 true true
node_A_2 true true
You can access the Cluster Setup wizard to change any of the values you entered for the admin storage
virtual machine (SVM) or node SVM by using the cluster setup command.
14. Confirm that you have four ports configured as cluster interconnects:
The following example shows output for two controller modules in cluster_A:
157
Mirroring the root aggregate on the new controller
You must mirror the root aggregate to provide data protection when you are adding a
controller to a MetroCluster configuration.
This task must be performed on the new controller module.
This mirrors the aggregate, so it consists of a local plex and a remote plex located at the remote
MetroCluster site.
You can configure intercluster LIFs on dedicated ports. Doing so typically increases the
available bandwidth for replication traffic.
Steps
1. List the ports in the cluster:
158
cluster01::> network port show
Speed
(Mbps)
Node Port IPspace Broadcast Domain Link MTU Admin/Oper
------ --------- ------------ ---------------- ----- -------
------------
cluster01-01
e0a Cluster Cluster up 1500 auto/1000
e0b Cluster Cluster up 1500 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
e0e Default Default up 1500 auto/1000
e0f Default Default up 1500 auto/1000
cluster01-02
e0a Cluster Cluster up 1500 auto/1000
e0b Cluster Cluster up 1500 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
e0e Default Default up 1500 auto/1000
e0f Default Default up 1500 auto/1000
The following example shows that ports "e0e" and "e0f" have not been assigned LIFs:
159
network interface failover-groups create -vserver system_SVM -failover-group
failover_group -targets physical_or_logical_ports
The following example assigns ports "e0e" and "e0f" to the failover group "intercluster01" on the system
SVM "cluster01":
5. Create intercluster LIFs on the system SVM and assign them to the failover group.
160
9.5 and earlier network interface create -vserver system_SVM -lif LIF_name
-role intercluster -home-node node -home-port port
-address port_IP -netmask netmask -failover-group
failover_group
The following example creates intercluster LIFs "cluster01_icl01" and "cluster01_icl02" in the failover group
"intercluster01":
161
cluster01::> network interface show -service-policy default-intercluster
Logical Status Network Current
Current Is
Vserver Interface Admin/Oper Address/Mask Node Port
Home
----------- ---------- ---------- ------------------ -------------
------- ----
cluster01
cluster01_icl01
up/up 192.168.1.201/24 cluster01-01 e0e
true
cluster01_icl02
up/up 192.168.1.202/24 cluster01-02 e0f
true
The following example shows that the intercluster LIFs "cluster01_icl01" and "cluster01_icl02" on the SVM
"e0e" port will fail over to the "e0f" port.
162
Configuring intercluster LIFs on shared data ports
You can configure intercluster LIFs on ports shared with the data network. Doing so
reduces the number of ports you need for intercluster networking.
Steps
1. List the ports in the cluster:
163
cluster01::> network interface create -vserver cluster01 -lif
cluster01_icl01 -service-
policy default-intercluster -home-node cluster01-01 -home-port e0c
-address 192.168.1.201
-netmask 255.255.255.0
164
In ONTAP 9.5 and earlier:
The following example shows that the intercluster LIFs "cluster01_icl01" and "cluster01_icl02" on the "e0c"
port will fail over to the "e0d" port.
You must create a mirrored data aggregate on each node in the DR group.
About this task
• You should know what drives will be used in the new aggregate.
• If you have multiple drive types in your system (heterogeneous storage), you should understand how you
can ensure that the correct drive type is selected.
• Drives are owned by a specific node; when you create an aggregate, all drives in that aggregate must be
owned by the same node, which becomes the home node for that aggregate.
In systems using ADP, aggregates are created using partitions in which each drive is partitioned in to P1,
P2 and P3 partitions.
• Aggregate names should conform to the naming scheme you determined when you planned your
MetroCluster configuration.
Steps
1. Display a list of available spares:
165
2. Create the aggregate:
If you are logged in to the cluster on the cluster management interface, you can create an aggregate on
any node in the cluster. To ensure that the aggregate is created on a specific node, use the -node
parameter or specify drives that are owned by that node.
◦ Aggregate’s home node (that is, the node that owns the aggregate in normal operation)
◦ List of specific drives that are to be added to the aggregate
◦ Number of drives to include
For more information about these options, see the storage aggregate create man page.
You must add licenses for the new controller module for any ONTAP services that require
standard (node-locked) licenses. For features with standard licenses, each node in the
cluster must have its own key for the feature.
For detailed information about licensing, see the knowledgebase article 3013749: Data ONTAP 8.2 Licensing
166
Overview and References on the NetApp Support Site and the System Administration Reference.
Steps
1. If necessary, obtain license keys for the new node on the NetApp Support Site in the My Support section
under Software licenses.
If the site does not have the license keys you need, contact your sales or support representative.
You can optionally create unmirrored data aggregates for data that does not require the
redundant mirroring provided by MetroCluster configurations.
About this task
• You should know what drives or array LUNs will be used in the new aggregate.
• If you have multiple drive types in your system (heterogeneous storage), you should understand how you
can verify that the correct drive type is selected.
• Drives and array LUNs are owned by a specific node; when you create an aggregate, all drives in that
aggregate must be owned by the same node, which becomes the home node for that aggregate.
• Aggregate names should conform to the naming scheme you determined when you planned your
MetroCluster configuration.
• Disks and aggregates management contains more information about mirroring aggregates.
Steps
1. Enable unmirrored aggregate deployment:
3. Install and cable the disk shelves that will contain the unmirrored aggregates.
You can use the procedures in the Installation and Setup documentation for your platform and disk shelves.
167
4. Manually assign all disks on the new shelf to the appropriate node:
If you are logged in to the cluster on the cluster management interface, you can create an aggregate on
any node in the cluster. To verify that the aggregate is created on a specific node, you should use the
-node parameter or specify drives that are owned by that node.
You must also ensure that you are only including drives on the unmirrored shelf to the aggregate.
◦ Aggregate’s home node (that is, the node that owns the aggregate in normal operation)
◦ List of specific drives or array LUNs that are to be added to the aggregate
◦ Number of drives to include
◦ Checksum style to use for the aggregate
◦ Type of drives to use
◦ Size of drives to use
◦ Drive speed to use
◦ RAID type for RAID groups on the aggregate
◦ Maximum number of drives or array LUNs that can be included in a RAID group
◦ Whether drives with different RPM are allowed
For more information about these options, see the storage aggregate create man page.
168
Related information
Disk and aggregate management
After adding the controller module, you must install the latest firmware on the new
controller module so that the controller module functions properly with ONTAP.
Steps
1. Download the most current version of firmware for your system and follow the instructions for downloading
and installing the new firmware.
The following command refreshes the MetroCluster configuration on all of the nodes in the DR group
that contains controller_A_1:
The following example shows the network port usage on a four-node MetroCluster configuration:
169
cluster_A::> network port show
Speed (Mbps)
Node Port IPspace Broadcast Domain Link MTU Admin/Oper
------ --------- --------- ---------------- ----- ------- ------------
controller_A_1
e0a Cluster Cluster up 9000 auto/1000
e0b Cluster Cluster up 9000 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
e0e Default Default up 1500 auto/1000
e0f Default Default up 1500 auto/1000
e0g Default Default up 1500 auto/1000
controller_A_2
e0a Cluster Cluster up 9000 auto/1000
e0b Cluster Cluster up 9000 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
e0e Default Default up 1500 auto/1000
e0f Default Default up 1500 auto/1000
e0g Default Default up 1500 auto/1000
14 entries were displayed.
3. Verify the MetroCluster configuration from both sites in the MetroCluster configuration.
a. Verify the configuration from site A:
metrocluster show
metrocluster show
170
cluster_B::> metrocluster show
Cluster Entry Name State
------------------------- ------------------- -----------
Local: cluster_B Configuration state configured
Mode normal
AUSO Failure Domain auso-on-cluster-
disaster
Remote: cluster_A Configuration state configured
Mode normal
AUSO Failure Domain auso-on-cluster-
disaster
171
Steps
1. Enable storage failover:
Cluster high availability (HA) must be configured in a cluster if it contains only two nodes and it differs from
the HA provided by storage failover.
172
3. Repeat the previous step on the partner cluster.
4. Verify that the SVMs are in a healthy state:
Before performing this procedure, the MetroCluster FC configuration consists of four nodes, with one HA pair
at each site:
At the conclusion of this procedure, the MetroCluster FC configuration consists of two HA pairs at each site:
173
Both sites must be expanded equally. A MetroCluster FC configuration cannot consist of an uneven number of
nodes.
Steps
1. Use the procedure in Fabric-attached MetroCluster installation and configuration to create a cabling layout
for your switch type, using the port usage for an eight-node MetroCluster configuration.
The FC switch port usage must match the usage described in the procedure so that the Reference
Configuration Files (RCFs) can be used.
If your environment cannot be cabled in such a way that RCF files can be used, you must
manually configure the system according to instructions found in Fabric-attached
MetroCluster installation and configuration. Do not use this procedure if the cabling cannot
use RCF files.
174
Racking the new equipment
You must rack the equipment for the new nodes.
Steps
1. Use the procedure in Fabric-attached MetroCluster installation and configuration to rack the new storage
systems, disk shelves, and FC-to-SAS bridges.
metrocluster show
175
cluster_A::> metrocluster check run
Component Result
------------------- ---------
nodes ok
lifs ok
config-replication ok
aggregates ok
4 entries were displayed.
You need to respond with y when prompted to continue into advanced mode and see the advanced
mode prompt (*>).
Steps
176
1. Go to the Config Advisor download page and download the tool.
2. Run Config Advisor, review the tool’s output and follow the recommendations in the output to address any
issues discovered.
Steps
1. Log in to the cluster at Site_A.
2. Invoke an AutoSupport message indicating the start of the maintenance:
The maintenance-window-in-hours parameter specifies the length of the maintenance window and
can be a maximum of 72 hours. If the maintenance is completed before the time has elapsed, you can
issue the following command to indicating that the maintenance period has ended:
You must disconnect the existing controller modules from the FC switches in the fabric.
About this task
This task must be performed at each MetroCluster site.
Steps
1. Disable the HBA ports that connect the existing controller modules to the switch fabric undergoing
maintenance:
2. On the local FC switches, remove the cables from the ports for the existing controller module’s HBA, FC-
VI, and ATTO bridges.
You should label the cables for easy identification when you re-cable them. Only the ISL ports should
remain cabled.
177
Applying the RCF files and recabling the switches
You must apply the RCF files to reconfigure your zoning to accommodate the new nodes.
Steps
1. Locate the RCF files for your configuration.
You must use the RCF files for an eight-node configuration and that match your switch model.
2. Apply the RCF files, following the directions on the download page, adjusting the ISL settings as needed.
3. Ensure that the switch configuration is saved.
4. Reboot the FC switches.
5. Cable both the pre-existing and the new FC-to-SAS bridges to the FC switches, using the cabling layout
you created previously.
The FC switch port usage must match the MetroCluster eight-node usage described in Fabric-attached
MetroCluster installation and configuration so that the Reference Configuration Files (RCFs) can be used.
If your environment cannot be cabled in such a way that RCF files can be used then contact
technical support. Do NOT use this procedure if the cabling cannot use RCF files.
6. Verify that the ports are online by using the correct command for your switch.
7. Use the procedure in Fabric-attached MetroCluster installation and configuration to cable the FC-VI ports
from the existing and new controllers, using the cabling layout you created previously.
The FC switch port usage must match the MetroCluster eight-node usage described in Fabric-attached
MetroCluster installation and configuration so that the Reference Configuration Files (RCFs) can be used.
If your environment cannot be cabled in such a way that RCF files can be used then contact
technical support. Do NOT use this procedure if the cabling cannot use RCF files.
8. From the existing nodes, verify that the FC-VI ports are online:
9. Cable the HBA ports from the current and the new controllers.
10. On the existing controller modules, e-enable the ports connected to the switch fabric undergoing
maintenance:
178
11. Start the new controllers and boot them into Maintenance mode:
boot_ontap maint
12. Verify that only storage that will be used by the new DR group is visible to the new controller modules.
None of the storage that is used by the other DR group should be visible.
13. Return to the beginning of this process to re-cable the second switch fabric.
Before using a new controller module in the MetroCluster configuration, you must clear
the existing configuration.
Steps
1. If necessary, halt the node to display the LOADER prompt:
halt
set-defaults
saveenv
boot_ontap menu
wipeconfig
6. At the boot menu, select option 5 to boot the system into Maintenance mode.
If you are using AFF systems in a configuration with mirrored aggregates and the nodes
do not have the disks (SSDs) correctly assigned, you should assign half the disks on
each shelf to one local node and the other half of the disks to its HA partner node. You
should create a configuration in which each node has the same number of disks in its
179
local and remote disk pools.
About this task
The storage controllers must be in Maintenance mode.
This does not apply to configurations which have unmirrored aggregates, an active/passive configuration, or
that have an unequal number of disks in local and remote pools.
This task is not required if disks were correctly assigned when received from the factory.
Pool 0 always contains the disks that are found at the same site as the storage system that
owns them, while Pool 1 always contains the disks that are remote to the storage system that
owns them.
Steps
1. If you have not done so, boot each system into Maintenance mode.
2. Assign the disks to the nodes located at the first site (site A):
a. On the first node, systematically assign half the disks on each shelf to pool 0 and the other half to the
HA partner’s pool 0:
disk assign -disk disk-name -p pool -n number-of-disks
If storage controller Controller_A_1 has four shelves, each with 8 SSDs, you issue the following
commands:
b. Repeat the process for the second node at the local site, systematically assigning half the disks on
each shelf to pool 1 and the other half to the HA partner’s pool 1:
disk assign -disk disk-name -p pool
If storage controller Controller_A_1 has four shelves, each with 8 SSDs, you issue the following
commands:
3. Assign the disks to the nodes located at the second site (site B):
180
You should assign an equal number of disks to each pool.
a. On the first node at the remote site, systematically assign half the disks on each shelf to pool 0 and the
other half to the HA partner’s pool 0:
disk assign -disk disk-name -p pool
If storage controller Controller_B_1 has four shelves, each with 8 SSDs, you issue the following
commands:
b. Repeat the process for the second node at the remote site, systematically assigning half the disks on
each shelf to pool 1 and the other half to the HA partner’s pool 1:
If storage controller Controller_B_2 has four shelves, each with 8 SSDs, you issue the following
commands:
halt
boot_ontap menu
If the MetroCluster nodes do not have the disks correctly assigned, or if you are using
DS460C disk shelves in your configuration, you must assign disks to each of the nodes in
the MetroCluster configuration on a shelf-by-shelf basis. You will create a configuration in
which each node has the same number of disks in its local and remote disk pools.
181
About this task
The storage controllers must be in Maintenance mode.
If your configuration does not include DS460C disk shelves, this task is not required if disks were correctly
assigned when received from the factory.
Pool 0 always contains the disks that are found at the same site as the storage system that
owns them.
Pool 1 always contains the disks that are remote to the storage system that owns them.
If your configuration includes DS460C disk shelves, you should manually assign the disks using the following
guidelines for each 12-disk drawer:
This disk assignment pattern ensures that an aggregate is minimally affected in case a drawer goes offline.
Steps
1. If you have not done so, boot each system into Maintenance mode.
2. Assign the disk shelves to the nodes located at the first site (site A):
Disk shelves at the same site as the node are assigned to pool 0 and disk shelves located at the partner
site are assigned to pool 1.
a. On the first node, systematically assign the local disk shelves to pool 0 and the remote disk shelves to
pool 1:
If storage controller Controller_A_1 has four shelves, you issue the following commands:
b. Repeat the process for the second node at the local site, systematically assigning the local disk
182
shelves to pool 0 and the remote disk shelves to pool 1:
If storage controller Controller_A_2 has four shelves, you issue the following commands:
3. Assign the disk shelves to the nodes located at the second site (site B):
Disk shelves at the same site as the node are assigned to pool 0 and disk shelves located at the partner
site are assigned to pool 1.
a. On the first node at the remote site, systematically assign its local disk shelves to pool 0 and its remote
disk shelves to pool 1:
If storage controller Controller_B_1 has four shelves, you issue the following commands:
b. Repeat the process for the second node at the remote site, systematically assigning its local disk
shelves to pool 0 and its remote disk shelves to pool 1:
If storage controller Controller_B_2 has four shelves, you issue the following commands:
183
storage show shelf
halt
boot_ontap menu
In a MetroCluster configuration, the ha-config state of the controller module and chassis
components must be set to mcc so they boot up properly.
About this task
• The system must be in Maintenance mode.
• This task must be performed on each new controller module.
Steps
1. In Maintenance mode, display the HA state of the controller module and chassis:
ha-config show
2. If the displayed system state of the controller is not correct, set the HA state for the controller module:
3. If the displayed system state of the chassis is not correct, set the HA state for the chassis:
To join the new controllers to the cluster, you must boot each new controller module and
use the ONTAP cluster setup wizard to identify the cluster will join.
Before you begin
You must have cabled the MetroCluster configuration.
You must not have configured the Service Processor prior to performing this task.
Steps
184
1. If you have not already done so, power up each node and let them boot completely.
If the system is in Maintenance mode, issue the halt command to exit Maintenance mode, and then issue
the following command from the LOADER prompt:
boot_ontap
2. Enable the AutoSupport tool by following the directions provided by the system.
3. Respond to the prompts to configure the node management interface.
If not, you must issue the following command on each node, and then reboot the node:
This command configures high availability mode but does not enable storage failover. Storage failover is
automatically enabled when you issue the metrocluster configure command later in the
configuration process.
185
network port show
The following example shows output for two controllers in cluster_A. If it is a two-node MetroCluster
configuration, the output shows only one node.
6. Because you are using the CLI to set up the cluster, exit the Node Setup wizard:
exit
cluster setup
186
::> cluster setup
You can return to cluster setup at any time by typing "cluster setup".
To accept a default or omit a question, do not enter a value.
9. After you complete the Cluster Setup wizard and it exits, verify that the cluster is active and the node is
healthy:
cluster show
The following example shows a cluster in which the first node (cluster1-01) is healthy and eligible to
participate:
If it becomes necessary to change any of the settings you entered for the admin SVM or node SVM, you
can access the Cluster Setup wizard by using the cluster setup command.
You can configure intercluster LIFs on dedicated ports. Doing so typically increases the
available bandwidth for replication traffic.
Steps
1. List the ports in the cluster:
187
network port show
The following example shows that ports "e0e" and "e0f" have not been assigned LIFs:
188
cluster01::> network interface show -fields home-port,curr-port
vserver lif home-port curr-port
------- -------------------- --------- ---------
Cluster cluster01-01_clus1 e0a e0a
Cluster cluster01-01_clus2 e0b e0b
Cluster cluster01-02_clus1 e0a e0a
Cluster cluster01-02_clus2 e0b e0b
cluster01
cluster_mgmt e0c e0c
cluster01
cluster01-01_mgmt1 e0c e0c
cluster01
cluster01-02_mgmt1 e0c e0c
The following example assigns ports "e0e" and "e0f" to the failover group "intercluster01" on the system
SVM "cluster01":
189
cluster01::> network interface failover-groups show
Failover
Vserver Group Targets
---------------- ----------------
--------------------------------------------
Cluster
Cluster
cluster01-01:e0a, cluster01-01:e0b,
cluster01-02:e0a, cluster01-02:e0b
cluster01
Default
cluster01-01:e0c, cluster01-01:e0d,
cluster01-02:e0c, cluster01-02:e0d,
cluster01-01:e0e, cluster01-01:e0f
cluster01-02:e0e, cluster01-02:e0f
intercluster01
cluster01-01:e0e, cluster01-01:e0f
cluster01-02:e0e, cluster01-02:e0f
5. Create intercluster LIFs on the system SVM and assign them to the failover group.
9.5 and earlier network interface create -vserver system_SVM -lif LIF_name
-role intercluster -home-node node -home-port port
-address port_IP -netmask netmask -failover-group
failover_group
The following example creates intercluster LIFs "cluster01_icl01" and "cluster01_icl02" in the failover group
"intercluster01":
190
cluster01::> network interface create -vserver cluster01 -lif
cluster01_icl01 -service-
policy default-intercluster -home-node cluster01-01 -home-port e0e
-address 192.168.1.201
-netmask 255.255.255.0 -failover-group intercluster01
191
In ONTAP 9.5 and earlier:
The following example shows that the intercluster LIFs "cluster01_icl01" and "cluster01_icl02" on the SVM
"e0e" port will fail over to the "e0f" port.
You can configure intercluster LIFs on ports shared with the data network. Doing so
reduces the number of ports you need for intercluster networking.
Steps
1. List the ports in the cluster:
192
cluster01::> network port show
Speed
(Mbps)
Node Port IPspace Broadcast Domain Link MTU Admin/Oper
------ --------- ------------ ---------------- ----- -------
------------
cluster01-01
e0a Cluster Cluster up 1500 auto/1000
e0b Cluster Cluster up 1500 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
cluster01-02
e0a Cluster Cluster up 1500 auto/1000
e0b Cluster Cluster up 1500 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
193
3. Verify that the intercluster LIFs were created:
The following example shows that the intercluster LIFs "cluster01_icl01" and "cluster01_icl02" on the "e0c"
port will fail over to the "e0d" port.
194
cluster01::> network interface show -service-policy default-intercluster
–failover
Logical Home Failover Failover
Vserver Interface Node:Port Policy Group
-------- --------------- --------------------- --------------- --------
cluster01
cluster01_icl01 cluster01-01:e0c local-only
192.168.1.201/24
Failover Targets: cluster01-01:e0c,
cluster01-01:e0d
cluster01_icl02 cluster01-02:e0c local-only
192.168.1.201/24
Failover Targets: cluster01-02:e0c,
cluster01-02:e0d
On non-ADP systems, the RAID type of the aggregate can be modified from the default RAID-
DP to RAID4 before or after the aggregate is mirrored.
Steps
1. Mirror the root aggregate:
This mirrors the aggregate, so it consists of a local plex and a remote plex located at the remote
MetroCluster site.
2. Repeat the previous step for each node in the MetroCluster configuration.
You must run the metrocluster configure -refresh true command to start data
195
protection on the nodes that you have added to a MetroCluster configuration.
About this task
You issue the metrocluster configure -refresh true command once, on one of the newly added
nodes, to refresh the MetroCluster configuration. You do not need to issue the command on each of the sites
or nodes.
The metrocluster configure -refresh true command automatically pairs the two nodes with the
lowest system IDs in each of the two clusters as disaster recovery (DR) partners. In a four-node MetroCluster
configuration, there are two DR partner pairs. The second DR pair is created from the two nodes with higher
system IDs.
Steps
1. Refresh the MetroCluster configuration:
a. Enter advanced privilege mode:
The following example shows the MetroCluster configuration refreshed on both DR groups:
The following example shows the network port usage on a four-node MetroCluster configuration:
196
cluster_A::> network port show
Speed (Mbps)
Node Port IPspace Broadcast Domain Link MTU Admin/Oper
------ --------- --------- ---------------- ----- ------- ------------
controller_A_1
e0a Cluster Cluster up 9000 auto/1000
e0b Cluster Cluster up 9000 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
e0e Default Default up 1500 auto/1000
e0f Default Default up 1500 auto/1000
e0g Default Default up 1500 auto/1000
controller_A_2
e0a Cluster Cluster up 9000 auto/1000
e0b Cluster Cluster up 9000 auto/1000
e0c Default Default up 1500 auto/1000
e0d Default Default up 1500 auto/1000
e0e Default Default up 1500 auto/1000
e0f Default Default up 1500 auto/1000
e0g Default Default up 1500 auto/1000
14 entries were displayed.
3. Verify the MetroCluster configuration from both sites in the MetroCluster configuration:
a. Verify the configuration from site A:
metrocluster show
Configuration: IP fabric
197
cluster_B::> metrocluster show
Configuration: IP fabric
You must create a mirrored data aggregate on each node in the DR group.
About this task
• You should know what drives will be used in the new aggregate.
• If you have multiple drive types in your system (heterogeneous storage), you should understand how you
can ensure that the correct drive type is selected.
• Drives are owned by a specific node; when you create an aggregate, all drives in that aggregate must be
owned by the same node, which becomes the home node for that aggregate.
In systems using ADP, aggregates are created using partitions in which each drive is partitioned in to P1,
P2 and P3 partitions.
• Aggregate names should conform to the naming scheme you determined when you planned your
MetroCluster configuration.
Steps
1. Display a list of available spares:
If you are logged in to the cluster on the cluster management interface, you can create an aggregate on
any node in the cluster. To ensure that the aggregate is created on a specific node, use the -node
parameter or specify drives that are owned by that node.
◦ Aggregate’s home node (that is, the node that owns the aggregate in normal operation)
◦ List of specific drives that are to be added to the aggregate
◦ Number of drives to include
198
In the minimum supported configuration, in which a limited number of drives are
available, you must use the force-small-aggregate option to allow the creation of a three
disk RAID-DP aggregate.
For more information about these options, see the storage aggregate create man page.
Beginning with ONTAP 9.8, the storage bridge command is replaced with system bridge.
The following steps show the storage bridge command, but if you are running ONTAP 9.8 or
later, the system bridge command is preferred.
Step
1. From the ONTAP cluster prompt, add the bridge to health monitoring:
a. Add the bridge, using the command for your version of ONTAP:
199
9.4 and earlier storage bridge add -address bridge-
ip-address -name bridge-name
b. Verify that the bridge has been added and is properly configured:
It might take as long as 15 minutes to reflect all data because of the polling interval. The ONTAP health
monitor can contact and monitor the bridge if the value in the "Status" column is "ok", and other
information, such as the worldwide name (WWN), is displayed.
The following example shows that the FC-to-SAS bridges are configured:
controller_A_1::>
You can move a metadata volume from one aggregate to another aggregate in a
MetroCluster configuration. You might want to move a metadata volume when the source
aggregate is decommissioned or unmirrored, or for other reasons that make the
aggregate ineligible.
About this task
• You must have cluster administrator privileges to perform this task.
• The target aggregate must be mirrored and should not be in the degraded state.
• The available space in the target aggregate must be larger than the metadata volume that you are moving.
Steps
200
1. Set the privilege level to advanced:
Cluster_A::>
The following command identifies the aggregates in cluster_A that are eligible to host metadata volumes:
201
Cluster_A::*> metrocluster check config-replication show-aggregate-
eligibility
202
volume move show -volume vol_constituent_name
You can check that the components and relationships in the MetroCluster configuration
are working correctly. You should do a check after initial configuration and after making
any changes to the MetroCluster configuration. You should also do a check before a
negotiated (planned) switchover or a switchback operation.
About this task
If the metrocluster check run command is issued twice within a short time on either or both clusters, a
conflict can occur and the command might not collect all data. Subsequent metrocluster check show
commands do not show the expected output.
Steps
1. Check the configuration:
The command runs as a background job and might not be completed immediately.
Component Result
------------------- ---------
nodes ok
lifs ok
config-replication ok
aggregates ok
clusters ok
connections ok
6 entries were displayed.
203
2. Display more detailed results from the most recent metrocluster check run command:
The metrocluster check show commands show the results of the most recent metrocluster
check run command. You should always run the metrocluster check run command prior to using
the metrocluster check show commands so that the information displayed is current.
The following example shows the metrocluster check aggregate show command output for a
healthy four-node MetroCluster configuration:
204
controller_A_2 controller_A_2_aggr0
mirroring-status
ok
disk-pool-allocation
ok
ownership-state
ok
controller_A_2_aggr1
mirroring-status
ok
disk-pool-allocation
ok
ownership-state
ok
controller_A_2_aggr2
mirroring-status
ok
disk-pool-allocation
ok
ownership-state
ok
The following example shows the metrocluster check cluster show command output for a healthy
four-node MetroCluster configuration. It indicates that the clusters are ready to perform a negotiated
switchover if necessary.
205
Last Checked On: 9/13/2017 20:47:04
Steps
1. Go to the Config Advisor download page and download the tool.
2. Run Config Advisor, review the tool’s output and follow the recommendations in the output to address any
issues discovered.
Steps
206
1. Log in to the cluster at Site_A.
2. Invoke an AutoSupport message indicating the end of the maintenance:
• You must ensure that the old and new platform models are both supported by the IP switches.
• The new nodes must have enough storage to accommodate the data of the old nodes, along with
adequate disks for root aggregates and spare disks.
• This procedure is not supported on FAS2750 or AFF A220 systems.
207
Sending a custom AutoSupport message prior to maintenance
Before performing the maintenance, you should issue an AutoSupport message to notify NetApp technical
support that maintenance is underway. Informing technical support that maintenance is underway prevents
them from opening a case on the assumption that a disruption has occurred.
Steps
1. To prevent automatic support case generation, send an Autosupport message to indicate the upgrade is
underway.
a. Issue the following command:
This example specifies a 10 hour maintenance window. You might want to allow additional time,
depending on your plan.
If the maintenance is completed before the time has elapsed, you can invoke an AutoSupport message
indicating the end of the maintenance period:
Steps
1. Verify the operation of the MetroCluster configuration in ONTAP:
a. Check whether the system is multipathed:
c. Confirm the MetroCluster configuration and that the operational mode is normal:
metrocluster show
208
f. Run Config Advisor.
g. After running Config Advisor, review the tool’s output and follow the recommendations in the output to
address any issues discovered.
2. Verify that the cluster is healthy:
cluster_A::>
Node: node_A_1-old
Speed(Mbps) Health
Port IPspace Broadcast Domain Link MTU Admin/Oper Status
--------- ------------ ---------------- ---- ---- ----------- --------
e0a Cluster Cluster up 9000 auto/10000 healthy
e0b Cluster Cluster up 9000 auto/10000 healthy
Node: node_A_2-old
Speed(Mbps) Health
Port IPspace Broadcast Domain Link MTU Admin/Oper Status
--------- ------------ ---------------- ---- ---- ----------- --------
e0a Cluster Cluster up 9000 auto/10000 healthy
e0b Cluster Cluster up 9000 auto/10000 healthy
cluster_A::>
209
network interface show -vserver Cluster
Each cluster LIF should display true for Is Home and have a Status Admin/Oper of up/up
cluster_A::>
210
cluster_A::> network interface show -vserver Cluster -fields auto-revert
Logical
Vserver Interface Auto-revert
--------- ------------- ------------
Cluster
node_A_1-old_clus1
true
node_A_1-old_clus2
true
node_A_2-old_clus1
true
node_A_2-old_clus2
true
cluster_A::>
Steps
1. Remove the existing MetroCluster configuration from Tiebreaker, Mediator, or other software that can
initiate switchover.
metrocluster configuration-settings
mediator remove
2. Remove the existing MetroCluster configuration from any third-party application that can initiate switchover.
211
Preparing the new controller modules
You must prepare the four new MetroCluster nodes and install the correct ONTAP
version.
About this task
This task must be performed on each of the new nodes:
• node_A_3-new
• node_A_4-new
• node_B_3-new
• node_B_4-new
In these steps, you clear the configuration on the nodes and clear the mailbox region on new drives.
Steps
1. Rack the new controllers.
2. Cable the new MetroCluster IP nodes to the IP switches as shown in the MetroCluster installation and
configuration.
3. Configure the MetroCluster IP nodes using the following sections of the MetroCluster installation and
configuration.
a. Gathering required information
b. Restoring system defaults on a controller module
c. Verifying the ha-config state of components
d. Manually assigning drives for pool 0 (ONTAP 9.4 and later)
4. From Maintenance mode, issue the halt command to exit Maintenance mode, and then issue the
boot_ontap command to boot the system and get to cluster setup.
Steps
1. Add the new MetroCluster IP nodes to the existing MetroCluster configuration.
a. Join the first new MetroCluster IP node (node_A_1-new) to the existing MetroCluster IP configuration.
212
"help" or "?" - if you want to have a question clarified,
"back" - if you want to change previously answered questions, and
"exit" or "quit" - if you want to quit the cluster setup wizard.
Any changes you made before quitting will be saved.
This system will send event messages and periodic reports to NetApp
Technical
Support. To disable this feature, enter
autosupport modify -support disable
within 24 hours.
213
Do you want to create a new cluster or join an existing cluster?
{create, join}:
join
b. Join the second new MetroCluster IP node (node_A_2-new) to the existing MetroCluster IP
configuration.
2. Repeat these steps to join node_B_1-new and node_B_2-new to cluster_B.
Steps
1. On the new MetroCluster IP nodes, configure the intercluster LIFs using the following procedures:
214
cluster_A:> cluster peer show
Peer Cluster Name Cluster Serial Number Availability
Authentication
------------------------- --------------------- --------------
--------------
cluster_B 1-80-000011 Available ok
For more information on the MetroCluster configuration settings and connections, see the following:
215
cluster_A::> metrocluster configuration-settings dr-group show
cluster_A::>
5. Configure the MetroCluster IP interfaces for the newly joined MetroCluster IP nodes:
◦ Certain platforms use a VLAN for the MetroCluster IP interface. By default, each of the
two ports use a different VLAN: 10 and 20. You can also specify a different (non-default)
VLAN higher than 100 (between 101 and 4095) using the -vlan-id parameter in the
metrocluster configuration-settings interface create command.
◦ Beginning with ONTAP 9.9.1, if you are using a layer 3 configuration, you must also
specify the -gateway parameter when creating MetroCluster IP interfaces. Refer to
Considerations for layer 3 wide-area networks.
The following platform models can be added to the existing MetroCluster configuration if the VLANs used
are 10/20 or greater than 100. If any other VLANs are used, then these platforms cannot be added to the
existing configuration as the MetroCluster interface cannot be configured. If you are using any other
platform, the VLAN configuration is not relevant as this is not required in ONTAP.
216
You can configure the MetroCluster IP interfaces from either cluster. Also, beginning with
ONTAP 9.1.1, if you are using a layer 3 configuration, you must also specify the -gateway
parameter to create MetroCluster IP interfaces. Refer to Considerations for layer 3 wide-
area networks.
217
6. Verify the MetroCluster IP interfaces are created:
DR
Config
Group Cluster Node Network Address Netmask Gateway
State
----- ------- ------- --------------- --------------- ---------------
---------
1 cluster_A
node_A_1-old
Home Port: e1a
172.17.26.10 255.255.255.0 -
completed
Home Port: e1b
172.17.27.10 255.255.255.0 -
completed
node_A_2-old
Home Port: e1a
172.17.26.11 255.255.255.0 -
completed
Home Port: e1b
172.17.27.11 255.255.255.0 -
completed
cluster_B
node_B_1-old
Home Port: e1a
172.17.26.13 255.255.255.0 -
completed
Home Port: e1b
172.17.27.13 255.255.255.0 -
completed
node_B_1-old
Home Port: e1a
172.17.26.12 255.255.255.0 -
completed
Home Port: e1b
172.17.27.12 255.255.255.0 -
completed
2 cluster_A
node_A_3-new
Home Port: e1a
172.17.28.10 255.255.255.0 -
218
completed
Home Port: e1b
172.17.29.10 255.255.255.0 -
completed
node_A_3-new
Home Port: e1a
172.17.28.11 255.255.255.0 -
completed
Home Port: e1b
172.17.29.11 255.255.255.0 -
completed
cluster_B
node_B_3-new
Home Port: e1a
172.17.28.13 255.255.255.0 -
completed
Home Port: e1b
172.17.29.13 255.255.255.0 -
completed
node_B_3-new
Home Port: e1a
172.17.28.12 255.255.255.0 -
completed
Home Port: e1b
172.17.29.12 255.255.255.0 -
completed
8 entries were displayed.
cluster_A>
cluster_A::>
219
DR Source Destination
Group Cluster Node Network Address Network Address Partner Type
Config State
----- ------- ------- --------------- --------------- ------------
------------
1 cluster_A
node_A_1-old
Home Port: e1a
172.17.28.10 172.17.28.11 HA Partner
completed
Home Port: e1a
172.17.28.10 172.17.28.12 DR Partner
completed
Home Port: e1a
172.17.28.10 172.17.28.13 DR Auxiliary
completed
Home Port: e1b
172.17.29.10 172.17.29.11 HA Partner
completed
Home Port: e1b
172.17.29.10 172.17.29.12 DR Partner
completed
Home Port: e1b
172.17.29.10 172.17.29.13 DR Auxiliary
completed
node_A_2-old
Home Port: e1a
172.17.28.11 172.17.28.10 HA Partner
completed
Home Port: e1a
172.17.28.11 172.17.28.13 DR Partner
completed
Home Port: e1a
172.17.28.11 172.17.28.12 DR Auxiliary
completed
Home Port: e1b
172.17.29.11 172.17.29.10 HA Partner
completed
Home Port: e1b
172.17.29.11 172.17.29.13 DR Partner
completed
Home Port: e1b
172.17.29.11 172.17.29.12 DR Auxiliary
completed
DR Source Destination
220
Group Cluster Node Network Address Network Address Partner Type
Config State
----- ------- ------- --------------- --------------- ------------
------------
1 cluster_B
node_B_2-old
Home Port: e1a
172.17.28.13 172.17.28.12 HA Partner
completed
Home Port: e1a
172.17.28.13 172.17.28.11 DR Partner
completed
Home Port: e1a
172.17.28.13 172.17.28.10 DR Auxiliary
completed
Home Port: e1b
172.17.29.13 172.17.29.12 HA Partner
completed
Home Port: e1b
172.17.29.13 172.17.29.11 DR Partner
completed
Home Port: e1b
172.17.29.13 172.17.29.10 DR Auxiliary
completed
node_B_1-old
Home Port: e1a
172.17.28.12 172.17.28.13 HA Partner
completed
Home Port: e1a
172.17.28.12 172.17.28.10 DR Partner
completed
Home Port: e1a
172.17.28.12 172.17.28.11 DR Auxiliary
completed
Home Port: e1b
172.17.29.12 172.17.29.13 HA Partner
completed
Home Port: e1b
172.17.29.12 172.17.29.10 DR Partner
completed
Home Port: e1b
172.17.29.12 172.17.29.11 DR Auxiliary
completed
DR Source Destination
Group Cluster Node Network Address Network Address Partner Type
221
Config State
----- ------- ------- --------------- --------------- ------------
------------
2 cluster_A
node_A_1-new**
Home Port: e1a
172.17.26.10 172.17.26.11 HA Partner
completed
Home Port: e1a
172.17.26.10 172.17.26.12 DR Partner
completed
Home Port: e1a
172.17.26.10 172.17.26.13 DR Auxiliary
completed
Home Port: e1b
172.17.27.10 172.17.27.11 HA Partner
completed
Home Port: e1b
172.17.27.10 172.17.27.12 DR Partner
completed
Home Port: e1b
172.17.27.10 172.17.27.13 DR Auxiliary
completed
node_A_2-new
Home Port: e1a
172.17.26.11 172.17.26.10 HA Partner
completed
Home Port: e1a
172.17.26.11 172.17.26.13 DR Partner
completed
Home Port: e1a
172.17.26.11 172.17.26.12 DR Auxiliary
completed
Home Port: e1b
172.17.27.11 172.17.27.10 HA Partner
completed
Home Port: e1b
172.17.27.11 172.17.27.13 DR Partner
completed
Home Port: e1b
172.17.27.11 172.17.27.12 DR Auxiliary
completed
DR Source Destination
Group Cluster Node Network Address Network Address Partner Type
Config State
222
----- ------- ------- --------------- --------------- ------------
------------
2 cluster_B
node_B_2-new
Home Port: e1a
172.17.26.13 172.17.26.12 HA Partner
completed
Home Port: e1a
172.17.26.13 172.17.26.11 DR Partner
completed
Home Port: e1a
172.17.26.13 172.17.26.10 DR Auxiliary
completed
Home Port: e1b
172.17.27.13 172.17.27.12 HA Partner
completed
Home Port: e1b
172.17.27.13 172.17.27.11 DR Partner
completed
Home Port: e1b
172.17.27.13 172.17.27.10 DR Auxiliary
completed
node_B_1-new
Home Port: e1a
172.17.26.12 172.17.26.13 HA Partner
completed
Home Port: e1a
172.17.26.12 172.17.26.10 DR Partner
completed
Home Port: e1a
172.17.26.12 172.17.26.11 DR Auxiliary
completed
Home Port: e1b
172.17.27.12 172.17.27.13 HA Partner
completed
Home Port: e1b
172.17.27.12 172.17.27.10 DR Partner
completed
Home Port: e1b
172.17.27.12 172.17.27.11 DR Auxiliary
completed
48 entries were displayed.
cluster_A::>
223
disk show -pool Pool1
cluster_A::>
224
cluster_A::> aggr mirror -aggregate aggr0_node_A_1-new
Second Plex
cluster_A::>
mirrored,
normal
225
aggr0_node_A_2-old
349.0GB 16.84GB 95% online 1 node_A_2-old
raid_dp,
mirrored,
normal
aggr0_node_A_1-new
467.6GB 22.63GB 95% online 1 node_A_1-new
raid_dp,
mirrored,
normal
aggr0_node_A_2-new
467.6GB 22.62GB 95% online 1 node_A_2-new
raid_dp,
mirrored,
normal
aggr_data_a1
1.02TB 1.01TB 1% online 1 node_A_1-old
raid_dp,
mirrored,
normal
aggr_data_a2
1.02TB 1.01TB 1% online 1 node_A_2-old
raid_dp,
mirrored,
Steps
1. Create mirrored data aggregates on each of the new MetroCluster nodes:
226
You must create at least one mirrored data aggregate per site. It is recommended to have
two mirrored data aggregates per site on MetroCluster IP nodes to host the MDV volumes,
however a single aggregate per site is supported (but not recommended). It is support that
one site of the MetroCluster has a single mirrored data aggregate and the other site has
more than one mirrored data aggregate.
First Plex
Second Plex
227
-
data 4.20.18 SAS 546.9GB
547.1GB
data 4.20.19 SAS 546.9GB
547.1GB
data 4.20.16 SAS 546.9GB
547.1GB
cluster_A::>
metrocluster configure
The following example shows the MetroCluster configuration refreshed on both DR groups:
228
cluster_A::*> metrocluster node show
DR Configuration DR
Group Cluster Node State Mirroring Mode
----- ------- ------------------ -------------- ---------
--------------------
1 cluster_A
node_A_1-old configured enabled normal
node_A_2-old configured enabled normal
cluster_B
node_B_1-old configured enabled normal
node_B_2-old configured enabled normal
2 cluster_A
node_A_3-new configured enabled normal
node_A_4-new configured enabled normal
cluster_B
node_B_3-new configured enabled normal
node_B_4-new configured enabled normal
8 entries were displayed.
cluster_A::*>
4. Move the MDV_CRS volumes from the old nodes to the new nodes in advanced privilege.
a. Display the volumes to identify the MDV volumes:
If you have a single mirrored data aggregate per site then move both the MDV volumes
to this single aggregate. If you have two or more mirrored data aggregates, then move
each MDV volume to a different aggregate.
The following example shows the MDV volumes in the volume show output:
229
cluster_A::> volume show
Vserver Volume Aggregate State Type Size
Available Used%
--------- ------------ ------------ ---------- ---- ----------
---------- -----
...
cluster_A MDV_CRS_2c78e009ff5611e9b0f300a0985ef8c4_A
aggr_b1 - RW -
- -
cluster_A MDV_CRS_2c78e009ff5611e9b0f300a0985ef8c4_B
aggr_b2 - RW -
- -
cluster_A MDV_CRS_d6b0b313ff5611e9837100a098544e51_A
aggr_a1 online RW 10GB
9.50GB 0%
cluster_A MDV_CRS_d6b0b313ff5611e9837100a098544e51_B
aggr_a2 online RW 10GB
9.50GB 0%
...
11 entries were displayed.mple
The following example shows the command and output for moving
"MDV_CRS_d6b0b313ff5611e9837100a098544e51_A" to aggregate "data_a3" on "node_A_3".
230
cluster_A::> vol move start -volume
MDV_CRS_d6b0b313ff5611e9837100a098544e51_A -destination-aggregate
data_a3 -vserver cluster_A
d. Use the volume show command to check that the MDV volume has been successfully moved:
The following output shows that the MDV volume has been successfully moved.
231
cluster_B::> cluster show -fields epsilon
node epsilon
---------------- -------
node_A_1-old true
node_A_2-old false
node_A_3-new false
node_A_4-new false
4 entries were displayed.
NetApp Support
232
By removing a DR Group, four nodes remain in the configuration.
233
b. Move all MDV_CRS metadata volumes to another DR group. Follow the steps in the following
procedure: Moving a metadata volume in MetroCluster configurations
c. Delete all MDV_aud metadata volumes that might exist in the DR group to be removed.
d. Delete all data aggregates in the DR group to be removed as shown in the following example:
234
metrocluster node show
The following example shows the removal of the DR group configuration on cluster_A.
235
cluster_A::*>
cluster_A::*>
236
These commands can be issued from either cluster and apply to the entire DR group spanning both the
clusters.
Repeat this step for the other local node in the old DR group.
7. Halt, power down, and remove the old controller modules and storage shelves.
Information Subject
MetroCluster Documentation • All MetroCluster information
237
Fabric-attached MetroCluster installation and • Fabric-attached MetroCluster architecture
configuration
• Cabling the configuration
• Configuring the FC-to-SAS bridges
• Configuring the FC switches
• Configuring the MetroCluster in ONTAP
MetroCluster Tiebreaker Software installation and • Monitoring the MetroCluster configuration with the
configuration MetroCluster Tiebreaker software
238
AFF and FAS Documentation Center • Hot-adding a disk shelf
• Hot-removing a disk shelf
The standard storage shelf
maintenance procedures can be used
with MetroCluster IP configurations.
239
Copyright Information
Copyright © 2022 NetApp, Inc. All rights reserved. Printed in the U.S. No part of this document covered by
copyright may be reproduced in any form or by any means-graphic, electronic, or mechanical, including
photocopying, recording, taping, or storage in an electronic retrieval system- without prior written permission of
the copyright owner.
Software derived from copyrighted NetApp material is subject to the following license and disclaimer:
THIS SOFTWARE IS PROVIDED BY NETAPP “AS IS” AND WITHOUT ANY EXPRESS OR IMPLIED
WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY
AND FITNESS FOR A PARTICULAR PURPOSE, WHICH ARE HEREBY DISCLAIMED. IN NO EVENT SHALL
NETAPP BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
NetApp reserves the right to change any products described herein at any time, and without notice. NetApp
assumes no responsibility or liability arising from the use of products described herein, except as expressly
agreed to in writing by NetApp. The use or purchase of this product does not convey a license under any
patent rights, trademark rights, or any other intellectual property rights of NetApp.
The product described in this manual may be protected by one or more U.S. patents, foreign patents, or
pending applications.
RESTRICTED RIGHTS LEGEND: Use, duplication, or disclosure by the government is subject to restrictions
as set forth in subparagraph (c)(1)(ii) of the Rights in Technical Data and Computer Software clause at DFARS
252.277-7103 (October 1988) and FAR 52-227-19 (June 1987).
Trademark Information
NETAPP, the NETAPP logo, and the marks listed at http://www.netapp.com/TM are trademarks of NetApp, Inc.
Other company and product names may be trademarks of their respective owners.
240