Updating the partner node in an HA pair
After updating the first node in an HA pair, you update its partner by initiating a takeover on it. The first node serves the partner's data while the partner node is upgraded or downgraded.
- Set the privilege level to advanced, entering y when prompted to continue: set -privilege advanced The advanced prompt (*>) appears.
- Set the new ONTAP software image to be the default image: system image modify {-node nodenameB -iscurrent false} -isdefault true The system image modify command uses an extended query to change the new ONTAP software image (which is installed as the alternate image) to be the default image for the node.
- Monitor the progress of the update: cluster image show-update-progress
- Verify that the new ONTAP software image is set as the default image: system image show
Example
In the following example, image2 is the new version of ONTAP and is set as the default image on the node:TDC-DM7K::*> system image show
Is Is Install
Node Image Default Current Version Date
-------- ------- ------- ------- --------- -------------------
TDC-7K-03
image1 false false X.X.X MM/DD/YYYY TIME
image2 true true Y.Y.Y MM/DD/YYYY TIME
TDC-7K-01
image1 false true X.X.X MM/DD/YYYY TIME
image2 true false Y.Y.Y MM/DD/YYYY TIME
4 entries were displayed. - Disable automatic giveback on the partner node if it is enabled: storage failover modify -node nodenameA -auto-giveback false If the cluster is an HA pair, a message is displayed warning you that disabling automatic giveback prevents the management cluster services from going online in the event of an alternating-failure scenario. Enter y to continue.
- Verify that automatic giveback is disabled for the partner node: storage failover show -node nodenameA -fields auto-giveback
Example
TDC-DM7K::> storage failover show -node node0 -fields auto-giveback
node auto-giveback
-------- -------------
TDC-7K-03 false
1 entry was displayed. - Run the following command twice to determine whether the node to be updated is currently serving any clients: system node run -node nodenameB -command uptime The uptime command displays the total number of operations that the node has performed for NFS, CIFS, FC, and iSCSI clients since the node was last booted. For each protocol, you must run the command twice to determine whether the operation counts are increasing. If they are increasing, the node is currently serving clients for that protocol. If they are not increasing, the node is not currently serving clients for that protocol.NoteYou should make a note of each protocol that has increasing client operations so that after the node is updated, you can verify that client traffic has resumed.
Example
The following example shows a node with NFS, CIFS, FC, and iSCSI operations. However, the node is currently serving only NFS and iSCSI clients.TDC-DM7K::> system node run -node TDC-7K-01 -command uptime
2:58pm up 7 days, 19:16 800000260 NFS ops, 1017333 CIFS ops, 0 HTTP ops, 40395 FCP ops, 32810 iSCSI ops
TDC-DM7K::> system node run -node TDC-7K-01 -command uptime
2:58pm up 7 days, 19:17 800001573 NFS ops, 1017333 CIFS ops, 0 HTTP ops, 40395 FCP ops, 32815 iSCSI ops - Migrate all of the data LIFs away from the node: network interface migrate-all -node nodenameB
- Verify the status of any LIFs that you migrated: network interface show For more information about parameters you can use to verify LIF status, see the network interface show man page.
Example
The following example shows that node1's data LIFs migrated successfully. For each LIF, the fields included in this example enable you to verify the LIF's home node and port, the current node and port to which the LIF migrated, and the LIF's operational and administrative status.TDC-DM7K::> network interface show -data-protocol nfs|cifs -role data -home-node
TDC-7K-01 -fields home-node,curr-node,curr-port,home-port,status-admin,status-oper
vserver lif home-node home-port curr-node curr-port status-oper status-admin
------- ------- --------- --------- --------- --------- ----------- ------------
vs0 data001 TDC-7K-01 e0a TDC-7K-03 e0a up up
vs0 data002 TDC-7K-01 e0b TDC-7K-03 e0b up up
vs0 data003 TDC-7K-01 e0b TDC-7K-03 e0b up up
vs0 data004 TDC-7K-01 e0a TDC-7K-03 e0a up up
4 entries were displayed. - Initiate a takeover: storage failover takeover -ofnode nodenameB -option allow-version-mismatch Do not specify the -option immediate parameter, because a normal takeover is required for the node that is being taken over to boot onto the new software image. If you did not manually migrate the LIFs away from the node, they automatically migrate to the node's HA partner so that there are no service disruptions.The node that is taken over boots up to the Waiting for giveback state.NoteIf AutoSupport is enabled, an AutoSupport message is sent indicating that the node is out of cluster quorum. You can ignore this notification and proceed with the update.
- Verify that the takeover was successful: storage failover show
Example
The following example shows that the takeover was successful. Node node1 is in the Waiting for giveback state, and its partner is in the In takeover state.TDC-DM7K::> storage failover show
Takeover
Node Partner Possible State Description
-------------- -------------- -------- -------------------------------------
TDC-7K-03 TDC-7K-01 - In takeover
TDC-7K-01 TDC-7K-03 false Waiting for giveback (HA mailboxes)
2 entries were displayed. - Wait at least eight minutes for the following conditions to take effect:
Client multipathing (if deployed) is stabilized.
Clients are recovered from the pause in I/O that occurs during takeover.
The recovery time is client-specific and might take longer than eight minutes, depending on the characteristics of the client applications.
- Return the aggregates to the partner node: storage failover giveback -ofnode nodenameB The giveback operation first returns the root aggregate to the partner node and then, after that node has finished booting, returns the non-root aggregates and any LIFs that were set to automatically revert. The newly booted node begins to serve data to clients from each aggregate as soon as the aggregate is returned.
- Verify that all aggregates are returned: storage failover show-giveback If the Giveback Status field indicates that there are no aggregates to give back, then all aggregates are returned. If the giveback is vetoed, the command displays the giveback progress and which subsystem vetoed the giveback operation.
- If any aggregates are not returned, perform the following steps:
- Review the veto workaround to determine whether you want to address the
veto
condition or override the veto. - If necessary, address the
veto
condition described in the error message, ensuring that any identified operations are terminated gracefully. - Rerun the storage failover giveback command.If you decided to override the
veto
condition, set the -override-vetoes parameter to true .
- Review the veto workaround to determine whether you want to address the
- Wait at least eight minutes for the following conditions to take effect:
Client multipathing (if deployed) is stabilized.
Clients are recovered from the pause in an I/O operation that occurs during giveback.
The recovery time is client specific and might take longer than eight minutes, depending on the characteristics of the client applications.
- Verify that the update was completed successfully for the node:
- Go to the advanced privilege level : set -privilege advanced
- Verify that update status is complete for the node: system node upgrade-revert show -node nodenameBThe status should be listed as complete.
If the status is not complete, from the node, run the system node upgrade-revert upgrade command. If the command does not complete the update, contact technical support.
- Return to the admin privilege level: set -privilege admin
- Verify that the node's ports are up: network port show -node nodenameB You must run this command on a node that has been upgraded to ONTAP 9.5.
Example
The following example shows that all of the node's data ports are up:TDC-DM7K::> network port show -node TDC-7K-01
Speed (Mbps)
Node Port IPspace Broadcast Domain Link MTU Admin/Oper
------ --------- ------------ ---------------- ----- ------- ------------
TDC-7K-01
e0M Default - up 1500 auto/100
e0a Default - up 1500 auto/1000
e0b Default - up 1500 auto/1000
e1a Cluster Cluster up 9000 auto/10000
e1b Cluster Cluster up 9000 auto/10000
5 entries were displayed. - Revert the LIFs back to the node: network interface revert * This command returns the LIFs that were migrated away from the node.
Example
TDC-DM7K::> network interface revert *
8 entries were acted on. - Verify that the node's data LIFs successfully reverted back to the node, and that they are up: network interface show
Example
The following example shows that all of the data LIFs hosted by the node is successfully reverted back to the node, and that their operational status is up:TDC-DM7K::> network interface show
Logical Status Network Current Current Is
Vserver Interface Admin/Oper Address/Mask Node Port Home
----------- ---------- ---------- ------------------ ------------- ------- ----
vs0
data001 up/up 192.0.2.120/24 TDC-7K-01 e0a true
data002 up/up 192.0.2.121/24 TDC-7K-01 e0b true
data003 up/up 192.0.2.122/24 TDC-7K-01 e0b true
data004 up/up 192.0.2.123/24 TDC-7K-01 e0a true
4 entries were displayed. - If you previously determined that this node serves clients, verify that the node is providing service for each protocol that it was previously serving: system node run -node nodenameB -command uptime The operation counts reset to zero during the update.
Example
The following example shows that the updated node has resumed serving its NFS and iSCSI clients:TDC-DM7K::> system node run -node TDC-7K-01 -command uptime
3:15pm up 0 days, 0:16 129 NFS ops, 0 CIFS ops, 0 HTTP ops, 0 FCP ops, 2 iSCSI ops - If this was the last node in the cluster to be updated, trigger an AutoSupport notification: autosupport invoke -node * -type all -message "Finishing_NDU" This AutoSupport notification includes a record of the system status just prior to update. It saves useful troubleshooting information in case there is a problem with the update process.
If the cluster is not configured to send AutoSupport messages, a copy of the notification is saved locally.
- Confirm that the new ONTAP software is running on both nodes of the HA pair: system node image show
Example
In the following example, image2 is the updated version of ONTAP and is the default version on both nodes:TDC-DM7K::*> system node image show
Is Is Install
Node Image Default Current Version Date
-------- ------- ------- ------- --------- -------------------
TDC-7K-03
image1 false false X.X.X MM/DD/YYYY TIME
image2 true true Y.Y.Y MM/DD/YYYY TIME
TDC-7K-01
image1 false false X.X.X MM/DD/YYYY TIME
image2 true true Y.Y.Y MM/DD/YYYY TIME
4 entries were displayed. - Reenable automatic giveback on the partner node if it was previously disabled: storage failover modify -node nodenameA -auto-giveback true
- Verify that the cluster is in quorum and that services are running by using the cluster show and cluster ring show (advanced privilege level) commands.You must perform this step before upgrading any additional HA pairs.
- Return to the admin privilege level: set -privilege admin
After you finish
Upgrade any additional HA pairs.