Skip to main content

Updating the partner node in an HA pair

After updating the first node in an HA pair, you update its partner by initiating a takeover on it. The first node serves the partner's data while the partner node is upgraded or downgraded.

  1. Set the privilege level to advanced, entering y when prompted to continue: set -privilege advanced

    The advanced prompt (*>) appears.
  2. Set the new ONTAP software image to be the default image: system image modify {-node nodenameB -iscurrent false} -isdefault true

    The system image modify command uses an extended query to change the new ONTAP software image (which is installed as the alternate image) to be the default image for the node.
  3. Monitor the progress of the update: cluster image show-update-progress
  4. Verify that the new ONTAP software image is set as the default image: system image show

    Example

    In the following example, image2 is the new version of ONTAP and is set as the default image on the node:
    TDC-DM7K::*> system image show
    Is Is Install
    Node Image Default Current Version Date
    -------- ------- ------- ------- --------- -------------------
    TDC-7K-03
    image1 false false X.X.X MM/DD/YYYY TIME
    image2 true true Y.Y.Y MM/DD/YYYY TIME
    TDC-7K-01
    image1 false true X.X.X MM/DD/YYYY TIME
    image2 true false Y.Y.Y MM/DD/YYYY TIME
    4 entries were displayed.

  5. Disable automatic giveback on the partner node if it is enabled: storage failover modify -node nodenameA -auto-giveback false

    If the cluster is an HA pair, a message is displayed warning you that disabling automatic giveback prevents the management cluster services from going online in the event of an alternating-failure scenario. Enter y to continue.
  6. Verify that automatic giveback is disabled for the partner node: storage failover show -node nodenameA -fields auto-giveback

    Example

    TDC-DM7K::> storage failover show -node node0 -fields auto-giveback
    node auto-giveback
    -------- -------------
    TDC-7K-03 false
    1 entry was displayed.

  7. Run the following command twice to determine whether the node to be updated is currently serving any clients: system node run -node nodenameB -command uptime

    The uptime command displays the total number of operations that the node has performed for NFS, CIFS, FC, and iSCSI clients since the node was last booted. For each protocol, you must run the command twice to determine whether the operation counts are increasing. If they are increasing, the node is currently serving clients for that protocol. If they are not increasing, the node is not currently serving clients for that protocol.
    Note
    You should make a note of each protocol that has increasing client operations so that after the node is updated, you can verify that client traffic has resumed.

    Example

    The following example shows a node with NFS, CIFS, FC, and iSCSI operations. However, the node is currently serving only NFS and iSCSI clients.
    TDC-DM7K::> system node run -node TDC-7K-01 -command uptime
    2:58pm up 7 days, 19:16 800000260 NFS ops, 1017333 CIFS ops, 0 HTTP ops, 40395 FCP ops, 32810 iSCSI ops

    TDC-DM7K::> system node run -node TDC-7K-01 -command uptime
    2:58pm up 7 days, 19:17 800001573 NFS ops, 1017333 CIFS ops, 0 HTTP ops, 40395 FCP ops, 32815 iSCSI ops

  8. Migrate all of the data LIFs away from the node: network interface migrate-all -node nodenameB
  9. Verify the status of any LIFs that you migrated: network interface show

    For more information about parameters you can use to verify LIF status, see the network interface show man page.

    Example

    The following example shows that node1's data LIFs migrated successfully. For each LIF, the fields included in this example enable you to verify the LIF's home node and port, the current node and port to which the LIF migrated, and the LIF's operational and administrative status.
    TDC-DM7K::> network interface show -data-protocol nfs|cifs -role data -home-node 
    TDC-7K-01 -fields home-node,curr-node,curr-port,home-port,status-admin,status-oper
    vserver lif home-node home-port curr-node curr-port status-oper status-admin
    ------- ------- --------- --------- --------- --------- ----------- ------------
    vs0 data001 TDC-7K-01 e0a TDC-7K-03 e0a up up
    vs0 data002 TDC-7K-01 e0b TDC-7K-03 e0b up up
    vs0 data003 TDC-7K-01 e0b TDC-7K-03 e0b up up
    vs0 data004 TDC-7K-01 e0a TDC-7K-03 e0a up up
    4 entries were displayed.

  10. Initiate a takeover: storage failover takeover -ofnode nodenameB -option allow-version-mismatch

    Do not specify the -option immediate parameter, because a normal takeover is required for the node that is being taken over to boot onto the new software image. If you did not manually migrate the LIFs away from the node, they automatically migrate to the node's HA partner so that there are no service disruptions.
    The node that is taken over boots up to the Waiting for giveback state.
    Note
    If AutoSupport is enabled, an AutoSupport message is sent indicating that the node is out of cluster quorum. You can ignore this notification and proceed with the update.
  11. Verify that the takeover was successful: storage failover show

    Example

    The following example shows that the takeover was successful. Node node1 is in the Waiting for giveback state, and its partner is in the In takeover state.
    TDC-DM7K::> storage failover show
    Takeover
    Node Partner Possible State Description
    -------------- -------------- -------- -------------------------------------
    TDC-7K-03 TDC-7K-01 - In takeover
    TDC-7K-01 TDC-7K-03 false Waiting for giveback (HA mailboxes)
    2 entries were displayed.

  12. Wait at least eight minutes for the following conditions to take effect:

    • Client multipathing (if deployed) is stabilized.

    • Clients are recovered from the pause in I/O that occurs during takeover.

      The recovery time is client-specific and might take longer than eight minutes, depending on the characteristics of the client applications.

  13. Return the aggregates to the partner node: storage failover giveback -ofnode nodenameB
    The giveback operation first returns the root aggregate to the partner node and then, after that node has finished booting, returns the non-root aggregates and any LIFs that were set to automatically revert. The newly booted node begins to serve data to clients from each aggregate as soon as the aggregate is returned.
  14. Verify that all aggregates are returned: storage failover show-giveback

    If the Giveback Status field indicates that there are no aggregates to give back, then all aggregates are returned. If the giveback is vetoed, the command displays the giveback progress and which subsystem vetoed the giveback operation.
  15. If any aggregates are not returned, perform the following steps:
    1. Review the veto workaround to determine whether you want to address the veto condition or override the veto.
    2. If necessary, address the veto condition described in the error message, ensuring that any identified operations are terminated gracefully.
    3. Rerun the storage failover giveback command.

      If you decided to override the veto condition, set the -override-vetoes parameter to true .
  16. Wait at least eight minutes for the following conditions to take effect:
    • Client multipathing (if deployed) is stabilized.

    • Clients are recovered from the pause in an I/O operation that occurs during giveback.

      The recovery time is client specific and might take longer than eight minutes, depending on the characteristics of the client applications.

  17. Verify that the update was completed successfully for the node:
    1. Go to the advanced privilege level : set -privilege advanced
    2. Verify that update status is complete for the node: system node upgrade-revert show -node nodenameB
      The status should be listed as complete.

      If the status is not complete, from the node, run the system node upgrade-revert upgrade command. If the command does not complete the update, contact technical support.

    3. Return to the admin privilege level: set -privilege admin
  18. Verify that the node's ports are up: network port show -node nodenameB

    You must run this command on a node that has been upgraded to ONTAP 9.5.

    Example

    The following example shows that all of the node's data ports are up:
    TDC-DM7K::> network port show -node TDC-7K-01
    Speed (Mbps)
    Node Port IPspace Broadcast Domain Link MTU Admin/Oper
    ------ --------- ------------ ---------------- ----- ------- ------------
    TDC-7K-01
    e0M Default - up 1500 auto/100
    e0a Default - up 1500 auto/1000
    e0b Default - up 1500 auto/1000
    e1a Cluster Cluster up 9000 auto/10000
    e1b Cluster Cluster up 9000 auto/10000
    5 entries were displayed.

  19. Revert the LIFs back to the node: network interface revert *

    This command returns the LIFs that were migrated away from the node.

    Example

    TDC-DM7K::> network interface revert *      
    8 entries were acted on.

  20. Verify that the node's data LIFs successfully reverted back to the node, and that they are up: network interface show

    Example

    The following example shows that all of the data LIFs hosted by the node is successfully reverted back to the node, and that their operational status is up:
    TDC-DM7K::> network interface show
    Logical Status Network Current Current Is
    Vserver Interface Admin/Oper Address/Mask Node Port Home
    ----------- ---------- ---------- ------------------ ------------- ------- ----
    vs0
    data001 up/up 192.0.2.120/24 TDC-7K-01 e0a true
    data002 up/up 192.0.2.121/24 TDC-7K-01 e0b true
    data003 up/up 192.0.2.122/24 TDC-7K-01 e0b true
    data004 up/up 192.0.2.123/24 TDC-7K-01 e0a true
    4 entries were displayed.

  21. If you previously determined that this node serves clients, verify that the node is providing service for each protocol that it was previously serving: system node run -node nodenameB -command uptime

    The operation counts reset to zero during the update.

    Example

    The following example shows that the updated node has resumed serving its NFS and iSCSI clients:
    TDC-DM7K::> system node run -node TDC-7K-01 -command uptime
    3:15pm up 0 days, 0:16 129 NFS ops, 0 CIFS ops, 0 HTTP ops, 0 FCP ops, 2 iSCSI ops

  22. If this was the last node in the cluster to be updated, trigger an AutoSupport notification: autosupport invoke -node * -type all -message "Finishing_NDU"

    This AutoSupport notification includes a record of the system status just prior to update. It saves useful troubleshooting information in case there is a problem with the update process.

    If the cluster is not configured to send AutoSupport messages, a copy of the notification is saved locally.

  23. Confirm that the new ONTAP software is running on both nodes of the HA pair: system node image show

    Example

    In the following example, image2 is the updated version of ONTAP and is the default version on both nodes:
    TDC-DM7K::*> system node image show
    Is Is Install
    Node Image Default Current Version Date
    -------- ------- ------- ------- --------- -------------------
    TDC-7K-03
    image1 false false X.X.X MM/DD/YYYY TIME
    image2 true true Y.Y.Y MM/DD/YYYY TIME
    TDC-7K-01
    image1 false false X.X.X MM/DD/YYYY TIME
    image2 true true Y.Y.Y MM/DD/YYYY TIME
    4 entries were displayed.

  24. Reenable automatic giveback on the partner node if it was previously disabled: storage failover modify -node nodenameA -auto-giveback true
  25. Verify that the cluster is in quorum and that services are running by using the cluster show and cluster ring show (advanced privilege level) commands.

    You must perform this step before upgrading any additional HA pairs.
  26. Return to the admin privilege level: set -privilege admin

After you finish

Upgrade any additional HA pairs.