Skip to main content

Performing aggregate healing and restoring mirrors (MetroCluster FC configurations)

After replacing hardware and assigning disks, you can perform the MetroCluster healing operations. You must then confirm that aggregates are mirrored and, if necessary, restart mirroring.

  1. Perform the two phases of healing (aggregate healing and root healing) on the disaster site:

    Example

    cluster_B::> metrocluster heal -phase aggregates

    cluster_B::> metrocluster heal -phase root aggregates

  2. Monitor the healing and verify that the aggregates are in either the resyncing or mirrored state: storage aggregate show -node local
    If the aggregate shows this state...Then...
    resyncingNo action is required.

    Let the aggregate complete resyncing.

    mirror degradedProceed to step 3.
    mirrored, normalNo action is required.
    unknown, offlineThe root aggregate shows this state if all the disks on the disaster sites were replaced.

    Example

    cluster_B::> storage aggregate show -node local 

    Aggregate Size Available Used% State #Vols Nodes RAID Status
    --------- -------- --------- ----- ------- ------ ---------- ------------
    node_B_1_aggr1
    227.1GB 11.00GB 95% online 1 node_B_1 raid_dp,
    resyncing
    NodeA_1_aggr2
    430.3GB 28.02GB 93% online 2 node_B_1 raid_dp,
    mirror
    degraded
    node_B_1_aggr3
    812.8GB 85.37GB 89% online 5 node_B_1 raid_dp,
    mirrored,
    normal
    3 entries were displayed.

    cluster_B::>

    In the following examples, the three aggregates are each in a different state:

    NodeState
    node_B_1_aggr1resyncing
    node_B_1_aggr2mirror degraded
    node_B_1_aggr3mirrored, normal
  3. If one or more plexes remain offline, additional steps are required to rebuild the mirror.

    In the preceding table, the mirror for node_B_1_aggr2 must be rebuilt.

    1. View details of the aggregate to identify any failed plexes: storage aggregate show -r -aggregate node_B_1_aggr2

      Example

      In the following example, plex /node_B_1_aggr2/plex0 is in a failed state:

      cluster_B::> storage aggregate show -r -aggregate node_B_1_aggr2

      Owner Node: node_B_1
      Aggregate: node_B_1_aggr2 (online, raid_dp, mirror degraded) (block checksums)
      Plex: /node_B_1_aggr2/plex0 (offline, failed, inactive, pool0)
      RAID Group /node_B_1_aggr2/plex0/rg0 (partial)
      Usable Physical
      Position Disk Pool Type RPM Size Size Status
      -------- ------------------------ ---- ----- ------ -------- -------- ----------

      Plex: /node_B_1_aggr2/plex1 (online, normal, active, pool1)
      RAID Group /node_B_1_aggr2/plex1/rg0 (normal, block checksums)
      Usable Physical
      Position Disk Pool Type RPM Size Size Status
      -------- ------------------------ ---- ----- ------ -------- -------- ----------
      dparity 1.44.8 1 SAS 15000 265.6GB 273.5GB (normal)
      parity 1.41.11 1 SAS 15000 265.6GB 273.5GB (normal)
      data 1.42.8 1 SAS 15000 265.6GB 273.5GB (normal)
      data 1.43.11 1 SAS 15000 265.6GB 273.5GB (normal)
      data 1.44.9 1 SAS 15000 265.6GB 273.5GB (normal)
      data 1.43.18 1 SAS 15000 265.6GB 273.5GB (normal)
      6 entries were displayed.

      cluster_B::>
    2. Delete the failed plex: storage aggregate plex delete -aggregate aggregate-name -plex plex
    3. Reestablish the mirror: storage aggregate mirror -aggregate aggregate-name
    4. Monitor the resynchronization and mirroring status of the plex until all mirrors are reestablished and all aggregates show mirrored, normal status: storage aggregate show