Skip to main content

Correcting misaligned spare partitions

When you add partitioned disks to an aggregate, you must leave a disk with both the root and data partition available as spare for every node. If you do not and your node experiences a disruption, ONTAP cannot dump the core to the spare data partition.

Before you begin

You must have both a spare data partition and a spare root partition on the same type of disk owned by the same node.
  1. Display the spare partitions for the node: storage aggregate show-spare-disks -original-owner node_name

    Note which disk has a spare data partition (spare_data) and which disk has a spare root partition (spare_root). The spare partition will show a non-zero value under the Local Data Usable or Local Root Usable column.
  2. Replace the disk with a spare data partition with the disk with the spare root partition: storage disk replace -disk spare_data -replacement spare_root -action start

    You can copy the data in either direction; however, copying the root partition takes less time to complete.
  3. Monitor the progress of the disk replacement: storage aggregate show-status -aggregate aggr_name
  4. After the replacement operation is complete, display the spares again to confirm that you have a full spare disk: storage aggregate show-spare-disks -original-owner node_name
    You should see a spare disk with usable space under both Local Data Usable and Local Root Usable.

Example

You display your spare partitions for node c1-01 and see that your spare partitions are not aligned:

c1::> storage aggregate show-spare-disks -original-owner c1-01

Original Owner: c1-01
Pool0
Shared HDD Spares
Local Local
Data Root Physical
Disk Type RPM Checksum Usable Usable Size
--------------------------- ----- ------ -------------- -------- -------- --------
1.0.1 SAS 7200 block 753.8GB 0B 828.0GB
1.0.10 SAS 7200 block 0B 73.89GB 828.0GB


You start the disk replacement job:

c1::> storage disk replace -disk 1.0.1 -replacement 1.0.10 -action start

While you are waiting for the replacement operation to finish, you display the progress of the operation:

c1::> storage aggregate show-status -aggregate aggr0_1

Owner Node: c1-01
Aggregate: aggr0_1 (online, raid_dp) (block checksums)
Plex: /aggr0_1/plex0 (online, normal, active, pool0)
RAID Group /aggr0_1/plex0/rg0 (normal, block checksums)
Usable Physical
Position Disk Pool Type RPM Size Size Status
-------- --------------------------- ---- ----- ------ -------- -------- ----------
shared 1.0.1 0 SAS 7200 73.89GB 828.0GB (replacing, copy in progress)
shared 1.0.10 0 SAS 7200 73.89GB 828.0GB (copy 63% completed)
shared 1.0.0 0 SAS 7200 73.89GB 828.0GB (normal)
shared 1.0.11 0 SAS 7200 73.89GB 828.0GB (normal)
shared 1.0.6 0 SAS 7200 73.89GB 828.0GB (normal)
shared 1.0.5 0 SAS 7200 73.89GB 828.0GB (normal)

After the replacement operation is complete, you confirm that you have a full spare disk:

ie2220::> storage aggregate show-spare-disks -original-owner c1-01

Original Owner: c1-01
Pool0
Shared HDD Spares
Local Local
Data Root Physical
Disk Type RPM Checksum Usable Usable Size
--------------------------- ----- ------ -------------- -------- -------- --------
1.0.1 SAS 7200 block 753.8GB 73.89GB 828.0GB