I have to move the root aggregate on a cluster node to different disks, and I have a couple questions for anyone who's done this before.
Background: we have a 4-node AFF cluster where support is ending on two nodes on 1/31. We are going to migrate all volumes to the remaining 2 nodes and move the shelves over after depopulating them. There is currently one shelf per node, all the same size, except for a single tiny additional shelf attached to the nodes we are depopulating. I'm going to use this as my "lifeboat" to move the root aggregate to from the first shelf I depopulate so the node continues to function after removing the shelf. We are running OnTAP 9.5P14.
My plan is to run the "system node migrate-root" command, however I can't find documentation indicating how this process works. At what point is there a reboot? How long does the process take? Does anyone have insight into this?
Also, I would like to use the minimal amount of disks, as I may need the rest of this small shelf to provide buffer space during later migrations, so I am hoping to only use three disks in the disklist parameter. Does anyone know if that will work in OnTAP 9.5? Do I have to use 5 disks? We're going to decom the shelf after this is all over so I'm not concerned with long-term.
If there's some way to partition the shelf and migrate the aggregate from an ADP to an ADP configuration, that would be even better. Would love to hear any suggestions or comments!
Thank you @SpindleNinja ! I've read that KB actually, and it is helpful, but it doesn't speak to whether or not you can configure three disks for the aggregate (which I hope is the case). Even though it describes in general what happens (i.e. creating new aggregate, backing up/restoring config, etc.) it seems a little vague on when / how the failover occurs. I just like to have as much advance knowledge as possible regarding what to expect so I can schedule an appropriate outage window for our CIFS Servers.
It mentions to note the min aggr root size via HWU and the that the min disk count based on raid type.
If I recall correctly (I was on the partner PS side of the world when I did it last).. it will do an ARL of the aggrs to the partner node and will do 2 reboots. You can also move your NAS lifs over before hand.
Thanks @SpindleNinja . I have a hard time navigating Hardware Universe ... I couldn't specifically figure out how to determine the number of disks in a non-partitioned root aggregate is supported. What is an ARL?
Thanks @SpindleNinja . I think I have a better idea. Minimum size is less than the usable storage on one disk, so I think 3 disks will work. I guess we'll see! Opening a case is probably a good idea, just wasn't sure if you could open a case just to vet a plan.
The process uses aggregate relocation. As far as the process goes after starting the command, all the data aggrs are relocated to the node not having the root change. This way the node being changed may reboot freely without impact. The cluster is temporarily disabled. A number of node reboots occur to provision, configure and restore to the new root aggr. At the end, the data aggregates are sent back to the original node. If I recall, the node reboots 3-4 times and the process takes somewhere between 20-40 minutes per controller.
If you have access to NetApp support, check out the tools page and the the Hands On Labs. Try the ONTAP 9.7 lab. After you setup the nodes and the cluster, you can actually try out the new root there using disks. My suggestion: simply create the base cluster and no data aggregates. you will need the disks to create a new root.
Logging into the "serial port" (mentioned in the doc for the lab) allows you to watch the process.
Honestly, the first time is a little scary. You just need to let the process run to completion and do not disturb it!
Very helpful @TMAC_CTG ! Thank you. I will actually have all data aggregates decommissioned on the node I'm doing this on by the time I run the command, so that should make it faster. I am concerned about this one comment:
The cluster is temporarily disabled.
Can you elaborate? I would assume you don't mean the entire cluster is literally disabled, maybe just the HA functionality on the two affected nodes? To clarify, this is a 4 node cluster. I don't want the entire remaining 3 nodes to be un-clustered during this!
The Hands on Lab is a great idea. I may try that. Thank you again!