We did it at a customer successfully but I can't recommend using it. SnapMirror Migrate does move the volume and all file handles, but it does not update exports...so once the migrate completed we quickly had to update exports and run exportfs to keep the mount. Data Motion for volumes in c-mode handles nas and san and that looks like the direction to get full support for migrating any volume... with 7-mode LUN volumes only. SnapMirror Migrate almost seems like it was never finished in that it doesn't update shares and exports after migration.
I have also used snapmirror migrate a few times, but only to prevent remounting filesystems on several hundreds of nfs clients. As Scott already mentioned with vmware you have to be really quick with re-exporting the filesystem or else your vm's will crash. With e.g. solaris or linux the nfs client is far more resilient and does not care about waiting for several minutes.
When you mentioned that the NFS will get disabled for the whole filer, do you mean that Admin has to do some thing or the system takes care of enabling the NFS automatically. And what is the typical delay between NFS disable and NFS enable phase?
it turns back on automatically, but the time is dependent on the mirror cutover of the snapmirror migrate volume. The migrated volume takes the vol fsid and file handles so clients don't know it moved...but it does not update the export the volume to the new location so that needs to be updated quickly right after the migrate finishes. It almost seems like snapmirror migrate was a partial vol move (most pieces there) several years ago... clients should handle the timeout but it isn't great that all volumes get affected since the protocol is stopped for the one migrating volume...then the single export update has to be updated as well so a couple of gotchas with this method.
But in a scenario where it is totally linux based environment accessing multiple NFS shares on Netapp, we can use it. As you know few minutes of NFS down won't affect the linux clients, they just hang for sometime, and then continue. Also even if the mount points still shown as old after the migrate, but that won't affect the runs going on.
So I call it pseudo-nondisruptive 🙂 for some complete NFS and linux environment. I donot know how a VMWare environment using NFS will behave.
As you know I was looking for smooth NFS volumen movement within a filer, and datamotion on 8.0 was no use for me, but thanks o you, atleast I can use the above method. I will also summarize in my original post once I do some good testing.
A general warning for anyone planning on using snapmirror migrate between systems running ONTAP 8.0.x and 8.1 7-mode systems: it does not work!
The snapmirror migrate aborts because it can't complete the file handle transfer to the new system, this is a know bug (547843). Luckily I found this one while performing my migration test and not during the actual controller hange (and migration of data to new disks). Only workaround at the moment was to upgrade the source controllers to ONTAP 8.1 and then execute the snapmirror migrates.
I agree that in most cases it will be intra-controller migrations and you will not run into this issue.
My case I was replacing the controllers/some of the disk shelves and the some disk shelves including data need to be reused. I had the new controllers up and running with the new disk shelves and a temporary host name/ip address. For the data on the disk shelves that had to be replaced I performed the base snapmirror transfer to the new controllers/disks.
During the change window, I performed the snapmirror migrate to the new controllers., shutdown the old controllers, connected the to be reused disk shelves including the data on them to the new controllers. Then the new controllers were renamed to the original controllers including IP addresses. Worked really well and this way we did not need to reboot 500 NFS clients
I also did many times NFS volume migration in the linux environment. Several times there was no issue as immediately updated the export but few times it gives error on the client PC due to stale file system need to unmount and remount it.