ONTAP Discussions
ONTAP Discussions
Hi All,
I have run into snapmirror transfer errors due to the destination is being full.
However the destination volume is not full, rather one of the two constituents is:
> volume show -is-constituent true
Vserver Volume Aggregate State Type Size Available Used%
--------- ------------ ------------ ---------- ---- ---------- ---------- -----
<--snip-->
aggr_01_2 online DP 100TB 16KB 99%
aggr_01_1 online DP 60TB 20.59TB 65%
When I increased the size of the Volume ontap added the additional space to the constituent that was not full while leaving the constituent that was full in the full state.
volume modify -vserver name -volume My_1_vault -size +20TB
<---snip--->
aggr_01_2 online DP 100TB 16KB 99% <<< NO increase
aggr_01_1 online DP 100.00TB 60.59TB 39% <<< 40TB increase
Is there any way to increase one specific constituent?
I was thinking that I could add a third constituent but it would still leave the full constituent full.
Thank you for any ideas to solve this!
Bela
Solved! See The Solution
Hi there,
You can modify the size of the constituent volumes with "volume modify -vserver vserver_name -volume fg_name -size new_size", however these are already 100TB, which is the maximum size for a volume prior to ONTAP 9.12.1P2. Once on that version or newer, you can expand to up to 300TB, as per instructions at https://kb.netapp.com/onprem/ontap/os/Do_transitioned_volumes_support_large_volumes_and_files
However, a better solution may to be rebalance the flexgroup constituents, which will move files from the full one to the empty one. This is available non-disruptively from.. 9.12.1P2. Instructions at https://docs.netapp.com/us-en/ontap/flexgroup/manage-flexgroup-rebalance-task.html
Hope this helps!
Hi there,
You can modify the size of the constituent volumes with "volume modify -vserver vserver_name -volume fg_name -size new_size", however these are already 100TB, which is the maximum size for a volume prior to ONTAP 9.12.1P2. Once on that version or newer, you can expand to up to 300TB, as per instructions at https://kb.netapp.com/onprem/ontap/os/Do_transitioned_volumes_support_large_volumes_and_files
However, a better solution may to be rebalance the flexgroup constituents, which will move files from the full one to the empty one. This is available non-disruptively from.. 9.12.1P2. Instructions at https://docs.netapp.com/us-en/ontap/flexgroup/manage-flexgroup-rebalance-task.html
Hope this helps!
Alex, is there any way to target the exact constituent volume and increase the size of that constituent volume without increasing the size of the entire flexgroup volume?
Hi AlexDawson
I am interested in knowing if the single constituent volume can be increased in space.
Many thanks in advance
Alessandro
Hi there,
It is important to keep all flexgroup constituent volumes the same size, so there is no supported way to shrink or expand a single constituent. If you find individual volumes being unreasonably used, it is suggested to rebalance the flexgroup.
Thanks!
Hello
I don't know how it happened, but a customer created a flexgroup via GUI, but it has constituent volumes of different sizes. The strange thing is that the "size" is different while the "available" is the same. What do you suggest we do? See below:
Vserver Volume Aggregate State Type Size Available Used%
--------- ------------ ------------ ---------- ---- ---------- ---------- -----
svm-nfs vol01_nfs__0001 aggr_data_01 online RW 5.19TB 365.2GB 93%
svm-nfs vol01_nfs__0002 aggr_data_01 online RW 8.37TB 365.2GB 95%
svm-nfs vol01_nfs__0003 aggr_data_03 online RW 1.92TB 365.2GB 81%
svm-nfs vol01_nfs__0004 aggr_data_01 online RW 1.95TB 365.2GB 81%
svm-nfs vol01_nfs__0005 aggr_data_01 online RW 1.77TB 365.2GB 79%
svm-nfs vol01_nfs__0006 aggr_data_03 online RW 2.19TB 365.2GB 83%
svm-nfs vol01_nfs__0007 aggr_data_03 online RW 1.87TB 365.2GB 80%
svm-nfs vol01_nfs__0008 aggr_data_01 online RW 9.28TB 365.2GB 96%
svm-nfs vol01_nfs__0009 aggr_data_03 online RW 9.45TB 365.2GB 96%
svm-nfs vol01_nfs__0010 aggr_data_01 online RW 1.88TB 365.2GB 81%
svm-nfs vol01_nfs__0011 aggr_data_01 online RW 9.23TB 365.2GB 96%
svm-nfs vol01_nfs__0012 aggr_data_03 online RW 1.87TB 365.2GB 80%