2011-06-27 07:53 PM
Have a FAS 2050 in production and a FAS2020 in DR both running Ontap version 7.3.4 running SnapMirror
I have a few questions:
Any help here would be much appreciated.
2011-06-27 11:48 PM
This error is correct, if you want to grow the destination volume beyond the 1TB Dedupe limit, you will have to:
1. break snapmirror
2. stop dedupe
3. "sis undo" dedupe (priv set diag)
4. and then update the snapmirror
WARNING - for this procedure:
To increase the size of an A-SIS enabled volume beyond the maximum limit for A-SIS, the A-SIS service must be turned off and the changes undone. Undoing A-SIS will re-inflate the file system and could require more disk space than is available in the A-SIS enabled volume. There is no way to expand the volume size until the undo is completed, so the recommended course of action is to create and use a temporary volume and migrate data necessary to free enough space for the re-inflation to complete.
The destination volume will always be non deduped if you want to stay with the fas2020 and >1TB.
- replace fas2020 with 2040 or 2050 or even bigger
- upgrade to Ontap 8.0.1 (or higher) where the limits are different (but you'll also need to replace the 2020 in this case, because this controller cannot run anything higher then 2.x)...
Hope this helps,
2011-06-28 12:07 AM
Correct me if I'm wrong here but dedup on the destination should not really do a lot as its turned on at Source so effectively the deduplicated blocks are being replicated as it is.
So you are getting the benefits anyway. I guess this just all depends on the rate of change and the frequency of the replication passes as to whether you are going to see full benefits of the dedup in the destination.
So if this is the case turning off dedup at destination, setting up another flexvol along with a new snapmirror job (schedule) initialisation, should give similar results again depending on rate of change for the data and the frequency of the replication passes.
This way we can allow the volume to grow in a suitable manner (> 1tb) and not limited via the constraints of FAS2020 and the ontap version at this point, or atleast until they hit the 2tb limit at the source on the FAS2050.
2011-06-28 12:40 AM
Yes you are correct with that the destination is not "involved" in the dedupe process in a Volume SnapMirror, however with Volume SnapMirror the lower volume size limit has higher priority, so you are "bound" to the 1TB limit of the 2020.
If you want to do the "trick" with the new volume on the destination, you would need to change it to a QTREE SnapMirror. Then you "loose" the bandwidth savings of transfering only deduped blocks, but you could grow the destination to the vol limit of the FAS2020 (until they hit the 2TB 2050 limit).
Hope this helps,
Excerpt of TR-3505:
Volume SnapMirror allows you to back up your data to another location for disaster recovery purposes.
Deduplication is supported with volume SnapMirror.
Volume SnapMirror operates at the physical block level; thus when deduplication is enabled on the source,
the data sent over the wire for replication is also deduplicated and therefore the savings are inherited at the destination.
This can significantly reduce the
amount of network bandwidth required during replication.
To run deduplication with volume SnapMirror:
- Deduplication can only be managed on the source system—the flexible volume at the destination
system inherits all the efficiency attributes and storage savings.
- Shared blocks are transferred only once, so deduplication reduces network bandwidth usage.
- The volume SnapMirror update schedule is not tied to the deduplication schedule.
- Maximum volume size limits for deduplicated volumes are constrained to the lower limit between the
source and the destination systems.
2011-06-28 01:25 AM
Again Thanks Peter,
I had over looked (putting it nicely on myself) the constraint of the 'Maximum volume size limits for deduplicated volumes are constrained to the lower limit between the
source and the destination systems'.
For now the best approach is stick with deduplication and restructure the data sets and create new volumes accordingly within the maximums and the client at some point will look at newer / higher performing filers.
Thanks for your time.