2014-06-06 09:25 AM
I'm working on a new SnapProtect 10 SP6 setup and battling a lack of information
Here is the scenario: I have an Office and a Datacenter connected with a 100Mbit WAN Link about 500km apart. I am trying to build a protection scenario where a CIFS Volume in the Office gets snapshots hourly / daily / weekly, and a copy is snapvaulted to the datacenter weekly. So far I have that working. Eventually I might add a cut to tape from the datacenter as an additional layer but for the moment it's not in play. The primary SnapProtect server / Media Agent is in the Datacenter.
My question has to do with catalogs. The volume in play has over 7 million files on it. If I configure the jobs to perform cataloging, the indexing process takes Many Many Hours to complete... when the goal is hourly snapshots, this is obviously unworkable.
At first I wondered if I even needed catalogs; but it seems like without them you lose a lot of the search capabilities when it comes to restores. If I know the filename(s) and locations of them I can obviously just mount a snap to restore, but we do frequently get much more fuzzy restore requests.
So I need to figure out a way to improve the performance of cataloging. Here's what I've tried:
I'm wondering if anyone has any tips or best practices or advice with regards to configuring SnapProtect jobs for cataloging. I see that one of the volumes I'm going to eventually need to throw in here has 26million files in it (almost 4 times the one I'm struggling with now!)
Any help would be appreciated.
2014-06-06 10:06 AM
Some more info (answering some of my own questions):
I started testing with a smaller volume (~800,000 files) and sat watching the CVNasFileScan.log file while a few things ran.
So maybe I'm overthinking this...if I get through one catalog of a volume with a big number of files, presumably each catalog process will just process changed files. That makes me happy but I'm confused still