The transition to NetApp MS Azure AD B2C is complete. If you missed the pre-registration, you will be invited to reigister at next log in.
Please note that access to your NetApp data may take up to 1 hour.
To learn more, read the FAQ and watch the video.
Need assistance? Complete this form and select “Registration Issue” as the Feedback Category.

ONTAP Discussions

OnTap 8.3.1P1 - MetroCluster - Raid Scrub and super high latency

mclawler75

I have two C-Dot FIlers at my location, one a 4 node MetroCluster, and one a standard C-Dot FIler.  We don't see issues with the Raid Scrub on the Standard C-Dot FIler (a two node 8060), but it beats the crap out of our 4 node MetroCluster (FAS8040's).  Is anyone else seeing this, the standard window for the Raid Scrub is 1am - 5am, if you look in the performance manager you should see a spike during those hours.

 

Hopeing I'm not the only one getting hit with this, to take care of the problem while Support digs into it we have simply killed the Raid Scrub options, not ideal I know, but we were leaving trucks at the shipping dock waiting for the Filer to serve up data.

3 REPLIES 3

ttran

Hi,

 

Would you happen to have "thorough scrub" enabled for any aggregates on your 4-node Metrocluster Cluster as this will definitely increase the length and system utilization for a scrub? In addition, the typical scrub will utilize more system resources in a Metrocluster configuration compared to a non-metrocluster configuration due to the remote mirror.

 

From the cluster shell you can run the following command to query if you have thorough scrub enabled for any aggregate(s):

> storage aggregate show -aggregate * -thorough-scrub on

 

If the feature is not enabled then you see the following output:

"There are no entries matching your query"

 

 

Team NetApp

mclawler75

No, that option is not set.

 

Its odd that the MC would see such a huge spike, I go from .05 (roughly) latency, to 10.x latency during this Raid Scrub.

 

On a standard c-dot filer I don't even see the job running other than its listed in the logs.

mclawler75

So, after working with the critical care team our issue has been tracked back to us over saturating the ATTO bridges.  The solution is to add another 4 more ATTO bridges to the mix.

Announcements
NetApp on Discord Image

We're on Discord, are you?

Live Chat, Watch Parties, and More!

Explore Banner

Meet Explore, NetApp’s digital sales platform

Engage digitally throughout the sales process, from product discovery to configuration, and handle all your post-purchase needs.

NetApp Insights to Action
I2A Banner
Public