<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Netapp-harvest + Graphite + grafana, throughput is way off in Active IQ Unified Manager Discussions</title>
    <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112787#M19958</link>
    <description>&lt;P&gt;It's been a week so not sure if you're still looking for some insight but I see part of your issue here.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you look carefully at your first image, with the raw graphite data, you'll see that while on the left you've highlighted write_data, in the legend for the graph you're showing the metrics:&lt;/P&gt;&lt;P&gt;...aggr.total_transfers &amp;nbsp;and ...aggr.Node02_SSD.total_transfers. &amp;nbsp; &amp;nbsp;Remember that the graphite interface will show each metric you double click, and it removes them when you again double click them. &amp;nbsp; Its easy to end up looking at a bunch of unrelated items this way.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In your grafana dashboard you're looking at Node.xxx.fcp.read_data and&amp;nbsp;&lt;SPAN&gt; Node.xxx.fcp.write_data. &amp;nbsp; &amp;nbsp; These don't measure the same things. &amp;nbsp; &amp;nbsp;&lt;/SPAN&gt;Aggregate total transfers does not equal protocol reads+writes for a node. &amp;nbsp; &amp;nbsp;Each aggregate is measuring its own operations to and from disk. &amp;nbsp; The protocols are each [FCP, iSCSI, CIFS, NFS] measuring their operations from and to&amp;nbsp;the client. &amp;nbsp; &amp;nbsp;Depending on what you're wanting to measure you'd look at one or the other or both together but they will not show the same values.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I hope that helps. &amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Fri, 20 Nov 2015 21:36:41 GMT</pubDate>
    <dc:creator>cbiebers</dc:creator>
    <dc:date>2015-11-20T21:36:41Z</dc:date>
    <item>
      <title>Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112553#M19919</link>
      <description>&lt;P&gt;This is an odd one, and I must admit I am very new to this whole tool set.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Followed this install process&lt;/P&gt;&lt;P&gt;&lt;A href="http://blog.pkiwi.com/netapp-advanced-performance-monitoring-with-harvest-graphite-and-grafana/" target="_blank"&gt;http://blog.pkiwi.com/netapp-advanced-performance-monitoring-with-harvest-graphite-and-grafana/&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;System is running RedHat 7.1 64X&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here is what I am seeing. &amp;nbsp;Throughput for everything seems to be badly off.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In &lt;SPAN&gt;Graphite I&lt;/SPAN&gt;&amp;nbsp;see the correct speeds. &amp;nbsp;But in&amp;nbsp;grafana, everything is off. &amp;nbsp;As best I can tell this seems to apply to all throughput values. &amp;nbsp;So NFS/CIFS/FCP/Disk...&lt;span class="lia-inline-image-display-wrapper" image-alt="graphite.png" style="width: 999px;"&gt;&lt;img src="https://community.netapp.com/t5/image/serverpage/image-id/4272i8E3AF39D71F023B5/image-size/large?v=v2&amp;amp;px=999" role="button" title="graphite.png" alt="graphite.png" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="grafana.png" style="width: 999px;"&gt;&lt;img src="https://community.netapp.com/t5/image/serverpage/image-id/4273i284B7D7828986313/image-size/large?v=v2&amp;amp;px=999" role="button" title="grafana.png" alt="grafana.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Any ideas?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 04 Jun 2025 22:46:51 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112553#M19919</guid>
      <dc:creator>DingbatCA</dc:creator>
      <dc:date>2025-06-04T22:46:51Z</dc:date>
    </item>
    <item>
      <title>Re: Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112787#M19958</link>
      <description>&lt;P&gt;It's been a week so not sure if you're still looking for some insight but I see part of your issue here.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you look carefully at your first image, with the raw graphite data, you'll see that while on the left you've highlighted write_data, in the legend for the graph you're showing the metrics:&lt;/P&gt;&lt;P&gt;...aggr.total_transfers &amp;nbsp;and ...aggr.Node02_SSD.total_transfers. &amp;nbsp; &amp;nbsp;Remember that the graphite interface will show each metric you double click, and it removes them when you again double click them. &amp;nbsp; Its easy to end up looking at a bunch of unrelated items this way.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In your grafana dashboard you're looking at Node.xxx.fcp.read_data and&amp;nbsp;&lt;SPAN&gt; Node.xxx.fcp.write_data. &amp;nbsp; &amp;nbsp; These don't measure the same things. &amp;nbsp; &amp;nbsp;&lt;/SPAN&gt;Aggregate total transfers does not equal protocol reads+writes for a node. &amp;nbsp; &amp;nbsp;Each aggregate is measuring its own operations to and from disk. &amp;nbsp; The protocols are each [FCP, iSCSI, CIFS, NFS] measuring their operations from and to&amp;nbsp;the client. &amp;nbsp; &amp;nbsp;Depending on what you're wanting to measure you'd look at one or the other or both together but they will not show the same values.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I hope that helps. &amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 20 Nov 2015 21:36:41 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112787#M19958</guid>
      <dc:creator>cbiebers</dc:creator>
      <dc:date>2015-11-20T21:36:41Z</dc:date>
    </item>
    <item>
      <title>Re: Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112788#M19959</link>
      <description>&lt;P&gt;Yep, still hunting for an answer.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Non of the throughput values seem correct in&amp;nbsp;grafana. &amp;nbsp;Not just that one. &amp;nbsp;I know I am pushing 400~600MB/s on average through my FAS8060, but grafana is showing 0.4~1.2MB/s. &amp;nbsp;So there is something off, I just have no clue where to look.&lt;/P&gt;</description>
      <pubDate>Fri, 20 Nov 2015 21:42:31 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112788#M19959</guid>
      <dc:creator>DingbatCA</dc:creator>
      <dc:date>2015-11-20T21:42:31Z</dc:date>
    </item>
    <item>
      <title>Re: Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112790#M19961</link>
      <description>&lt;P&gt;I just did a comparison to the values below in grafana with Brocade Switch View for the physical ports in my environment. &amp;nbsp;(Because I'm comparing to physical ports, I have to look at the node's physical port values, not the SVM LIF values, or I'd have to do a bunch of math. &amp;nbsp; This is the metric that you show on the left side in the initial screenshot of graphite, but not the metric you were actually displaying.)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I use the&amp;nbsp;&amp;nbsp;Network Port dashboard &amp;nbsp;and&amp;nbsp; reference these metrics in the Fibre Channel row. &amp;nbsp; (Metrics captured by choosing edit on the graph).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;netapp perf $Group $Cluster node $Node &lt;STRONG&gt;fcp_port&lt;/STRONG&gt; $Port write_data highestAverage($TopResources) aliasByNode(5, 7)&lt;BR /&gt;netapp perf $Group $Cluster node $Node &lt;STRONG&gt;fcp_port&lt;/STRONG&gt; $Port read_data highestAverage($TopResources) aliasByNode(5, 7)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When I check these values against the Brocade values I'm in alignment, with minor variation because the Brocade switchview GUI is displaying 30second averages to Harvest's 1 minute averages.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;What do you see when you look at the values above in Grafana either on your own dashboard, or using the one provided in the package?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;EDIT: &amp;nbsp;Note the bold piece is different than what you were showing in grafana.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 20 Nov 2015 23:35:48 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112790#M19961</guid>
      <dc:creator>cbiebers</dc:creator>
      <dc:date>2015-11-20T23:35:48Z</dc:date>
    </item>
    <item>
      <title>Re: Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112836#M19972</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In the netapp-harvest.conf file you will find a default key/value like this:&lt;/P&gt;&lt;P&gt;normalized_xfer &amp;nbsp; = mb_per_sec &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;What it will do is normalize all throughput numbers to MB/s. &amp;nbsp;So in Graphite and Grafana you are viewing in MB/s and not that of the native Data ONTAP counter manager counter being graphed. &amp;nbsp;I found normalizing data to be a much easier way of working; you can always scale back to whatever unit you want if needed for your use case.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regarding throughput being off, sometimes it is just user confusion because with cDOT the node that does the frontend protocol work is not necessarily the same that does the backend volume work. &amp;nbsp;Depending on the object you're looking at you may see frontend or backend numbers. &amp;nbsp;In the default "node" dashboard you will see "protocol &lt;STRONG&gt;backend&lt;/STRONG&gt; drilldown" and then things like "FCP &lt;STRONG&gt;frontend&lt;/STRONG&gt; drilldown" to show these both.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;So in the "frontend" views you see very detailed information about the IOPs arriving that node. &amp;nbsp;Those IOPs are then translated into WAFL messages and sent to the backend (on the same or different node) to be serviced. &amp;nbsp;At the "backend" the messages are tagged with protocol but otherwise are only tracked as read/write/other vs much more detail tracked at the "frontend" node. &amp;nbsp;If all traffic is direct (IOPs arrive on a LIF on the same node that owns the volume) then the "frontend" and "backend" numbers should agree,&amp;nbsp;but if you have indirect traffic they will be different.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Maybe you can check&amp;nbsp;your setup taking the above info into account and let us know if that helped? &amp;nbsp;&lt;/P&gt;&lt;P&gt;--If it does, please also "accept as answer"&amp;nbsp;the post that answered your question so that others will see the Q/A is answered.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Cheers,&lt;BR /&gt;Chris Madden&lt;/P&gt;&lt;P&gt;Storage Architect, NetApp EMEA (and author of Harvest)&lt;/P&gt;&lt;P&gt;Blog:&amp;nbsp;&lt;A href="http://blog.pkiwi.com/" target="_blank"&gt;It all begins with data&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2015 11:04:53 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112836#M19972</guid>
      <dc:creator>madden</dc:creator>
      <dc:date>2015-11-23T11:04:53Z</dc:date>
    </item>
    <item>
      <title>Re: Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112855#M19978</link>
      <description>&lt;P&gt;I feel like a n00b. &amp;nbsp;Thanks for pointing me to the "normalized_xfer"&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;cat netapp-harvest.conf | grep normalized_xfer&lt;/P&gt;&lt;P&gt;normalized_xfer &amp;nbsp; = mb_per_sec &amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;normalized_xfer &amp;nbsp; = gb_per_sec &amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;normalized_xfer &amp;nbsp; = gb_per_sec &amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;normalized_xfer &amp;nbsp; = gb_per_sec &amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Changed to&amp;nbsp;&lt;SPAN&gt;mb_per_sec, for the 3 involving netapp and everthing is now lining up perfectly!&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2015 16:52:40 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112855#M19978</guid>
      <dc:creator>DingbatCA</dc:creator>
      <dc:date>2015-11-23T16:52:40Z</dc:date>
    </item>
    <item>
      <title>Re: Netapp-harvest + Graphite + grafana, throughput is way off</title>
      <link>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112869#M19980</link>
      <description>&lt;P&gt;Great to hear! &amp;nbsp;The default dashboards assume you normalize perf info to mb_per_sec and capacity info (UCUM) to gb_per_sec so it was probably a copy/paste mistake between pollers of different server types. I can see myself doing this too so I'll have to think about improving usability here...&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;&lt;P&gt;Chris&lt;/P&gt;</description>
      <pubDate>Mon, 23 Nov 2015 20:19:04 GMT</pubDate>
      <guid>https://community.netapp.com/t5/Active-IQ-Unified-Manager-Discussions/Netapp-harvest-Graphite-grafana-throughput-is-way-off/m-p/112869#M19980</guid>
      <dc:creator>madden</dc:creator>
      <dc:date>2015-11-23T20:19:04Z</dc:date>
    </item>
  </channel>
</rss>

