Subscribe

Re: netapp-harvest ocum capacity metrics

Hi Moshe,

 

Good to hear you found the issue with regards to capacity metrics.  

 

For the LUN WARNING message every minute this can be encountered if you have a cluster which has LUNs but none of them are online.  These show up in a list of LUNs but do not have performance metrics available because they are not active.  This bug was in my backlog to fix and will be in the next Toolchest release.

 

Cheers,

Chris

 

 

 

 

Re: netapp-harvest ocum capacity metrics

Hi Chris,

 

I have the same problem. No error in the log but I get no data.

I have no idea what mhnap-cls is...

 

 

Re: netapp-harvest ocum capacity metrics

Hi @p_w

 

There should be something in the poller log (/opt/netapp-harvest/log) explaining why data can't be collected or forwarded to Graphite.  You could try restarting the poller (/opt/netapp-harvest/netapp-manager -restart) to see if something is logged at startup that would indicate a config error.  The "mhnap-cls" mentioned in the earlier post was the hostname from that other user and not something related to Harvest.

 

Cheers,
Chris Madden

Storage Architect, NetApp EMEA (and author of Harvest)

Blog: It all begins with data

 

If this post resolved your issue, please help others by selecting ACCEPT AS SOLUTION or adding a KUDO or both!

 

Re: netapp-harvest ocum capacity metrics

Hello,

 

My problem is similar, but I did not manage to solve it. The OCUM worker exits with an error.

Here is the OCUM worker log file :

cat /opt/netapp-harvest/log/NDC-PS-OCU-CDOT_netapp-harvest.log

[2016-07-06 17:07:28] [NORMAL ] WORKER STARTED [Version: 1.2.2] [Conf: netapp-harvest.conf] [Poller: NDC-PS-OCU-CDOT]
[2016-07-06 17:07:28] [NORMAL ] [main] Poller will monitor a [OCUM] at [10.197.208.166:443]
[2016-07-06 17:07:28] [NORMAL ] [main] Poller will use [password] authentication with username [netapp-harvest] and password [**********]
[2016-07-06 17:07:33] [WARNING] [sysinfo] Discovered [ndc-ps-cluster1] on OCUM server with no matching conf section; to collect this cluster please add a section
[2016-07-06 17:07:33] [WARNING] [sysinfo] Discovered [ndc-ps-cluster2] on OCUM server with no matching conf section; to collect this cluster please add a section
[2016-07-06 17:07:33] [NORMAL ] [main] Collection of system info from [10.197.208.166] running [6.4P1] successful.
[2016-07-06 17:07:33] [ERROR  ] [main] No best-fit collection template found (same generation and major release, minor same or less) found in [/opt/netapp-harvest/template/default].  Exiting;

 

Here is the monitored hosts section from my netapp-harvest.conf

##
## Monitored host examples - Use one section like the below for each monitored host
##

#====== 7DOT (node) or cDOT (cluster LIF) for performance info ================
#
[ndc-ps-fas01]
hostname       = 10.197.208.20
site           = Grenoble_BTIC

[ndc-ps-fas02]
hostname       = 10.197.208.30
site           = Grenoble_BTIC

[ndc-ps-fas3240-1]
hostname       = 10.197.208.80
site           = Grenoble_BTIC

[ndc-ps-fas3240-2]
hostname       = 10.197.208.90
site           = Grenoble_BTIC

[ntap-energetic]
hostname       = 10.197.189.167
site           = Grenoble_BTIC
username       = netapp-harvest
password       = XXXXXXX

[ndc-ps-cluster1]
hostname       = 10.197.208.29
site           = Grenoble_BTIC

[ndc-ps-cluster2]
hostname       = 10.197.208.32
site           = Grenoble_BTIC

#====== OnCommand Unified Manager (OCUM) for cDOT capacity info ===============
#
[NDC-PS-OCU-CDOT]
hostname          = 10.197.208.166
site              = Grenoble_BTIC
host_type         = OCUM
data_update_freq  = 900
normalized_xfer   = gb_per_sec

 

Thanks for your help.

Best regards.

Re: netapp-harvest ocum capacity metrics

[ Edited ]

 

@patrice_dronnier 

if you had this working with OCUM 6.3 or 6.2 ? 

 

[2016-07-06 17:07:33] [NORMAL ] [main] Collection of system info from [10.197.208.166] running [6.4P1] successful.
[2016-07-06 17:07:33] [ERROR  ] [main] No best-fit collection template found (same generation and major release, minor same or less) found in [/opt/netapp-harvest/template/default].  Exiting;

 

 

Re: netapp-harvest ocum capacity metrics

@patrice_dronnier @SrikanthReddy

 

Harvest v1.2.2 does work with OCUM 6.4 and a default template will be included in the next release.

 

For now you can simply copy the OCUM 6.3 one to 6.4:

 

# cp /opt/netapp-harvest/template/default/ocum-6.3.0.conf /opt/harvest-harvest/template/default/ocum-6.4.0.conf
# netapp-manager -start

Cheers,
Chris Madden

Storage Architect, NetApp EMEA (and author of Harvest)

Blog: It all begins with data

 

If this post resolved your issue, please help others by selecting ACCEPT AS SOLUTION or adding a KUDO or both!

Re: netapp-harvest ocum capacity metrics

 

I had a question to Chris about Multiple sites/clusters  on one OCUM server 

Only One OCUM server with one site is enough as confirmed by Chris 

 

"You should only add one poller for each OCUM server.  The Poller connects to the OCUM server and discovers all clusters known to it.  The poller then checks the conf file for FILER poller entries to learn the cluster -> site mapping for each cluster it discovered. "

Re: netapp-harvest ocum capacity metrics

Hello Chris,

With some delay, thanks for your answer.

Your solution solved my problem.

Best regards.

Patrice

Re: netapp-harvest ocum capacity metrics

I hope it's OK to post a reply/question to a thread that is marked closed.... I am having similar problems of not seeing capacity data on the capacity graphs of the NetAppHarvest dashboards ... and am having a hard time grasping whether I have things set up right or not.

 

Here is my OCUM poller entry from /opt/netapp-harvest/netapp-harvest.conf:

[OCUM_CEPINVAP296]
hostname = cepinvap296.centura.org
site = CENTURA
host_type = OCUM
data_update_freq = 300
auth_type = password
username = netapp-harvest
password = XXXXXX
normalized_xfer = gb_per_sec
template = ocum-opm-hierarchy.conf
graphite_root = netapp-capacity.Clusters.{display_name}
graphite_meta_metrics_root = netapp-capacity-poller.{site}

 

Here is an example poller entry for a CDOT NetApp within the same /opt/netapp-harvest/netapp-harvest.conf file that the OCUM poller entry resides:

[r1ppkrntap01]
hostname = r1ppkrntap01.corp.centura.org
site = CENTURA
host_enabled = 1

 

  • NOTE: All other CDOT entries look the same, except, oviously, the hostname

 

 

 

In our OCUM (6.4) server, all the CDOT NetApp clusters are defined in lower case

 

On my Harvest linux machine, these are the directories that have been created by Harvest:

/var/lib/graphite/whisper/netapp/perf/CENTURA/

/var/lib/graphite/whisper/netapp/capacity/CENTURA/

/var/lib/graphite/whisper/netapp-performance/Clusters/

/var/lib/graphite/whisper/netapp-capacity/Clusters/

 

All directories have 34 subdirs - one for each NetApp cluster... except the .../netapp/capacity/... directory... it only has 18 subdirs

 

When I do a whisper-dump on any of the files in cluster subdirs of .../netapp/capacity/..., I see one entry, way back in 2016, and all other entries are "0,          0"

 

When I do whisper-dump on any of the files in cluster subdirs of .../netapp-capacity/... I see up-to-date entries with valid values

 

And also, the capacity graphs in graphana all seem to use queries from the .../netapp/capacity/... data:

  • Example (the Top SVM Capacity Used Percent graph from the "Cluster Group" dashboard):
    • aliasByNode(sortByMaxima(highestAverage(netapp.capacity.$Group.$Cluster.svm.*.vol_summary.afs_used_percent, $TopResources)), 3, 5)

I would think these graphs would need to pull data from netapp-capacity.Clusters.$Cluster.

 

In my OCUM Harvest Poller log file, I think I have everything set up, as I see this, whenever I start it up:

[2017-04-20 11:58:36] [NORMAL ] WORKER STARTED [Version: 1.2.2] [Conf: netapp-harvest.conf] [Poller: OCUM_CEPINVAP296]
[2017-04-20 11:58:36] [NORMAL ] [main] Poller will monitor a [OCUM] at [cepinvap296.centura.org:443]
[2017-04-20 11:58:36] [NORMAL ] [main] Poller will use [password] authentication with username [netapp-harvest] and password [**********]
[2017-04-20 11:58:39] [WARNING] [sysinfo] Discovered [rclinvntap01] on OCUM server with no matching conf section; to collect this cluster please add a section
[2017-04-20 11:58:39] [WARNING] [sysinfo] Discovered [rclascntap01] on OCUM server with no matching conf section; to collect this cluster please add a section
[2017-04-20 11:58:39] [NORMAL ] [main] Collection of system info from [cepinvap296.centura.org] running [6.4] successful.
[2017-04-20 11:58:39] [NORMAL ] [main] Using specified collection template: [ocum-opm-hierarchy.conf]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1psahntap01] for host [r1psahntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.cepmmcntap01] for host [cepmmcntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.b1pinvntap01] for host [b1pinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.impascntap01] for host [impascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1psmcntap01] for host [r1psmcntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1psfmntap01] for host [r1psfmntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1pluhntap01] for host [r1pluhntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.b1pascntap01] for host [b1pascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1psumntap01] for host [r1psumntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.vnpascntap01] for host [vnpascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1pstmntap01] for host [r1pstmntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.cepascntap01] for host [cepascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.eppascntap01] for host [eppascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.uatinvntap01] for host [uatinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1pcrantap01] for host [r1pcrantap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1psncntap01] for host [r1psncntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.v1pascntap01] for host [v1pascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1ppahntap01] for host [r1ppahntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.impinvntap01] for host [impinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1paahntap01] for host [r1paahntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1pbwhntap01] for host [r1pbwhntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.cepinvntap01] for host [cepinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.vnpinvntap01] for host [vnpinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1ppkrntap01] for host [r1ppkrntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.eppinvntap01] for host [eppinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.v1pinvntap01] for host [v1pinvntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1ppenntap01] for host [r1ppenntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1pmmcntap01] for host [r1pmmcntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1plahntap01] for host [r1plahntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.r1pschntap01] for host [r1pschntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Calculated graphite_root [netapp-capacity.Clusters.g3pascntap01] for host [g3pascntap01]
[2017-04-20 11:58:39] [NORMAL ] [main] Using graphite_meta_metrics_root [netapp-capacity-poller.CENTURA]
[2017-04-20 11:58:39] [NORMAL ] [main] Startup complete. Polling for new data every [300] seconds.

Re: netapp-harvest ocum capacity metrics

Hi @RickStratton

 

 

The dashboards imported by Harvest assume you use Harvest to collect both performance and capacity information.  By default information is submitted for perf under netapp.perf.<site>.<cluster> and capacity under netapp.capacity.<site>.<cluster>.  To use these defaults your OCUM entry should look like:

 

[OCUM_CEPINVAP296]
hostname = cepinvap296.centura.org
site = CENTURA
host_type = OCUM
data_update_freq = 300
auth_type = password
username = netapp-harvest
password = XXXXXX

With a config like above the default dashboards should show information.

 

Your current config uses a customized config that aligns the Harvest submitted capacity metrics hiearchy (netapp-capacity.Clusters.*) with the OPM 'external data provider' submitted metrics (netapp-performance.Clusters.*).  Harvest does not provide any default dashboards that are aligned with the OPM hierarchy so you'd have to create your own Grafana dashboard panels if you go this route.

 

 

Hope this helps!

 

Cheers,
Chris Madden

Solution Architect - 3rd Platform - Systems Engineering NetApp EMEA (and author of Harvest)

Blog: It all begins with data

 

If this post resolved your issue, please help others by selecting ACCEPT AS SOLUTION or adding a KUDO or both!