2015-09-28 06:39 AM
On 2 of cDOT 2-nodes-clusters on that I am supposed to administer, I discovered TWO cluster-mgmt interfaces.
Does it make any sense according to Netapp recommendations ? I was pretty sure that only ONE cluster-mgmt per cluster was allowed.
Is there a reason to worry about side-effects ?
Solved! SEE THE SOLUTION
2015-09-28 06:57 AM - edited 2015-09-28 11:47 PM
Below, the output of "network interface show" for one of the 2 clusters.
Among side-effects, I already see one with OnCommand Unified Manager: even configured with the first cluster-mgmt address,after some time it takes into account the second (unreachable) address.
Logical Status Network Current Current Is
Vserver Interface Admin/Oper Address/Mask Node Port Home
----------- ---------- ---------- ------------------ ------------- ------- ----
clf3220ftv01_mgmt up/up 172.22.34.40/27 ndf3220ftv01 e0a true
clf3220ftv01_mgmt2203 up/up 10.199.203.2/24 ndf3220ftv01 e1b-2203 true
intcl_lif1 up/up 172.22.34.39/27 ndf3220ftv01 e0a true
ndf3220ftv01_clus1 up/up 169.254.123.87/16 ndf3220ftv01 e1a true
ndf3220ftv01_clus2 up/up 169.254.43.247/16 ndf3220ftv01 e2a true
ndf3220ftv01_mgt up/up 172.22.34.41/27 ndf3220ftv01 e0M true
ndf3220ftv01_snapmirror up/up 10.99.100.120/27 ndf3220ftv01 e2b-413 true
intcl_lif2 up/up 172.22.34.49/27 ndf3220ftv02 e0a true
ndf3220ftv02_clus1 up/up 169.254.243.218/16 ndf3220ftv02 e1a true
ndf3220ftv02_clus2 up/up 169.254.219.118/16 ndf3220ftv02 e2a true
ndf3220ftv02_mgt up/up 172.22.34.42/27 ndf3220ftv02 e0M true
ndf3220ftv02_snapmirror up/up 10.99.100.121/27 ndf3220ftv02 e1b-413 true
<EDIT: removing entries containing sensitive information>
57 entries were displayed.
2015-09-28 07:07 AM
Well, while it is certainly unusual, I do not see anything wrong in having multiple interfaces with cluster management role. Of course, there could be just one routing table for admin SVM, so you really can only access one of them from different subnet.
2015-10-04 10:04 PM
From an installation perspective, it depends on the installation engineer as to which ports are used for cluster-mgmt but typically the favored port is e0a on this particular controller instead of e0M. Also, port e0a is the node-mgmt port as well which is fine. I've also seen customers try and use an ifgrp (e0a,e0b) for cluster-mgmt but there's no reason to do this either because 'cluster-mgmt' only hosts system manager and SSH traffic and in the event of a port failure, it will move over to a different port anyway. Can you provide the output of the 'net port show' command if possible?