Talk and ask questions about NetApp FAS and AFF series unified storage systems. Talk with other members how to optimize these powerful data storage systems.
Talk and ask questions about NetApp FAS and AFF series unified storage systems. Talk with other members how to optimize these powerful data storage systems.
Hi Team we have an issue on one of our netapp appliances. One of the controllers keeps faulting every 2 days. Only a hard reset seems to solve the issue temporarily. Some of the resources do stay online but the rest do not failover to the other controller. kindly Advise
... View more
Need some help with a FAS2520 2-node switchless cluster. Since some time one of the nodes appears in red (node down) in OnCommandSystemManager. Here's some commands output that I've managed to run (I'm a NetApp newbie, hope these make sense):
ntapcl-bul-sf::*> cluster show Node Health Eligibility Epsilon -------------------- ------- ------------ ------------ ntapcl-bul-sf-01 false true false ntapcl-bul-sf-02 true true false 2 entries were displayed.
ntapcl-bul-sf::*> storage failover show Takeover Node Partner Possible State Description -------------- -------------- -------- ------------------------------------- ntapcl-bul-sf-01 ntapcl-bul-sf- - Unknown 02 ntapcl-bul-sf-02 ntapcl-bul-sf- false In takeover, Auto giveback deferred 01 2 entries were displayed. ntapcl-bul-sf::*>
ntapcl-bul-sf::*> network port show Warning: Unable to list entries for vifmgr on node "ntapcl-bul-sf-01": RPC: Couldn't make connection. Node: ntapcl-bul-sf-02 Ignore Speed(Mbps) Health Health Port IPspace Broadcast Domain Link MTU Admin/Oper Status Status --------- ------------ ---------------- ---- ---- ----------- -------- ------ a0a Default Default up 1500 auto/1000 healthy false e0M Default Default up 1500 auto/1000 healthy false e0a Default - up 1500 auto/1000 healthy false e0b Default - up 1500 auto/1000 healthy false e0c Default Default down 1500 auto/10 - false e0d Cluster Cluster up 9000 auto/10000 healthy false e0e Default Default down 1500 auto/10 - false e0f Cluster Cluster up 9000 auto/10000 healthy false 8 entries were displayed.
ntapcl-bul-sf::*> network interface show Logical Status Network Current Current Is Vserver Interface Admin/Oper Address/Mask Node Port Home ----------- ---------- ---------- ------------------ ------------- ------- ---- Cluster ntapcl-bul-sf-01_clus1 up/- 169.254.49.169/16 ntapcl-bul-sf-01 e0f false ntapcl-bul-sf-01_clus2 up/- 169.254.163.32/16 ntapcl-bul-sf-01 e0f true ntapcl-bul-sf-02_clus1 up/up 169.254.87.210/16 ntapcl-bul-sf-02 e0d true ntapcl-bul-sf-02_clus2 up/up 169.254.190.147/16 ntapcl-bul-sf-02 e0f true SVM_AL SVM_AL-ISCSI-1 up/- 172.16.6.32/16 ntapcl-bul-sf-01 a0a true SVM_AL-ISCSI-2 up/up 172.16.6.33/16 ntapcl-bul-sf-02 a0a true SVM_AL-MGMT up/up 172.16.6.29/16 ntapcl-bul-sf-02 a0a true Logical Status Network Current Current Is Vserver Interface Admin/Oper Address/Mask Node Port Home ----------- ---------- ---------- ------------------ ------------- ------- ---- SVM_AL SVM_AL-NFS-1 up/up 172.16.6.30/16 ntapcl-bul-sf-02 a0a false SVM_AL-NFS-2 up/up 172.16.6.31/16 ntapcl-bul-sf-02 a0a true SVM_BL SVM_BL-ISCSI-1 up/- 172.16.6.27/16 ntapcl-bul-sf-01 a0a true SVM_BL-ISCSI-2 up/up 172.16.6.28/16 ntapcl-bul-sf-02 a0a true SVM_BL-MGMT up/up 172.16.6.24/16 ntapcl-bul-sf-02 a0a false SVM_BL-NFS-1 up/up 172.16.6.25/16 ntapcl-bul-sf-02 a0a false SVM_BL-NFS-2 up/up 172.16.6.26/16 ntapcl-bul-sf-02 a0a true ntapcl-bul-sf ntapcl-bul-sf-01_mgmt1 up/- 172.16.6.20/16 ntapcl-bul-sf-01 e0M true Logical Status Network Current Current Is Vserver Interface Admin/Oper Address/Mask Node Port Home ----------- ---------- ---------- ------------------ ------------- ------- ---- ntapcl-bul-sf ntapcl-bul-sf-02_cluster_mgmt up/up 172.16.6.19/16 ntapcl-bul-sf-02 e0M false ntapcl-bul-sf-02_mgmt1 up/up 172.16.6.21/16 ntapcl-bul-sf-02 e0M true 17 entries were displayed. ntapcl-bul-sf::*>
I've tried to reboot the failed node from command line, but:
ntapcl-bul-sf::*> system node reboot -node ntapcl-bul-sf-01 Warning: Rebooting or halting node "ntapcl-bul-sf-01" in an HA-enabled cluster may result in client disruption or data access failure. To ensure continuity of service, use the "storage failover takeover" command. Are you sure you want to reboot node "ntapcl-bul-sf-01"? {y|n}: y Warning: Unable to list entries on node ntapcl-bul-sf-01. RPC: Couldn't make connection Error: command failed: RPC: Couldn't make connection
Looks like a communication issue, there is no ping to any of the IPs of the failed node - not from a workstation, not from the second node.
What would be the right thing to do next? Is it safe to turn the power off/on , or reconnect all network cables one by one? Or something else?
Thanks in advance for the help!
... View more
Hello NA Community, I had to power off our old FAS2520 for a maintenance and I think all the SSDs of our flashpool hit the bug 1335350 where they failed after a power cycle and I'm pretty sure we exceeded 70,000 power on hours. The bug details says to contact support, unfortunately this is now an outdated system with an outdated ONTAP version (otherwise I wouldn't have had that issue), so I'm not even able to create a support case for that system anymore. Did someone here encounter this issue and was there a way to make the SSDs that exceed 70k power on hours work again? I assume updating the FAS2520 now wouldn't be useful as the system will be unable to flash a new firmware on the SDDs that are marked as failed/broken. Thank you in advance.
... View more
Hi, in my lab i'm using a single controller FAS2750 and i'm trying to use the e0a and e0b to connect them to a ethernet switch, but i don't get a link. When using the same cable and make a bridge from e0a to e0b i'm getting a link, but not when connecting to the switch. Same on the switch, when i connecting both ports i'm getting also a link. Switch and NetApp are showing all SFP informations of the connected SFP. I have already tryed different DAC cables and SFP modules with optical cables. The switch is a Mellanox MSN2010. Do i missed something? Kind regards Stefan
... View more
A300 have 4*DS224C(shelf ID 0,1,2,3) in single stack. Data aggregates are spread among all disks in this stack. During maintenance window, I plan to move the shelf id 3 to a new stack within the same A300, set new shelf ID as 10 and power on shelf and controller. Does it cause any impact to data? Is there any document or KB for this?
... View more