If the controllers aren't getting populated then there could be some sort of communication error. Check the event logs -> AppWatch - this will show any errors if any. If its clean, try the following. 1. close SCOM 2. go to Start -> Run -> and enter Microsoft.MOM.UI.Console.exe /ClearCache (This will clear out anything that has been cached in SCOM) 3. Open a console window to a specific storage array. (SSH, Telnet) 4. Run Discovery Wizard again for a specific controller (Specify Ip Address for controller above) 5. We should be able to see any errors if any on the console. 6. Try to run the Manage Storage Credentials from SCOM -> Monitoring -> DataOnTap Folder-> Controllers
... View more
Doesn't look like the controllers were discovered. Did you go through the "discovery wizard" to pick the controllers up via snmp? You'll need to do this first for anything to show up.
... View more
It may be a good idea to open a case for this so we can start getting the information logged. If it turns out to be a bug then we'll have all the info ready for the dev team to take a look. Thanks!
... View more
Hi WH, Here is the bit from the AW2.1.1 BPG which will be released soon. This is the bare minimum for AppWatch to have basic monitoring functionality. Any advanced features such as PRO, Cloning, etc will not be covered with these roles. 1.1 BEST PRACTICES for NETAPP STORAGE MINIMAL ACCESS CONTROL In some IT environments, a detailed assignment of the minimal permissions is required. Table 3 describes the capabilities that are needed to connect to the storage system from ApplianceWatch PRO and gather monitoring data by using a local account on the storage system. This set of capabilities is purely for monitoring ApplianceWatch PRO basic functions and does not include any of the advanced features. This local Data ONTAP account will need to be assigned a customized role and contain the following capabilities. Note: These are the minimum requirements for basic monitoring only and do not contain any active management, cmdlets , or SCVMM PRO functionality. Table 3) Minimum capabilities for NetApp storage users for monitoring with ApplianceWatch PRO . NetApp Storage Capabilities login-http-admin api-system-get-version api-system-get-info api-system-get-vendor-info api-cf-status api-system-get-ontapi-version api-vfiler-list-info api-ems-autosupport-log api-aggr-list-info api-volume-list-info api-lun-list-info api-disk-list-info api-storage-shelf-list-info api-license-list-info api-lun-map-list-info api-volume-autosize-get api-aggr-options-list-info api-qtree-list, api-storage-shelf-environment-list-info api-lun-get-space-reservation-info api-volume-options-list-info api-perf-object-get-instances api-snmp-get api-snapmirror-get-status Example: Sample command to add/modify a custom role. useradmin role modify scom-user-roles -a login-http-admin,api-system-get-version,api-system-get-info,api-system-get-vendor-info,api-cf-status,api-system-get-ontapi-version,api-vfiler-list-info,api-ems-autosupport-log,api-aggr-list-info,api-volume-list-info,api-lun-list-info,api-disk-list-info,api-storage-shelf-list-info,api-license-list-info,api-lun-map-list-info,api-volume-autosize-get,api-aggr-options-list-info,api-qtree-list,api-storage-shelf-environment-list-info,api-lun-get-space-reservation-info,api-volume-options-list-info,api-perf-object-get-instances,api-snmp-get,api-snapmirror-get-status
... View more
Have you tried to map the problem LUN to another server? What type of data is on the LUN? If you take a snapshot and map it and the problem still exists I would say that the data on the LUN may be the issue here. Manual copy would be slow but you could run it over night via ndmp copy, rsync, or other copy tool from the local nodes to a freshly mapped lun. or The snapshot route, take a snapshot with SnapDrive, and map to server or other server and see if problem exists.
... View more
If you've changed the passwords since installation, you may need to recheck and reenter the passwords in SCOM. Go to the action account and reenter the credentials. Run manual discovery of AppWatch from discovered inventory.
... View more
Can you go to administration tab and click "connected management groups" Can you connect? credentials ok here? EDIT: Also can you check the credentials for Action Account under SCOM Administration -> Run As configuration -> Accounts
... View more
Hi Daniel, SDW 6.3 was just release and is still in "First Customer Shipped" status as opposed to 6.2.1 which is official "GA" . There would be no reason to upgrade to 6.3 p1 unless you were looking for a specific burt fix or functionality. Major difference between 6.3 would be the integration with VSC (aka SMVI) , SMSQL, and NFS datastores. SnapDrive 6.3 for Windows includes the following major new and enhanced features: Support for performing Snapshot copy operations on virtual machine disks (VMDKs) in an NFS or VMFS datastore, which requires Virtual Storage Console 2.0.1 or later. Please check the availability of Virtual Storage Console 2.0.1 and install it before using this feature. Support for space reclamation when using VMDKs in an NFS datastore Support for DataFabric Manager caching with RBAC Automatic download of the storage system access control (storacl.exe) tool as part of your SnapDrive 6.3 for Windows installation Bug fixes For more information, see the SnapDrive 6.3 for Windows Release Notes. If you need any of this then you would need to upgrade, otherwise I would stick with 6.2.1 for the burt fixes available.
... View more
http://now.netapp.com/NOW/download/software/snapdrive_win/6.2.1/ SnapDrive® 6.2.1 for Windows® - Description Features of This Release SnapDrive 6.2.1 for Windows is an update to SnapDrive 6.2 and includes fixes for the following bugs: Bug ID 393518: SnapDrive 6.2 fails to create RDM LUNs with ESX iSCSI when FCP license is not set on the filer. For more information, go to Bug Tools on the NOW site at http://now.netapp.com and search for bug ID 393518. Bug ID 395057: SMHV does not work on Japanese Windows OS Bug ID 397424: SDW: LUN restore fails IOCTL_DISK_SET_DRIVE_LAYOUT_EX on reversed SnapMirror destination Bug ID 397462: Windows 2008R2 120s Volume Arrival Timeout when connecting to LUN Clone on the same host Bug ID 397866: Snapshot mounts are very slow. Bug ID 408511: ESX iSCSI RDM igroup has initiators from other hosts. SnapDrive does not support this configuration. Bug 424071: SDW Workaround: SME remote verify fails on SM dst with SDW ZAPI snap-list-info response volume offline Bug 426196: SnapDrive for Windows creates multiple FlexClone volumes for LUNs on same Snapshot copy Bug 440822: The SnapDrive service in an ESX VM takes too long to start up in large ESX environments SnapDrive 6.2.1 for Windows continues to support the following major new and enhanced features introduced in SnapDrive 6.2 for Windows: Support for dynamic addition and removal of Hyper-V®; pass-through disks using Windows Server 2008 R2 Support for Microsoft® cluster shared volumes using Windows Server 2008 R2 Support for restoring data at the file level Support for RBAC with Operations Manager Support for RDM LUNs using iSCSI hardware initiators and VMware® ESX iSCSI initiators Support for creating FC RDM LUNs across two virtual Microsoft cluster machines using Windows Server 2003 and Windows Server 2008 systems Support for IPv6 with ESX 4.0 Bug fixes Limitations SnapDrive 6.2.1 currently has the following limitations: iSCSI RDM LUNs are not supported across two physical Microsoft cluster machines. Customers using SnapDrive 5.0 and earlier on Windows Server 2003 IA64 systems must upgrade to Windows Server 2008 on IA64 systems before they can use SnapDrive 6.2.1. A LUN managed by SnapDrive cannot be configured as a "dynamic" disk (a storage device that is divided into volumes rather than partitions); it can serve only as a "basic" disk (a storage device for host-side application data). A LUN cannot be configured as an extended partition; SnapDrive supports only a single, primary partition on a LUN. A LUN created with FilerView® or at the storage system command line can be managed only if certain steps (documented in the SnapDrive for Windows Installation and Administration Guide) are taken to prepare these disks for SnapDrive. Conversion of a LUN with an MBR partition to a GPT partition is not supported. Permanent clones are not supported. Snapshot™ copies that are created from the storage system console are not supported; because this can lead to inconsistencies within the NTFS file system, only SnapDrive should be used to create Snapshot copies of LUNs. Volume-level SnapVault® configurations are not supported (however, qtree SnapVault configurations are supported). Considerations when using Microsoft iSCSI Software Initiator include the following: In an MSCS configuration that uses volume mountpoints, install Microsoft hotfix 925677 on all cluster nodes if you encounter the issue described in Knowledge Base article KB925677. In a noncluster configuration, the Exchange® service and the SQL service must be configured with a dependency on the iSCSI Software Initiator service.
... View more
AppWatch does alerts for Thin Provisioned Volume Autogrow for SCVMM/Hyper-V environments - any Hyper-V VM on NetApp Storage monitored by AppWatch will generate this alert and can automatically be alerted via email,sms, etc with the "custom subscription" function within SCOM. For Non-SCVMM/Hyper-V environments you'll need to use the SNMP traps like Adai suggested.
... View more
You should be able to install with that but keep in mind that SDW won't see anything until VSC 2.0.1 is installed. You've got the setup correct but VSC is the communicator to the VMware bits so SDW won't be aware of any disks until VSC is installed and configured.
... View more
Just to clarify, SDW 6.3 will be able to identify NFS datastores for installation on VMDKs with VSC 2.0.1. You won't be able to install on a NFS share. As for integration , with this solution (SMSQL,VSC,SDW) you'll be able to quiesce your SQL database and truncate logs just like SMSQL currently works on physical hosts. To have it all in single pane of glass is something we are working towards and will start to see more and more in the future.
... View more
David, This is a question that is being looked at internally and our SQL experts should address this in a TR or BPG soon. I will see if we can get some info posted on this thread.
... View more
Sure not a problem. The reference I noted had the exact error and mentioned that the VMkernel was sending a command "nmp_CompleteCommandForPath: Command 0x3c" which translates to a "read buffer" which isn't supported in OnTap. Not sure why you're getting it in w2k3 and not w2k8 but there has been recent activity that had issues with w2k3 & ESX 4.1 Perhaps a difference in drivers used? Again not sure if its related but worth a read if you're on 4.1 w/ w2k3. -> http://now.netapp.com/NOW/cgi-bin/bol?Type=Detail&Display=447078 Also, I'd recommend opening a NGS case if its causing problems in your environment. They'll have direct access to escalations if needed.
... View more
Did some searching and found the following on a internal Burt which had some good info. It looks like its a command that OnTap doesn't support and the error is produced. 0x3C Read Buffer 0x05/0x20/0x00 OnTap does not have any buffer for them to read. We have no plans to have a buffer to ever support this command.
... View more
I would suggest opening a support ticket so NGS can help you troubleshoot. The "failed to enumerate" does sound like a dns or communication issue as many have noted here. Try the following. 1. Check DNS resolution on the host to storage and storage to host. As you have 4 hyper-v nodes, ensure that all ip addresses are properly resolved. 2. Check the snapdrive preferred ip address setting and ensure you have the correct management ip on the arrays. 3. Open a console to the storage array and try to create a disk. You should be able to see any permission errors on the console if it happens to be a permissions problem.
... View more
Just a note. Amrita posted on a seperate thread that VSC 2.0.1 would be shipped around the end of the month so may need to wait just a bit longer. http://communities.netapp.com/message/40797#40797
... View more
Unfortunately you cannot telnet into the node that has been taken over. Its essentially in either "waiting for giveback" mode or its completely down. If you have access to MYautosupport you should be able to take a look at all the logs avaialble for both nodes. Again , I would strongly recommend getting NGS involved if you haven't done so already. Good luck!
... View more
I would open a support ticket with NGS to ensure the failed node is ok. You don't want to try doing a giveback and cause other problems without diagnosing the failed node. If it did a failover there is probably a good reason why it did so would strongly recommend having NGS take a look at the logs to be sure you don't have any failed loops, h/w, etc.
... View more