3 weeks ago - last edited 3 weeks ago
My name is Carlos and I'm a System Administrator at a payroll company in Canada.
Recently, we purchased a NetApp e-2824 to be connected to our 2 VMware servers in order to improve performance and availability.
However, when we finally finished the installation and started doing some testing we noticed that the performance was very low.
The SAN has 2 x controllers working in load balancing, and each of those has 2 x 10GB/s ports.
We basically connected that SAN to a 24 port x 1GB/s switch and connected our VMware servers 1GB/s ports to the same switch.
We also isolated this system in a specific VLAN with an IP addressing different from the other network connected to the same switch.
Below there is a summary of the connections:
SAN Controller A por e0a (10GB/s) <------------> Switch port (1GB/s) [VLAN 2] Switch port (1GB/s) <------------> VMware server 1 eth0
SAN Controller A por e0b (10GB/s) <------------> Switch port (1GB/s) [VLAN 2] Switch port (1GB/s) <------------> VMware server 1 eth1
SAN Controller B por e0a (10GB/s) <------------> Switch port (1GB/s) [VLAN 2] Switch port (1GB/s) <------------> VMware server 2 eth0
SAN Controller B por e0b (10GB/s) <------------> Switch port (1GB/s) [VLAN 2] Switch port (1GB/s) <------------> VMware server 2 eth1
We're using iSCSI and Jumbo Frames (in all equipment).
I believe the performance is low because we're using 1GB/s links among all equipment, but I'm not familiar with storage installations neither with iSCSI.
Anybody knows if there is any way to make this system perform good with the current hardware?
Or we'll have to upgrade all switch and server's NIC's to 10GB/s?
3 weeks ago
You are correct - since the switch only does 1Gbit connectivity, the devices, including E-Series controller, will only do 1Gbit over any connection, meaning at most ~90MByte/sec speeds. If throughput is much below 70Mbyte/sec, you probably need to check and doublecheck all of your jumbo frame settings.
If you connect multiple connections on the vmware servers, you can use MPIO to balance IO somewhat over multiple connections, but it won't produce significant levels of improved throughput. For that, you would need a 10Gbit switch and cards for each of your vmware servers
Hope this helps!
3 weeks ago
Thanks for your answer.
Initially, we've installed the SAN directly connected to the servers, but we've been told that this kind of installation wouldn't work.
It's actually worked, but then we decided to connect those to one of our switchs to see if the performance would be better.
The performance stayed the same on both architectures, but if we go with the second one we'll need to upgrade the switch as well.
My plan would be to install 10GB/s ethernet modules on the servers and roll back the installation to connect SAN and servers directly.
Anybody knows if doing that I'll be able to get those connected at 10GB/s? (I think it should work)
Then, if that works, we'll setup a project to add 2 x 10GB/s switches (for HA) for the next 6 months in order to make the installation better.
3 weeks ago
On the E-Series of controllers, we do support Direct Connect for iSCSI for vmware, meaning you don't need a switch, but having one does allow for better resiliency and multipathing in your storage network. To get 10Gb connectivity, you will need appropriate cables (optical, twinax or 10GBaseT), supported by both controller and server, and ideally any future switch you get.
Hope this helps!
3 weeks ago
Just to check if I understood it correctly.
If I want connect the SAN directly to the VMware servers, I'll need to add a pair of 10GB/s ports to each of those servers and also buy the specific cables to interconnect them, right?
2 weeks ago
When you say performance is low, what do you mean?
If you mean throughput is pegged at less than 100MB/s then yes, the switch you are using is the bottlneck. You could confirm this in your switch web interface or cli and checking port stats.
However if you are seeing high latency then there could be a number of reasons for the issue:
Check and recheck jumbo frame config!
If you are seeing high read latency only then you may want to disable delayed ack on you esxi hosts:
You can check the throughput you are geting with a tool like Atto disk benchmark in one of your VMs. This will generate lots of storage traffic which can be handy for finding issues and also check if changes you have made have had an impact.
2 weeks ago - last edited 2 weeks ago
Regarding performance, I meant the experience in general.
Our servers have a localstorage and I presented the storage to those.
When I turn on a VM hosted in the localstorage it takes maybe 20 seconds to start.
When I do the same on a VM hosted in the SAN, it takes probably more than a minute.
I configured jumbo frames on all devices to optimize performance, but I think it's a matter of connection speed among all devices.
2 weeks ago
That sounds more like a latency issue than a throughput issue as booting a server is unlikely to be maxing out a 1GB link.
Take a look at the delayed ack config.
Beyond that, I think you need to get familiar with the stats counters in vmware rather than relying solely on observing system performance.
Log in to vcenter, browse to your VM>monitor>perfromance then select disk from the drop down.
By default, this shows you the latency and throughput of your storage.
For flash storage, even over a 1GB link, you shouldn't be seeing much over 1ms in latency. For disk storage with no flash caching, expect to see not too much difference from your internal storage on the host (~20ms if busy, <10ms ideally)
I suspect that you will be seeing much higher values (ie. disable delayed ack)
Handy hint: vmware stats are accurate but the interface is slow and clunky. Get Veeam One (free), install on a windows box and point it at your virtual infrastructure. It will scoop up all your stats and give you a nice snappy interface to view up to 1 week of data. There are also a host of alarms you can confugure if you don't already have monitoring in place.