So we recently implemented a new ESX/iSCSI environment and we're seeing worse performance than we were in our multipathed 1gbe environment.
We have 3 x Dell R910 Servers with 2 x Intel Dual Port 82598EB 10gb CX4 NICs each.
Each of the servers are connected to a Force10 C300 switch. We are using 1 port on each NIC of each server for iSCSI traffic and each one is connected to a seperate Force10 C300. The C300's are then cross connected.
We have 2 x Dell Equallogic PS6510 arrays also cross connected to the two C300 chassis.
We have a seperate vSwitch configured on each host and are using a VMK for each physical iSCSI NIC. We are using jumbo frames and are able to vmkping a jumbo frame to the portal and each target address on the SANs with no problem.
When running performance tests in redhat using dbench we are only seeing around 550mb/s throughput either directly to the local vmdk disk which is stored on a round robin multipathed datastore or using a software iSCSI initiator within Redhat also multipathing.
We hooked up our old SANs for comparison and in an otherwise identical configuration we're seeing a consistent 630mb/s out of our ESX hosts utilizing 4 x 1gbe from the host to the switch and 3 x 1gbe to the PS5000.
So, with only 4 1gbe NICs on the host, and 3 1gbe NICs on an array with 32 less drives we are seeing an 80mb/s INCREASE in performance over our new 10gbe hardware... what gives?
When switching the path selection policy from Round Robin to Fixed on the datastore we see speeds drop to around 130mb/s
I feel like we must be missing some huge configuration step to see such terrible performance.
Ideas?