Hello,
I just want to ask community, is there anybody see similar behavior like my environment. And how you treat it?
I have lot of clusters on a datacenter. Some of the clusters on Cisco UCS, some on HP BL460 blades. Some of the clusters connected on NetApp SAN, some on EMC SAN.
All these devices connected on 4 or 6 EMC SAN switches. All devices in same physical data center.
Each cluster has about 15 - 25 hosts, and average 15VMs per host.
I always see weird storage latency on all farms. From vmkernel logs, it indicates it's LUN timeout, datastore heartbeat lost, performance degrading...etc.
It more like too many VMs running on shared storage lead to it cannot provide enough IOPS to VMs.
Do you see similar problem like mine? How you fix it?
My understand is that virtual disks share same datastore, datastore of a cluster reside on a physical LUN, the LUN shares same physical storage pool with other luns, the storage pool consist of lot of physical disks.
so if some virtual disks are busy, it will lead to the lun slow, then lead to storage pool slow, then impact other luns in same pool, then virtual disks on other luns will be slow down.
Is that correct?