I’ve been going back and forth with this issue for some time but honestly I have no idea if the vCenter telemetry is something to rely on. I’m experiencing rather high latency on the storage on my VMs, most of them idle, only vCenter and virtual firewall generate some IOPS, 5 are shut down, other 3 VMs are linux machines that idle for 99%, even though they can spike 100ms per IO. Today I have decided to migrate a VM storage to another server to find that higher disk utilization reduces the latency on the host, how that makes any sense? I’m using P420 in RAID 10 with 4x4TB 7k SAS HDDs.
Host latency:
If it’s constantly reading the same data it’s stored in cache, which is significantly faster than reading from the actual drive. Because the latency is average and cache is very fast it lowers the latency shown in that graph.
15-30ms latency seems reasonable for that hardware (P420 in RAID 10 with 4x4TB 7k SAS HDDs.)
Basically your single SAS disk can do ~150IOPS random, so worst case of 4k random reads you will get ~600KBps.
For the migration if it’s sequential (depending on filesystem layout), the performance can be different, up to maximum streaming performance of like 100MBs for large sequential reads.
Then 2x for your RAID config.
I have an r710 with 8 WD black 500gb disk in a raid 5, No vcenter. From a cold boot of any flavor VM there is severe disk latency.
Only way to get past it in my case is a large disk operation, is xfsdump of / solves the latency.
I had the same issue, only using a 930-8i w/ 2M cache. Honestly performance sucked on all of my VMs. I reinstalled the server with Rocky Linux 8.8 and KVM using the same array and performance was acceptable (the array was configured as a LVM volume). I then added a NVMe drive as a LVM cache and performance was much better (good enough for my homelab). Too bad, since I really prefer VMware.