Our setup: -Windows VMs running on Cisco B200 M4 blades -Storage is connected via multipath 8 GB/s fiber to Pure1 storage LUNs -Networking is connected through the Cisco UCS switches to the MS350 switch stack (8 members) via dual 10G fiber links, vSphere is setup to utilize multiple uplinks per virtual switch - The MS350 stack is doing layer 3 and the only reason traffic would need to hit our active/standby MX100's is if they need to talk to one of our other sites, or the internet. No traffic shaping is in place which would purposely limit per-client bandwidth. Other notes - Our Pure1 array is barely being hit. IOPS are in the 10k-20k range on average, latency under 2 ms, bandwidth usually under 100 MB/s. It's capable of 3 GB/s. I've been trying to figure out why file transfers between the servers are only getting up to around 150 MB/s. I have tried tweaking advanced network settings like tcp/ip offload, etc etc, made sure there are no mixed up MTU sizes anywhere, but I would think for a 10G network running on flash storage would be quicker than this. Just trying to figure out where the bottleneck is and where I might want to look next. Thanks!
... View more