I've been researching this product for a couple weeks now due to the fact that it's one of the first hits on google when speaking of active/Active HA storage..
Im setup just like the following page running WS2012 hyper-v clustering with mirrored iscsi sans running the starwinds software:
http://highavailability.starwindsoftware.com/
IM experiencing very poor write performance when running throughput testing on a single VM with nothing else running on either the storage nodes or the hyper-v nodes. I'm using Atto and crystaldiskmark
The storage nodes have 2 iscsi nics a piece as well as 2nics dedicated to the sync channel. Both storage nodes run a 4drive RAID0 for testing purposes. All nics are 1gig nics and the Hyper-v nodes are using MPIO to connect to the storage. Given the equipment, the storage nodes can spit out 4gig/sec should it be asked to. Obviously the storage subsystem in a 4 drive raid0 is more than fast enough for testing. Both arrays do ~+400MB/sec read/writes.
The Hyper-v boxes are rather handicap though, boasting only a single 1gigabit nic for iscsi, and 1 Gbit nic a piece for VM traffic, csv/liv migration...and a management nic. So 4 nics total, traffic has been segmented off appropriately with no cluster or client traffic allowed over the iscsi channel.
Knowing this bottleneck, the theoretical max performance the vm can hope to achieve is ~100MB/s to around 125MB/s writes if your lucky and have a 100% efficient 1Gbe nic.
This theory holds true if on one of the storage nodes I shutdown both iscsi nics plus the starwinds sync channel. This forces all traffic to the other storage nodes and respective iscsi target. Again a single storage node has 2 multipathing nics so a target gets 2 data streams from 2 nics.
The problem is if I run both storagenodes full bore active/active and syncing, the write performance within the Vm dives Down to 70MB/sec. And this is with all 4 storage nics doing Round robin multipathing. If I take 1 nic from each storage node away, the performance gets even worse, I'm talking 50-60MB/second writes in this active/active SAN.
So where am I missing my mark? This is just a single VM pulling from a single lun with nothing on it. As raw as you can get. No bells and whistles, the lab was set up with simplicity in mind but the storage nodes were upgraded in both networking and storage subsystem due to the performance issues I'm seeing.
The 4drive raid0 is obviously not the issue, the network can handle a full 1gbit/s(handicapped by the switch that handled iscsi traffic, it could do 4gbit/s if I had 10gig switches laying around). My question is, is that if the active/active nature of the product is causing such a massive performance penalty due to having to simultaneously read/write and sync at the at the same time?
Again if I point the vm at a single storage node, with multipathing over dual nics, the VMs write, pin at 100MB/s, the limit of the switch and the 1gig nic dedicated to iscsi. I'm curious as to what's going on. I've watch both storage nodes iscsi multipathing nics, and it's working properly, with each nic showing equal traffic spread across all 4 iscsi nics(2 in each storage node)... Given the chance it will push 4Gbit/sec.
In this config doing active/active, raid0 and 4-Gbit nics doing MPIO, I'd expect the the VM to max out the network which it limited to 1Gbit/sec or ~ 100MB/sec as such, should not dip below 100MB/s in benchmarking.
Read performance is skewed due to csv caching and such, and those speeds are insane, it's the writes that I'm concerned about...
If this get solved we deploy and purchased the product, if not. Well, it is what it is. But I'd rather stick with this product due to having worked with it so much now.
Any help would be appreciated. By all means if something is obviously out of place here please bring it to my attention.
The Latest Gartner® Magic Quadrant™Hyperconverged Infrastructure Software