1) Good. Wire speed. Congratulations to StarWind team.
2) In HA all writes are executed TWICE. First write is issued by hypervisor to HA node 1 and then HA node 1 submits the same write to HA node 2 and only after it's completion whole storage cluster reports status "OK" to hypervisor. So for the same amount of time we need to move doubled data.
This cannot happen PHYSICALLY (for single non-overlapped operation of course, see below). This means HA performance is 50% for non-HA performance. Add some time for sync, delays and re-buffering and we'll get 40-45% (exactly what you have).
Windows Monitor shows virtual 2 Gbps network usage adding two full-duplex pathes (1 Gbps for Tx and 1 Gbps for Rx). For read and for write (both directions). As with write most of the data goes only one direction you cannot see more then 50% of the network utilization.
Both numbers you see are MAXIMUM ones. Or close to them.
3) Good. You have properly configured cross-link doing wire speed. Congratulations to you as an IT engineer.
Couple of remarks however.
1) In a RL I/O operation is quite seldom executed alone. Most of time bunch of writes go at the same time. So performance is going to be higher then 50% as we have a pipeline loaded. So question raises here: what are you using to measure your write performance? If the tool issues non-overlapped I/O you'll have poor results. With something like ATTO Disk Benchmark or Intel I/O Meter you'll be OK (default pattern is 4 concurrent I/Os). As I've told modern OSes don't use write-wait-write-wait scenario, they do write-write-write-wait-wait-wait. So please give a try to referenced tools or do execute some real scenario.
2) In a RL reads and writes are combined. So network usage is going to be higher for mixed read/write scenario (say you have file server or SQL server configured inside VM). Please give a try to combined load as well.
3) You can dramatically boost even single non-overlapped I/O mode by using doubled-tripled 1 GbE links between HA nodes or installing 10 GbE connection between them. You'll get something like 70-75% of non-HA config for purely sequential writes in non-overlapped mode. Give it a try as well.
heitor_augusto wrote:Ok, let me simplify the things:
* ESXi 4.1 host connected a non-HA LUN (size: 2 GB, write-back cache: 4 GB) give me a performance of 117 MB/s in the VM.
* ESXi 4.1 host connected a HA LUN (size: 2 GB, write-back cache: 4 GB) give me a performance of 47 MB/s in the VM. The graph of the synchronization network in Resource Monitor of Windows 2008 does not exceed 50%.
* The test with the NTTtcp shows a performance of 939.078 Mbits in a synchronization network between storages.