VM Performance during full sync
Posted: Wed Apr 06, 2011 2:10 pm
Dear all,
At present i have one test VM on our eventual, intended production VM cluster with HA 16tb starwind licenses.
I have starwind 5.6 configured as follows:
2x SAN machines (DL370's, G7, 12gb RAM, 500gb 7200 8x2.5" disks in raid10) 6 nics on each machine configured with jumbo frames and starwind recommended ip enhancements as follows:
-Storage NIC's x3 on 3 separate subnets (my plan is to provide different CSV's over different nics to avoid saturation - is this dumb?), all 1gb. no df gw or dns. VLAN500 with IP's in these ranges 192.168.240.0/27, 192.168.240.32/27, 192.168.240.64/27, the san storage NIC's correspond and are at the other end of these /27 ranges.
-Management NIC x1 on LAN (VLAN1), 1gb, 10.0.1.235/236
-Heartbeat NIC crossover to other SAN server, 1gb - No VLAN 192.168.241.3/4
-Sync NIC x1 on crossover to other SAN server, 1gb - No VLAN 192.168.242.3/4
2x Hyper-V R2 servers, 6 nics on each, configured with jumbo frames and starwind recommended ip enhancements as follows:
-VM switch on 2x adaptors, teamed. 1gb each, 2gb teamed - no ip.
-Storage NIC's, x2 on seperate subnets to correspond with first two storage NIC's on SAN servers, 1gb on VLAN500 (sit in same subnets as corresponding SAN NIC's)
-Heartbeat NIC x1 on crossover to other VM host, 1gb - No VLAN
-Management NIC x1, 1gb - (VLAN1) 10.0.1.250/251
All targets are configured as HA, the quorum has 64mb of write back and the 100gb csv has 1024mb of write back. Sync and hearbeat are configured. the windows initiator is configured with mpio for each target, all cluster tests pass with green, live migration works well, performance seems great under load.
I have no performance problems in terms of speed to/from SAN, i see 115mb constant on all the interfaces moving iscsi traffic under normal circumstances, seems to work great.
When i do a gentle shutdown of the san it sync's back again fine, there are no performance issues during fast sync. If there is a harsh shutdown, the VM's continue to run but any operation beyond a ping is next to impossible while they are connected during a full-sync. If we are to lose one side of the san we need things to continue to run during the slow-sync. My question is, what am i doing wrong? Clearly connectivity remains but it's too clogged up to be useful.
While the slow sync occurs the sync nic is 60% utilised (despite managing tcp wirespeed when not dealing with sync traffic).
I read elsewhere in the forum http://www.starwindsoftware.com/forums/ ... tml#p12694, that you should put the heartbeat on a vlan on the client connection nic, by this do you mean the management connection OR the iscsi connections that the VMHosts connect to? (see below quoted section from url)
Thanks,
Jim
At present i have one test VM on our eventual, intended production VM cluster with HA 16tb starwind licenses.
I have starwind 5.6 configured as follows:
2x SAN machines (DL370's, G7, 12gb RAM, 500gb 7200 8x2.5" disks in raid10) 6 nics on each machine configured with jumbo frames and starwind recommended ip enhancements as follows:
-Storage NIC's x3 on 3 separate subnets (my plan is to provide different CSV's over different nics to avoid saturation - is this dumb?), all 1gb. no df gw or dns. VLAN500 with IP's in these ranges 192.168.240.0/27, 192.168.240.32/27, 192.168.240.64/27, the san storage NIC's correspond and are at the other end of these /27 ranges.
-Management NIC x1 on LAN (VLAN1), 1gb, 10.0.1.235/236
-Heartbeat NIC crossover to other SAN server, 1gb - No VLAN 192.168.241.3/4
-Sync NIC x1 on crossover to other SAN server, 1gb - No VLAN 192.168.242.3/4
2x Hyper-V R2 servers, 6 nics on each, configured with jumbo frames and starwind recommended ip enhancements as follows:
-VM switch on 2x adaptors, teamed. 1gb each, 2gb teamed - no ip.
-Storage NIC's, x2 on seperate subnets to correspond with first two storage NIC's on SAN servers, 1gb on VLAN500 (sit in same subnets as corresponding SAN NIC's)
-Heartbeat NIC x1 on crossover to other VM host, 1gb - No VLAN
-Management NIC x1, 1gb - (VLAN1) 10.0.1.250/251
All targets are configured as HA, the quorum has 64mb of write back and the 100gb csv has 1024mb of write back. Sync and hearbeat are configured. the windows initiator is configured with mpio for each target, all cluster tests pass with green, live migration works well, performance seems great under load.
I have no performance problems in terms of speed to/from SAN, i see 115mb constant on all the interfaces moving iscsi traffic under normal circumstances, seems to work great.
When i do a gentle shutdown of the san it sync's back again fine, there are no performance issues during fast sync. If there is a harsh shutdown, the VM's continue to run but any operation beyond a ping is next to impossible while they are connected during a full-sync. If we are to lose one side of the san we need things to continue to run during the slow-sync. My question is, what am i doing wrong? Clearly connectivity remains but it's too clogged up to be useful.
While the slow sync occurs the sync nic is 60% utilised (despite managing tcp wirespeed when not dealing with sync traffic).
I read elsewhere in the forum http://www.starwindsoftware.com/forums/ ... tml#p12694, that you should put the heartbeat on a vlan on the client connection nic, by this do you mean the management connection OR the iscsi connections that the VMHosts connect to? (see below quoted section from url)
Any advice appreciated I've been banging my head against a wall on this for 4 days now. My only remaining issue is the VM performance during a full sync, i was expecting it to be viable for production use given everything else i'm reading so am keen to know what i've configured incorrectly.1. On each server's client connection NIC - deploy 2 VLANs resulting in 2 subnet IP separation (e.g 10.10.10.1 and 10.10.20.1)
10.1 will be used for heartbeat, 20.1 will be used for client connections.
Your config should now have 3 addresses to connect to e.g.
10.10.100.1 - Server1 Sync
10.10.20.1 - Server1 Client
10.10.10.1 - Server1 Heartbeat
10.10.100.2 - Server2 Sync
10.10.20.2 - Server2 Client
10.10.10.2 - Server2 Heartbeat
Thanks,
Jim