Page 1 of 1

V8.... Wow.... Feedback

Posted: Sat Nov 14, 2015 5:06 am
by Sienar
So I setup a small lab environment and have been attempting to test out V8. I started yesterday with build 8198 because I'd downloaded it several weeks ago and had only just got around to it. Seemed be doing ok, had a 5GB LUN replicated between the 2 nodes and a 2012 R2 cluster using that for quorum. I noticed today that build 8730 was available. Is this your current production, selling version of Virtual SAN? You guys are making giant changes between builds of the same version (V8). IE, you completely removed the concept of clusters. That seems like a drastic change that would go into V9, not just another spin of V8. Also, build 8730 seems to have broken device creation. I can no longer create large LSFS devices because it won't let you turn off dedupe and I don't have 1TB of ram in my test machines. Looks like I'm rolling back to 8198 because of that, but the fun there is that it's killed the replication for my little quorum device so I'm probably going to have to delete and recreate that. Good thing there isn't important data in there...

Re: V8.... Wow.... Feedback

Posted: Mon Nov 16, 2015 3:52 pm
by darklight
I believe deduplication is even not enabled by default. Give it a double-check and you should be fine.
lsfs.jpg
lsfs.jpg (41.56 KiB) Viewed 8261 times
Cluster concept was something I never used to try since it looks more like cosmetics/management thing rather. Basically, I always set up everything manually, it's just kind of feeling to know exactly how it's configured. Didn't even notice that it is not present anymore :)

Re: V8.... Wow.... Feedback

Posted: Wed Nov 18, 2015 12:09 am
by Sienar
It shows unchecked. But the behavior of the following screens in the wizard indicate that the checkbox is ignored. It won't let me create a 10TB, thin provisioned, LSFS LUN. Both builds I've tried have proven way to flaky. I've scrapped the test and will just stick with MS DFS-R with a manual failover using a secondary IP that I can just move between the file servers. Would've loved for virtual SAN to work though.

Re: V8.... Wow.... Feedback

Posted: Wed Nov 18, 2015 5:22 am
by oxyi
Am I reading it correctly? are you saying in the newer version of virtual SAN, there is no more clustering , like 2 nodes, 3 nodes?

Re: V8.... Wow.... Feedback

Posted: Wed Nov 18, 2015 10:53 am
by darklight
Sienar, take a look to LSFS technical description:
https://knowledgebase.starwindsoftware. ... scription/
Limits and requirements
Required free RAM (it is not related to L1 cache):
4.6 GB of RAM per 1 TB initial LSFS size (with deduplication disabled)
7.6 GB of RAM per 1 TB initial LSFS size (with deduplication enabled)
LSFS maximum size is 11 TB
Basically you need to have 46GB of RAM for 10TB LSFS device.

oxyi, yeah, that's correct. AFAIK they (temporarily) removed cluster option from management console for some redesigning/revamping.

Re: V8.... Wow.... Feedback

Posted: Thu Nov 19, 2015 12:55 am
by Sienar
That's pretty odd. It let me create them in the previous build on a system with only 32GB of ram. 46GB of RAM just to mount a measly 10TB is not production ready...

Re: V8.... Wow.... Feedback

Posted: Thu Nov 19, 2015 3:36 pm
by anton (staff)
It's not what he told. Clustering is there it's just management organized in different way. We're still working on the visualization here.
oxyi wrote:Am I reading it correctly? are you saying in the newer version of virtual SAN, there is no more clustering , like 2 nodes, 3 nodes?

Re: V8.... Wow.... Feedback

Posted: Thu Nov 19, 2015 4:00 pm
by anton (staff)
It really depends on how you look at the things and what's the competitor.

Bad news: Other guys have even higher memory requirements for logging and in-line dedupe. So Storage Spaces Direct now need ±10GB of RAM for 1TB of flash cache (not talking about actual storage!) and flash cache is mandatory with S2D.
ZFS needs ±200-300-400 bytes for every hash block below so it's ±1:20 ratio if we use 4KB blocks like we use. It's 100GB of RAM for 2TB of 4KB in-line deduplicated storage. We'll ask for 14GB only!

Good news: We're working on moving hash tables and index tables from RAM to flash. First will go index tables as they are same for dedupe and non-dedupe modes so we'll drop from 7GB per 1TB to 3.5GB and then to zero. First iteration is VERY close to being released.

We'll still keep it as an option as if you want to dedupe NVMe storage you'll be willing to play 1-2-3TB of flash per node and that means RAM-based index and hash tables are OK.

P.S. ZFS will do the same as well keeping hash tables in flash and not in RAM.

Please stay need.
Sienar wrote:That's pretty odd. It let me create them in the previous build on a system with only 32GB of ram. 46GB of RAM just to mount a measly 10TB is not production ready...