I was wondering if there is a recommendation for DeDupe based volumes? I ran into an issue where testing on a 408GB disk, and created a 350GB StarWind DeDupe volume. I was hoping that it would use only up to 350GB and no more, so I'd have ~60GB free just in case. I see now that there is an .SPBITMAP, .SPDATA, and an .SPMETADATA, as well as a .DDDB, and I ended up with a drive full issue in Windows, which took my target offline. Test environment so no big deal, but I want to get a handle on what I might have done wrong or with incorrect assumptions before I retry.
I see from other posts and from the performance whitepaper, that 4K DeDupe works better than 64K DeDupe. This of course makes perfect sense, as it is far more likely to find identical 4K blocks than to hope to find blocks 16x the size without a single byte difference. It also makes sense that the block size should be the same throughout, to avoid a situation similar to MBR mis-aligning. So I would want 4KB blocks on the Windows 2008R2 Starwind host, and on my ESXi 5 boxes I would want 1MB VMFS 5 datastores, and inside my VM's, I would want 4KB blocks, correct?
My biggest concern is ensuring that the Starwind host volume doesn't fill to capacity and take the volume(s) offline.
Also in the performance white paper there is some indication that there is a way to measure the DeDupe size and ratio. Is this done by simply looking at the size of one of the 4 volumes above or is there a tool within the StarWind console I haven't found to tell me this? It looks like if you delete blocks (ie: a VM) it does not reclaim the space inside the VMFS LUN (or inside of StarWind), which is typical of thin provisioned disks and I understand that. But this would skew DeDupe calculations if one is simply looking at the high-water mark of the files on the StarWind host volume.
Just looking to make sure I'm doing this right, before I jump both feet into it.
Thanks!
The Latest Gartner® Magic Quadrant™Hyperconverged Infrastructure Software