Hi,
Firstly, to explain my setup. Server 2012 R2 hyper-V hosts in a cluster (2 node) using a CSV NTFS.
A Guest Cluster for File Services using 2012 R2 and a shared VHDX, would like to use ReFS.
Current File Server is a single VM 2008 R2, data stored on NTFS (VHDX)
I am migrating a file server (approx. 1TB of data) but I want my new FS to be nice and resilient and highly available, I have a few questions which I am not too clear on the answers for and was wondering if anyone can shed some light on it. So far I have determined that deduplication is not supported on ReFS fine, data integrity is my top priority Id only save about 10% anyway.
I have also read that DFSR is not supported but I am not clear on whether this is replication TO the ReFS volume, or FROM the ReFS volume or both my existing file system is NTFS. I am unsure how to move 1TB of data retaining all ACLS and file structure from NTFS to ReFS, my idea was to use DFSR and once data is replicated I kill the original connection and switch it over out of hours, but now I am not sure. Robocopy is out of the question for me because the rate of data change is too frequent at the source, the number of files is just ridiculous to trawl through, and the time it will take is far too long. Im assuming an image based backup of a volume will restore the file system, Im thinking Windows Server Backup might work just to restore the root folder share, but Ive not gone from one file system to another using that. I also do want to use DFS as a namespace for connecting to the folder share, instead of using \\servername\sharename I want to use \\DFSNamespace\Sharename which will be used in mapped drives on the clients will either replication or namespace work on ReFS or neither, the technet documentation isnt too clear to me.
Secondly, I understand that having Checksums on files proves something about the file being integral. Is that right? If thats the case, is this something that ReFS does an equivalent thing for, or am I getting myself confused between these two things. If checksums would be useful for certain files even on ReFS, how would I generate these and use them when opening and saving certain files (there are a subset of files such as finance and design files which are critical to our business, we really want to make sure these files are not corrupted).
Finally, an idea to make a CSV or cluster available storage more resilient to failures or corruption:
I will be using both a General File server and a SOFS for application data. General File server uses an available storage disk in the cluster and is attached the general file server role, whereas SOFS uses CSV, these will be based on shared VHDX files. In both cases I am wondering how to make the underlying VHDX more resilient from corruption.
I have read that one way ReFS corrects corrupted blocks of data is that it will get clean blocks from a mirrored disk in storage spaces, so my question is this (and is it technically possible, and a supported configuration?)
Can I attach two shared VHDX files for the available storage in the cluster, create a mirrored storage space and add this space to the cluster instead of a single disk will this work?
Can I do the same above, but then add that mirrored space as a CSV.
My process would be this:
Add shared VHDX to both VMs
Using server manager: Create a storage space
Create a mirrored disk within the space
Create a volume
Format it as ReFS
Add it to the cluster as available storage
Add it is as CSV (for the SOFS scenario)
Will this work?
I have also noticed there is an area in Failover Cluster Manager to create and add storage pools, this is a feature I have never been able to use before because of not having the correct disks is this what I am after to do the above?
Many thanks in anticipation
Steve