![]() ![]() The optimization policy specifies the files that should be considered for Data Deduplication. To preserve access semantics, Data Deduplication uses a file system filter (Dedup.sys) to redirect reads to optimized content completely transparently to the user or application that makes the read request.Ī file is considered optimized (or deduplicated) by Data Deduplication if it has been chunked, and its unique chunks have been stored in the chunk store. Data Deduplication is supported on NTFS formatted volumes.Ī file system filter is a plugin that modifies the default behavior of the file system. The file system is the software and on-disk data structure that the operating system uses to store files on storage media. This is the part of the file that Data Deduplication optimizes. The file stream is the main content of the file. ![]() For instance, Date Created, Last Read Date, Author, etc. The chunk store is an organized series of container files in the System Volume Information folder that Data Deduplication uses to uniquely store chunks.Īn abbreviation for Data Deduplication that's commonly used in PowerShell, Windows Server APIs and components, and the Windows Server community.Įvery file contains metadata that describes interesting properties about the file that are not related to the main content of the file. The Unoptimization job, which is a special job that should only be run manually, undoes the optimization done by deduplication and disables Data Deduplication for that volume.Ī chunk is a section of a file that has been selected by the Data Deduplication chunking algorithm as likely to occur in other, similar files. Additionally, Data Deduplication keeps backup copies of popular chunks when they are referenced more than 100 times in an area called the hotspot. When possible, Data Deduplication can automatically use volume features (such as mirror or parity on a Storage Spaces volume) to reconstruct the corrupted data. The Integrity Scrubbing job identifies corruption in the chunk store due to disk failures or bad sectors. The Garbage Collection job reclaims disk space by removing unnecessary chunks that are no longer being referenced by files that have been recently modified or deleted. The optimization process that Data Deduplication uses is described in detail in How does Data Deduplication work?. The Optimization job deduplicates by chunking data on a volume per the volume policy settings, (optionally) compressing those chunks, and storing chunks uniquely in the chunk store.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |