Fact-checked by Grok 2 weeks ago

Data scrubbing

Data scrubbing is an correction that uses a background task to periodically inspect main or for errors, then corrects them using redundant such as checksums, bits, or mirrored copies. This process helps prevent silent , such as , by detecting and repairing issues before they accumulate into uncorrectable errors, ensuring long-term in systems like arrays and file systems. The primary purpose of data scrubbing is to maintain reliability in storage environments where media degradation or transmission errors can occur undetected. It is commonly implemented in redundant storage systems, including RAID configurations, modern file systems like ZFS and Btrfs, and hardware such as ECC memory and FPGAs, reducing the mean time to data loss. By proactively verifying data against redundancy mechanisms, scrubbing enhances fault tolerance without interrupting normal operations, though it may increase temporary I/O load during execution.

Fundamentals

Definition

Data scrubbing is a background process in computing systems that periodically inspects data stored in memory or on storage devices for errors by reading the data and verifying its integrity using redundant information, such as checksums or parity bits. This technique leverages error-correcting codes () to identify and, where possible, correct discrepancies without interrupting normal operations. A core aspect of data scrubbing is its proactive approach to detecting silent data corruption, where errors like bit flips or media degradation occur undetected and could accumulate over time into uncorrectable failures if left unaddressed. By systematically scanning or during idle periods, scrubbing ensures that such latent errors are identified early, allowing redundant mechanisms to reconstruct accurate data before they propagate. In contrast to data cleaning processes in databases, which focus on correcting semantic inconsistencies, duplicates, or formatting issues in datasets during extract-transform-load (ETL) workflows, data scrubbing specifically targets low-level and to combat physical degradation. The practice emerged in the early 2000s amid advancements in redundant storage architectures, designed to mitigate issues like and silent failures in large-scale archival systems.

Purpose and Benefits

Data scrubbing serves as a critical mechanism to mitigate the risks of in storage systems, particularly arising from hardware failures such as latent sector errors (LSEs) and silent during periods of inactivity or long-term archival. By proactively scanning and verifying using redundancy mechanisms like or error-correcting codes, scrubbing identifies and repairs these issues before they escalate into unrecoverable losses, addressing vulnerabilities from factors including and infrequent access patterns in large-scale environments. The primary benefits of data scrubbing include reduced system downtime through the timely correction of correctable errors, preventing them from compounding into multi-bit uncorrectable that could necessitate extensive efforts. This proactive approach significantly enhances overall reliability in mission-critical applications, such as servers and archives, where is paramount, by extending the mean time between (MTBF) and minimizing the impact of correlated disk errors. In RAID configurations, for instance, scrubbing ensures that single-sector issues are resolved prior to a disk , thereby averting complete reconstruction and associated operational disruptions. In large-scale systems, the quantifiable impact of scrubbing is evident in its ability to detect rare but critical s, with uncorrectable bit rates typically around 1 in 10^{14} bits read, allowing systems to identify potential failures annually in petabyte-scale deployments and prevent events that could otherwise occur multiple times per century without intervention. Furthermore, in modern environments, scrubbing contributes to cost savings by reducing the frequency and complexity of post-corruption recovery operations, optimizing resource utilization in distributed infrastructures where SSD retention s are prevalent.

Principles

In the context of storage systems, the principles of data scrubbing focus on maintaining through systematic handling.

Error Detection

detection forms a critical component of the data scrubbing process, enabling the of silent data corruptions, latent sector s, and bit flips in systems without . techniques leverage mathematical s to compute and compare signatures of data blocks, flagging inconsistencies that indicate . Cyclic redundancy checks () are widely employed due to their ability to detect burst s up to the length of the with high probability; for instance, a 32-bit can achieve a of 6 for data lengths up to 16,360 bits, providing robust protection against random bit flips common in . Checksums, such as Fletcher's , offer computationally efficient alternatives by iteratively summing data bytes in two running totals, detecting all single-bit s and most multi-bit s within the checksum length, though they are less effective against certain burst patterns compared to . Similarly, Adler-32, a variant of Fletcher's checksum, enhances performance for longer data streams by using modulo-65521 arithmetic, making it suitable for verifying during periodic scans in resource-constrained environments. Hash functions, including cryptographic ones like SHA-256, provide stronger for larger datasets, ensuring that even subtle alterations are detected with negligible false positives. In systems with , bits enable block-level detection by appending a bit that maintains overall even or odd across the data. This simple yet effective method computes the as the (XOR) of all data bits, allowing detection of odd-numbered bit during reads. P = D_1 \oplus D_2 \oplus \cdots \oplus D_n where P is the parity bit and D_i are the individual data bits. If the recomputed mismatches the stored value, an is flagged, though alone cannot pinpoint the exact location and misses even-numbered . These techniques are applied during background operations to minimize performance impact while ensuring data fidelity over time. Scanning approaches in data scrubbing dictate how storage media is traversed to apply these detection methods. Sequential scrubbing reads the entire dataset in logical block order, verifying each sector using , checksums, or during system periods to catch latent in cold data. This method maximizes coverage but can introduce if scrubbing rates exceed available time; optimal rates, such as 20 /hour, balance detection speed with foreground workload interference. Targeted or hot-spot monitoring, in contrast, prioritizes regions with higher risk—such as aging disk areas or those with prior latent sector —by partitioning into segments and sampling adaptively, often using staggered patterns across multiple regions to exploit spatial locality. Staggered scrubbing, for example, divides disks into 128 or more regions and scrubs corresponding segments in rounds, reducing the mean time to detect clustered by up to 40% compared to pure sequential methods while maintaining low overhead (around 2% with 1 MB segments). These approaches ensure comprehensive identification without exhaustive full scans. Post-2020 developments have integrated for enhanced in enterprise storage, complementing traditional techniques by analyzing access patterns, error histories, and metadata to predict and flag potential integrity issues proactively. Multi-tiered models, such as autoencoders and forests, identify outliers in logs that signal silent corruptions, improving detection accuracy in intelligent storage arrays by up to 25% over rule-based methods alone. Upon error detection, these mechanisms inform subsequent correction efforts to restore .

Error Correction

Error correction in data scrubbing involves repairing detected errors by leveraging built-in redundancy to restore without relying on external backups. Common mechanisms include from blocks in redundant arrays or from mirrored copies in duplication-based systems. In -based systems, such as those using XOR operations across blocks, corrupted is recovered by recalculating the original value from the remaining healthy blocks and the existing information. For mirrored setups, correction simply replaces the erroneous block with an identical copy from the redundant mirror, ensuring immediate availability of accurate . In error-correcting code (ECC) memory, decoding identifies and flips single-bit errors by computing a value from checks embedded in the . The correction process typically follows these steps: first, the affected block is isolated to prevent further reads or writes that could propagate the error; second, the correct data is recomputed using from healthy replicas, such as or mirrors; finally, the repaired data is rewritten to the original location or a new one, with verification checksums applied to confirm integrity. This sequence minimizes disruption, as scrubbing operates in the background, but care is taken to avoid "parity pollution," where uncorrected errors inadvertently corrupt during recomputation. Advanced techniques enable online correction without system downtime, particularly through (CoW) mechanisms that ensure updates. In CoW systems, modifications create new block copies while preserving originals until verification, allowing seamless repair of corruptions during active operations by redirecting pointers to corrected versions post-recomputation. This approach maintains consistency even amid concurrent access, reducing the risk of partial failures. A foundational example of correction is the , which corrects single-bit errors in memory. The S is calculated as S = H \cdot r \pmod{2}, where H is the parity-check matrix and r is the received codeword vector (equivalent to H \cdot E \pmod{2} with E as the error vector, since valid codewords yield zero syndrome). The resulting S gives the position of the error bit, which is then flipped to correct it. For a (7,4) with parity-check : H = \begin{pmatrix} 1 & 0 & 1 & 0 & 1 & 0 & 1 \\ 0 & 1 & 1 & 0 & 0 & 1 & 1 \\ 0 & 0 & 0 & 1 & 1 & 1 & 1 \end{pmatrix} Suppose the original codeword is the all-zero word (0,0,0,0,0,0,0) and there is an error in the third bit, yielding received word r = (0,0,1,0,0,0,0). Computing S = H \cdot r^T \pmod{2} yields S = (1,1,0)^T, interpreted as 011 (with the first component as LSB), or 3, indicating the error in bit 3. Flipping bit 3 corrects the word back to all zeros. As of 2020, research has explored models like autoencoders for to predict SSD failures, enabling preemptive correction of latent errors in NAND flash and improving reliability.

Storage Applications

RAID

In RAID configurations, data scrubbing involves periodic full-array reads to verify the consistency of and information across all disks, identifying and correcting silent or bit errors before they lead to failures during reconstruction. This process also detects and remaps defective sectors on individual drives, enhancing overall array reliability by proactively addressing issues like media errors or parity mismatches without interrupting normal operations. Data scrubbing primarily applies to redundant RAID levels, such as 5 and 6, where parity-based mechanisms allow for during the read-verify cycle. In 1 and 10, scrubbing focuses on consistency by comparing data across mirrored pairs to resolve discrepancies. Common implementations include PowerEdge servers' Patrol Read feature, which has provided automated background scrubbing for arrays since the early 2000s via PERC controllers, scanning for and repairing potential disk errors continuously or on schedule. In environments, the MD subsystem supports scrubbing through tools, often automated via jobs for weekly or monthly checks since kernel version 2.6. Scrubbing is typically scheduled monthly to balance error detection with minimal performance impact, involving logging of detected bit errors in parity-based arrays before triggering a full rebuild. As of 2025, NVMe RAID controllers, such as Broadcom's 9600 series, extend scrubbing support to SSD-based arrays, incorporating offload technologies like KIOXIA's RAID Offload to efficiently verify data integrity without excessive host CPU overhead.

File Systems

In file systems, data scrubbing serves as a background to ensure the integrity of and file blocks by leveraging built-in checksums, thereby detecting potential corruption caused by , hardware faults, or silent data errors. This process is particularly vital in environments where is stored long-term on disk arrays, operating atop underlying storage layers such as to validate logical structures without relying solely on physical . The general scrubbing process in file systems involves systematically reading all allocated blocks and metadata, computing or verifying checksums against stored values, and initiating repairs where possible using redundancy or backups, all while the file system remains mounted and operational. This online approach minimizes disruption, often integrating with volume managers like LVM to snapshot volumes temporarily for safe verification without interrupting user access. For instance, in systems supporting metadata checksums, scrubbing can flag inconsistencies in inodes, directory entries, or block group descriptors, prompting corrective actions like rewriting affected structures. A key challenge in implementing data scrubbing within file systems, especially (CoW) designs, lies in balancing the benefits against I/O overhead, as the process generates substantial read traffic that can compete with foreground workloads and exacerbate fragmentation in CoW trees. Scheduling scrubs during low-activity periods or throttling their rate helps mitigate performance impacts, though this requires careful configuration to maintain proactive detection. Examples of scrubbing in non-specialized file systems include ext4's e2scrub_all tool, introduced in e2fsprogs 1.45.0 in March 2019, which performs offline metadata checks on mounted ext4 volumes hosted on LVM logical volumes by creating read-only snapshots and running non-repairing scans; any detected issues necessitate taking the file system offline for e2fsck repairs. Similarly, Apple's APFS, deployed automatically since macOS High Sierra in 2017, employs noncryptographic checksums for ongoing metadata integrity verification on internal storage, ensuring crash consistency and structural soundness without explicit user-initiated scrubbing for user data.

File System Implementations

Btrfs

Btrfs employs a (CoW) design that facilitates through per-block applied to both data and blocks. By default, it uses the CRC32C , a 32-bit checksum that is computed before writing blocks to disk and verified upon reading, enabling precise fault isolation to specific blocks rather than entire files or volumes. This mechanism supports online repair by identifying corrupted data without halting filesystem operations. The primary tool for data scrubbing in Btrfs is the btrfs scrub start command, introduced in Linux kernel version 3.0 in July 2011. When executed on a mounted filesystem, it initiates a comprehensive scan of all data and metadata across subvolumes and underlying devices, recomputing and comparing checksums to detect discrepancies such as bit rot, media errors, or metadata corruption. If redundancy exists—such as in RAID1 or RAID10 profiles—Btrfs automatically attempts repairs by replacing erroneous blocks with verified copies from replicas, logging the outcomes for review. The process operates in the background by default, with options to specify devices, set I/O priorities, or run read-only (though read-only mode on writable filesystems may still trigger writes due to design constraints). Btrfs uniquely integrates functionality at the filesystem level through configurable profiles, allowing scrubbing to natively handle redundancy without relying on separate volume managers like MD . Administrators can pause or resume interrupted scrubs—enhanced in versions starting around 6.x for better handling of events like suspends or freezes—and monitor progress or repair statistics via btrfs scrub status, which reads from persistent logs updated every 5 seconds. To mitigate performance impacts, scrubbing can be throttled using I/O limits introduced in 5.14, targeting about 80% device bandwidth on idle systems. A full scrub on a 1 TB volume typically requires 1-2 hours on modern hardware, though actual times depend on disk speed, configuration, and data density; it excels at proactively detecting silent corruption before it affects accessibility. Recent enhancements in Btrfs 6.x kernels as of 2025, such as improved signal handling, freezing support, and performance optimizations in 6.16, enable more efficient resumption and reduce overhead for ongoing scrubs.

ReFS

The Resilient File System (), developed by for Windows environments, incorporates data scrubbing through a background process known as the data integrity scanner or scrubber, which can be enabled via Task Scheduler. When enabled, this mechanism periodically scans volumes to verify checksums embedded in integrity streams, which protect both file data and against . Upon detecting latent errors, the scrubber proactively initiates repairs using features such as block cloning or mirror copies, ensuring data resilience without manual intervention. A key aspect of scrubbing is the flag, which allows administrators to exclude specific files from the process. This attribute is particularly valuable for applications like databases that employ their own checks, preventing unnecessary overhead from the . streams, enabled by default on volumes, compute and store checksums to facilitate these verifications, extending protection to as well. Introduced with , scrubbing operates on a configurable schedule managed via Task Scheduler, defaulting to a monthly (every four weeks) run when enabled to balance integrity checks with system performance. It integrates seamlessly with Storage Spaces, leveraging virtualized storage layouts like mirrors and for automated repairs. The process handles single-block errors by replacing corrupted sectors with valid copies from redundant sources, while logging all detections and repairs in the Event Viewer under the \Windows\DataIntegrityScan channel for monitoring and auditing. Support for tiered storage ensures scrubbing spans across fast SSD tiers and slower HDD tiers without disruption. By 2025, scrubbing has seen enhanced integration in , particularly with version 24H2 and later builds, enabling native booting from volumes and support for consumer SSDs through features like Dev Drive. Additionally, 2025 introduces improvements such as deduplication and NVMe-oF support, enhancing scrubbing efficiency in environments. This expansion broadens scrubbing's applicability beyond servers to and scenarios, maintaining the system's focus on proactive error correction.

ZFS

ZFS implements a robust model through end-to-end checksums computed using the on all and blocks within a pool. These checksums enable the detection of silent at any point in the stack, from the to the physical disks. Pool-wide scrubbing is initiated via the zpool scrub command, which systematically traverses the entire pool to validate . The scrubbing process reads every in the , recomputes its , and compares it against the stored value; discrepancies trigger automatic repair using redundant copies in configurations such as mirrors or RAID-Z vdevs. If a mismatch is found, reconstructs the correct data from available replicas and rewrites the affected , ensuring self-healing without user intervention. The operation supports pausing and resuming after interruptions, allowing it to complete reliably even on large s. ZFS supports scheduled or continuous scrubbing at the pool level through automation tools, providing ongoing integrity verification beyond manual pool scrubs. It handles replication of critical via blocks—multiple on-disk copies of pool and filesystem —to enhance repair reliability during scrubs. Originally introduced in in 2005, ZFS has been ported to since 2008 and to via the ZFS on project, which continues to evolve with features like improved asynchronous scrubbing in versions 2.2 and later, and RAID-Z expansion in 2.3 as of 2025. Typical scrub rates in pools range from 100 to 500 MB/s, depending on configuration, pool utilization, and I/O contention, with higher speeds achievable on SSD-based or well-tuned HDD arrays. This process has proven effective in detecting subtle corruptions, such as "scribbling" caused by in disk controllers or SSDs, where erroneous overwrites occur without traditional error reporting.

Hardware Applications

Memory

Memory scrubbing refers to the systematic process of periodically reading data from , such as or , verifying it against embedded error-correcting codes (), and rewriting corrected data to mitigate soft errors. These transient faults, which do not cause permanent damage, can arise from cosmic rays or alpha particles striking cells, leading to bit flips at rates of approximately 10 to 100 failures in time (FIT) per in . In high-reliability environments like servers and data centers, scrubbing prevents the accumulation of correctable errors that could escalate into uncorrectable failures, ensuring without interrupting normal operations. The primary technique involves hardware-based scrubbers integrated into memory controllers, which perform background operations like read-check-write cycles to scan and correct errors at predefined intervals, often on the order of hours to days depending on system configuration and error rates. For instance, patrol scrubbing in server systems proactively detects and corrects single-bit errors before they are accessed, using the memory controller's ECC logic to compute syndromes and flip erroneous bits. Software-based scrubbing complements this by leveraging operating system kernels to initiate memory scans during idle periods, though it may introduce minor performance overhead compared to dedicated hardware. Both approaches rely on single-error correction, double-error detection (SECDED) schemes, where a Hamming code or similar parity mechanism generates syndromes during reads; a non-zero syndrome identifies the error location, enabling correction by inverting the affected bit before rewriting. In server environments, ECC-protected dual in-line memory modules (DIMMs) have employed scrubbing since the 1990s, coinciding with the shift to synchronous DRAM (SDRAM) architectures that prioritized reliability for mission-critical workloads. Modern implementations, such as those in DDR5 memory for data centers, incorporate on-die ECC with built-in scrubbing features to handle increased error rates from higher densities, particularly in AI training systems where large-scale DRAM usage amplifies soft error risks. These advancements, as detailed in a 2023 review of memory reliability, availability, and serviceability (RAS) technologies, underscore scrubbing's role in sustaining performance for compute-intensive applications.

FPGAs

Field-Programmable Gate Arrays (FPGAs) rely on SRAM-based configuration memory to define their programmable logic, making them susceptible to single event upsets (SEUs) from , which can corrupt the bitstream and alter circuit behavior. FPGA scrubbing addresses this by periodically reconfiguring the configuration memory to detect and correct errors, restoring the device to its intended state without halting user logic in most cases. This technique is essential for maintaining reliability in radiation-prone environments, where uncorrected errors could propagate to functional failures. Common scrubbing methods include full reconfiguration, which overwrites the entire from a golden copy, and partial dynamic reconfiguration, which targets specific frames or regions to minimize overhead and downtime. Partial methods, supported by tools like ISE since the early 2000s, leverage frame-level access for efficiency. (TMR) is often integrated with scrubbing, using majority voting to mask errors alongside periodic reconfiguration for correction, significantly improving in designs. The scrubbing process typically involves readback verification of the against a golden reference, error detection via built-in error-correcting code (), and correction through targeted reloads; in space applications, intervals are tuned to minutes or daily cycles to prevent error accumulation exceeding the mitigation window. In aerospace and satellite systems, scrubbing is a standard practice, as exemplified by NASA's use in Virtex FPGAs, where external scrubbing via SelectMAP outperforms internal methods for handling multi-bit upsets (MBUs). Configuration memory SEU rates in geostationary orbit (GEO) can reach up to 4 errors per day for devices like the Virtex-5 SX55, necessitating scrubbing rates that outpace upset accumulation to limit system failure probability. Recent advancements in AMD/Xilinx Versal adaptive SoCs (first introduced in 2018), feature XilSEM scrubbing with enhanced TMR-based protection, offering up to 30 times greater SEU mitigation than prior generations and enabling reliable edge AI inferencing in space with single event functional interrupt (SEFI) rates as low as 1 per 200 years in GEO during solar minimum.

References

  1. [1]
    [PDF] Data Cleaning: Problems and Current Approaches. - SciSpace
    Data cleaning, also called data cleansing or scrubbing, deals with detecting and removing errors and incon- sistencies from data in order to improve the quality ...
  2. [2]
    Data Cleaning | NNLM
    Jun 14, 2022 · Similar Terms. Deduplication. Data Standardization. Data Cleansing. Data Scrubbing. Tools. Tidyverse is a collection of open source R packages ...
  3. [3]
    [PDF] Nguyen-Hoang, Tu Anh Data quality management in big data
    Jun 1, 2025 · Automated data cleaning and cleansing: Five research studies on data cleaning have made significant contributions, each with its own.
  4. [4]
    [PDF] A Systematic Review of Tools for AI-Augmented Data Quality ... - arXiv
    Jun 29, 2025 · This study reviews tools for AI-augmented data quality management, finding 10 tools that met criteria, but most focus on data cleansing rather ...
  5. [5]
    [PDF] A Clean-Slate Look at Disk Scrubbing - USENIX
    Disk scrubbing is a background process that reads disks during idle periods to detect irremediable read er- rors in infrequently accessed sectors.
  6. [6]
    Scrubbing with partial side information for radiation-tolerant memory
    ### Summary of Scrubbing Definition in Memory Context with ECC
  7. [7]
    [PDF] An Analysis of Data Corruption in the Storage Stack - USENIX
    Abstract. An important threat to reliable storage of data is silent data corruption. In order to develop suitable protection.
  8. [8]
    [PDF] Practical Scrubbing: Getting to the bad sector at the right time
    To protect against data loss due to LSEs, most commercial storage systems use a “scrubber”: a background process that periodically performs full-disk scans to ...
  9. [9]
    [PDF] Data Cleaning: Problems and Current Approaches
    Data cleaning, also called data cleansing or scrubbing, deals with detecting and removing errors and ... Definition of transformation workflow and mapping ...
  10. [10]
    [PDF] Disk Scrubbing Large Archival Storage Systems
    By scrubbing all of the data stored on all of the disks, we can detect block failures and compensate for them by rebuilding the affected blocks.Missing: definition | Show results with:definition
  11. [11]
    [PDF] What We Learned from 10K SSD-Related Storage System Failures
    Jul 12, 2019 · Classic techniques like data scrubbing can effectively mitigate retention issue by scanning and checking data in- tegrity. However, it is ...
  12. [12]
    [PDF] Software and Digital Systems Program - Data Integrity Techniques
    Sep 25, 2014 · With “scrubbing” (testing that the error detection mechanisms are still functioning ... – CRCs provide burst error detection up to CRC size.
  13. [13]
    Ensuring Data Integrity in Storage: Techniques and Applications
    Checksums are generated using a hash function. The use of cryptographic hash functions has become a standard in Internet applications and protocols.
  14. [14]
    [PDF] A Clean-Slate Look at Disk Scrubbing - USENIX
    Most existing systems perform sequential disk scrub- bing, meaning that they access disk sectors by increas- ing logical block address, and use a scrubbing ...
  15. [15]
    Enhancing ML-based anomaly detection in data management for ...
    This study introduces a multi-tiered machine learning-based approach to detect anomalies, specifically targeting security threats, performance irregularities, ...
  16. [16]
    [PDF] Parity Lost and Parity Regained - Computer Sciences Dept.
    The scrub should fix the inconsistency (by recomputing the contents of the parity disk) because inconsistent data and parity lead to data corruption if a second ...
  17. [17]
    [PDF] Lecture 18: Error Detection and Correction
    Nov 27, 2021 · ▷ Hamming codes use m equations to correct one error in 2m - 1 bits. ▷ Most useful when the error rate is low. ECC RAM uses Hamming codes.
  18. [18]
    [PDF] End-to-end Data Integrity for File Systems: A ZFS Case Study
    COW transactions for atomic updates: ZFS maintains data consistency in the event of system crashes by using a copy-on-write transactional update model. ZFS ...
  19. [19]
    Hamming error-correcting codes - Project Nayuki
    Apr 18, 2025 · For each parity group i, assign it the value 0 if the check passes or 2i if the check fails. Then, sum up all these values to yield the syndrome ...
  20. [20]
    Improving the accuracy, adaptability, and interpretability of SSD ...
    Oct 12, 2020 · We propose the use of 1-class isolation forest and autoencoder-based anomaly detection techniques for predicting previously unseen SSD failure ...
  21. [21]
    5.4.16.10. Scrubbing a RAID Logical Volume
    RAID scrubbing is the process of reading all the data and parity blocks in an array and checking to see whether they are coherent.
  22. [22]
    Dell PowerEdge RAID Controller 11 User's Guide PERC H755 ...
    The Patrol read feature is designed as a preventative measure to ensure physical disk health and data integrity. Patrol read scans and resolves potential ...Missing: scrubbing | Show results with:scrubbing
  23. [23]
    Data Scrubbing | DSM - Synology Knowledge Center
    Go to the Storage page. · Click the Schedule Data Scrubbing button. · Tick the Enable data scrubbing schedule checkbox. · Select and prioritize the storage pools ...
  24. [24]
    RAID - ArchWiki
    RAID is a storage technology that combines multiple disk drive components (typically disk drives or partitions thereof) into a logical unit.Convert a single drive system... · Install Arch Linux with Fake RAID · Talk:RAID
  25. [25]
    Scrubbing a RAID Logical Volume - Naver Blog
    May 7, 2015 · In Dell PowerEdge RAID environments, a feature called "patrol read" can perform data scrubbing and preventive maintenance. Linux MD RAID ...
  26. [26]
    Replacing a Linux RAID Drive - irq5.io
    Jul 13, 2017 · Most Linux distributions provide the raid-check script for periodic RAID scrubbing. This is basically a background cron job that tells the ...<|control11|><|separator|>
  27. [27]
    How to prevent silent data corruption by using Data Scrubbing ...
    Jul 20, 2022 · Go to Storage & Snapshots. · Click the Global Settings icon · Go to the Storage tab. · Enable RAID scrubbing schedule and configure the Frequency.
  28. [28]
    [PDF] Introducing KIOXIA Data Scrubbing Offload Technology
    KIOXIA's offload technology moves data scrubbing to SSDs, offloading compute and DRAM demands, freeing up resources and improving performance.
  29. [29]
    NVME | SSD RAID and HBA | Storage Adapters - Broadcom Inc.
    Designed to deliver the best possible performance for NVMe SSD-based storage systems, the 9600 family delivers 4X the bandwidth, 6X the RAID IOPs, 25X lower and ...<|control11|><|separator|>
  30. [30]
    Filesystem Data Integrity: Checksums, Scrubbing, and Silent ...
    Understand how modern filesystems protect against data corruption with checksums, scrubbing, and error correction. Explore integrity mechanisms through ...
  31. [31]
    Improving ext4: bigalloc, inline data, and metadata checksums
    Nov 29, 2011 · ZFS also does periodic scrubbing to validate checksums. Improving ext4: bigalloc, inline data, and metadata checksums. Posted Nov 30, 2011 1 ...
  32. [32]
    e2scrub_all(8) — e2fsprogs - testing - Debian Manpages
    Jul 30, 2025 · Searches the system for all LVM logical volumes containing an ext2, ext3, or ext4 file system, and checks them for problems.Missing: introduction Linux 4.0 2015
  33. [33]
    E2fsprogs Release Notes - SourceForge
    E2fsprogs 1.04 (May 16, 1996). First "official" (1.03 was a limited release only) to support building e2fsprogs under Linux 2.0 kernels (as well as late ...E2fsprogs 1.45.5 (January 7... · E2fsprogs 1.45.1 (May 12, 2019)
  34. [34]
    Signed system volume security - Apple Support
    May 13, 2022 · Since its introduction, APFS has provided file-system metadata integrity using noncryptographic checksums on the internal storage device.
  35. [35]
    Checksumming - BTRFS documentation!
    Data and metadata are checksummed by default. The checksum is calculated before writing and verified after reading the blocks from devices.Missing: per- | Show results with:per-
  36. [36]
    Linux_3.0 - Linux Kernel Newbies
    Linux 3.0 released on 21 Jul, 2011. Summary: Besides a new version numbering scheme, Linux 3.0 also has several new features: Btrfs data scrubbing and ...
  37. [37]
    Scrub - BTRFS documentation! - Read the Docs
    Scrub is a validation pass over all filesystem data and metadata that detects data checksum errors, basic super block errors, basic metadata block header errors ...Missing: 6. enhancements incremental
  38. [38]
    btrfs-scrub(8) - Linux manual page - man7.org
    btrfs scrub is used to scrub a mounted btrfs filesystem, which will read all data and metadata blocks from all devices and verify checksums.Missing: 6. incremental
  39. [39]
    btrfs: scrub: enhance freezing and signal handling - LWN.net
    Oct 16, 2025 · ... paused scrub will still make the ioctl executing process trapped inside kernel space. ... Qu Wenruo (3): btrfs: scrub: add cancel/pause ...<|separator|>
  40. [40]
    Resilient File System (ReFS) overview - Microsoft Learn
    Jul 28, 2025 · ... ReFS introduces a data integrity scanner, known as a scrubber. This scrubber periodically scans the volume, identifying latent corruptions ...Key benefits · Supported deployments
  41. [41]
    ReFS integrity streams - Microsoft Learn
    Nov 1, 2024 · This scrubber periodically scans the volume, identifies latent corruptions, and proactively triggers a repair of any corrupt data. Note. The ...How it works · Performance
  42. [42]
    File Attribute Constants (WinNT.h) - Win32 apps | Microsoft Learn
    Sep 23, 2025 · This value is reserved for system use. FILE_ATTRIBUTE_NO_SCRUB_DATA; 131072 (0x00020000). The user data stream not to be read by the background ...
  43. [43]
    [MS-FSA]: Appendix A: Product Behavior | Microsoft Learn
    Mar 7, 2024 · 1: Only ReFS supports FILE_ATTRIBUTE_INTEGRITY_STREAM. <52> Section 2.1.5.1.1: Only NTFS and ReFS support FILE_ATTRIBUTE_NO_SCRUB_DATA. <53> ...
  44. [44]
    ReFS with Integrity Streams on Windows 11 - Can it be trusted?
    Jan 16, 2022 · You can go to Task Scheduler -> Microsoft -> Windows -> Data Integrity, and pick the 2nd task from the list. Set a weekly schedule and you ...
  45. [45]
    What's new in Windows Server 2025 | Microsoft Learn
    Feb 28, 2025 · Dev Drive now supports block cloning starting with Windows 11 24H2 and Windows Server 2025. Because Dev Drive uses the Resilient File System ( ...DTrace on Windows · Server 2022 · Mica material
  46. [46]
    Windows 11 Setup will let you choose between NTFS and ReFS ...
    Mar 27, 2025 · Based on changes seen in newer builds, it looks like Windows Setup will allow you to choose between NTFS (the default) and ReFS, the modern file format.Missing: integration | Show results with:integration
  47. [47]
    Windows 11 24H2/25H2 Supports Being Installed To And Booting ...
    Sep 27, 2025 · Version 24H2/25H2 seems to fully support being installed to and booting from an ReFS system drive. Everything seems to work fine, tried ...
  48. [48]
    zpool-scrub.8 — OpenZFS documentation
    DESCRIPTION. Begins a scrub or resumes a paused scrub. The scrub examines all data in the specified pools to verify that it checksums correctly.
  49. [49]
    Controlling ZFS Data Scrubbing
    Performing routine scrubbing guarantees continuous I/O to all disks on the system. Routine scrubbing has the side effect of preventing power management from ...
  50. [50]
    History - - OpenZFS
    May 25, 2014 · 2005 – Source code was released as part of OpenSolaris. 2006 – Development of FUSE port for Linux started. 2007 – Apple started porting ZFS to ...
  51. [51]
    [PDF] ZFS BEST PRACTICES. | FreeBSD Foundation
    Important data in ZFS is stored multiple times, in what are called ditto blocks. Pool-wide data has three ditto blocks (so is stored three times), and file ...<|separator|>
  52. [52]
    ZFS scrub rates, speeds, and how fast is fast
    Sep 11, 2015 · The fastest rate a raidz scrub can report is your total disk bandwidth across all disks and the fastest rate a mirror scrub can report is your single disk ...
  53. [53]
    [PDF] Soft Memory Errors and Their Effect on Sun Fire Systems
    Generally speaking, cosmic ray soft errors occur in DRAM memory at a rate of ~10 to 100 FIT/MB (1 FIT = 1 device fail in 1 billion hours). So a system with 10 ...
  54. [54]
    DRAM Errors and Cosmic Rays: Space Invaders or Science Fiction?
    Jul 23, 2024 · It is widely accepted that cosmic rays are a plausible cause of DRAM errors in high-performance computing (HPC) systems.
  55. [55]
    A study of DRAM failures in the field
    Nov 10, 2012 · The hardware scrubber periodically reads every location in memory. Its goal is to correct any latent (unaccessed) correctable errors before a ...
  56. [56]
    Explaining the ECC Background Scrubbing Feature of the DDRMC
    The Enable Background Scrubbing feature issues a periodic Read Check Write operation which checks the existing data and ECC syndrome in the memory for ECC ...
  57. [57]
    What to do in response to repeat DRAM ECC error notifications for ...
    Aug 31, 2017 · As the scrubbing consists of normal read and write operations, it may increase power consumption for the memory compared to non-scrubbing ...
  58. [58]
    Can Linux scrub memory? - Super User
    Dec 28, 2011 · The answer is yes, and it is done transparently (provided you have ECC memory to detect errors, and your kernel version is at least 2.6.30 to continue to ...
  59. [59]
    [PDF] Software-Defined Error-Correcting Codes - UCLA NanoCAD Lab
    Feb 27, 2016 · Reliability management techniques such as memory page retirement [30] or scrubbing. [31] are opportunistic and incur little or no hardware cost.<|separator|>
  60. [60]
    The evolution of server RAM – from memory tubes to DDR5
    Jun 17, 2025 · The transition to SDRAM with ECC came in the course of the 1990s. In contrast to DRAM (asynchronous architecture), 'Synchronous DRAM' works ...
  61. [61]
  62. [62]
    (PDF) Review of Memory RAS for Data Centers - ResearchGate
    Jul 31, 2025 · This paper describes an overview of the current RAS technologies and trends in memory for data centers, which includes an analysis of conventional ECC ...
  63. [63]
    [PDF] Effectiveness of Internal vs. External SEU Scrubbing Mitigation ...
    The purpose of configuration mitigation is to protect the FPGA configuration from upsets. This approach requires the system to have the ability to write into ...
  64. [64]
    [PDF] Differentiating Scrub Rates Between Space-Flight Applications and ...
    Apr 12, 2013 · – Reduces the accumulation bit error rate. – Does not correct functional upsets. • Scrub Rate (dC/dt) must be fast enough to beat accumulation.
  65. [65]
    Design Techniques for Xilinx Virtex FPGA Configuration Memory ...
    Jan 15, 2013 · Several techniques exist for coping with radiation ... scrubbing, which allows error mitigation and prevents failures due to error accumulation.
  66. [66]
    [PDF] AMD XQR Versal Adaptive SoCs Enable Next - Indico at ESA / ESTEC
    Mar 25, 2025 · in AMD Versal Adaptive SoCs. • Design uses AMD Versal AI Edge VE2302. • Real-time on-orbit anomaly detection for up to 80 channels of ...