Best practices for large-scale R2 backup (≈250 TB, long-running)
🌍CDN✅Solved
Hey all ! (I'm using a translator)
I’m planning a one-time large backup operation from Cloudflare R2 to another object storage (Backblaze B2), and I’d really appreciate some guidance to make sure I handle this as safely as possible.
Context: - Around 250 TB total - Multimedia content uploaded by my customers, mostly delivered using HLS (m3u8 playlists with many small .ts segments) - The R2 bucket is strictly read-only during the operation (no deletes, no sync) - Using rclone in copy mode only
Important detail: this operation is specifically meant to create a backup. At the moment, I don’t have any other copy of this data, so my main concern is avoiding anything that could trigger throttling, temporary blocks, or account issues during the transfer.
Planned approach: - Stable and continuous bandwidth (roughly 60–120 MB/s total, no bursts) - Limited concurrency (2–3 parallel rclone processes, low transfers/checkers) - Folder-by-folder copy (per logical content unit), no global bucket scans - No fast-list, no aggressive retries
Goal: - Keep traffic predictable and non-abusive - Safely complete a full backup without putting the source data at risk - Minimize the chance of interruptions during a long-running transfer (~2–3 weeks)
Before starting, I just want to double-check: - Is this approach aligned with R2 best practices for large-scale backup operations? - Are there specific limits or patterns I should be extra careful about when doing a long-running backup like this?