Skip to main content

How to Recover Interrupted or Failed Cloud Transfers with RcloneView

· 5 min read
Tayson
Senior Engineer

Network drops, API timeouts, laptop sleeps, and power outages interrupt cloud transfers. RcloneView and rclone have built-in mechanisms to resume safely without re-uploading everything from scratch.

Transferring terabytes to the cloud is not a five-minute operation. During long-running jobs, connectivity interruptions are almost inevitable. The good news is that rclone's intelligent transfer engine — which RcloneView uses under the hood — is designed for exactly this scenario. Copy and Sync operations are inherently idempotent: re-running them skips files already transferred and resumes from where things broke.

RcloneView app preview

Manage & Sync All Clouds in One Place

RcloneView is a cross-platform GUI for rclone. Compare folders, transfer or sync files, and automate multi-cloud workflows with a clean, visual interface.

  • One-click jobs: Copy · Sync · Compare
  • Schedulers & history for reliable automation
  • Works with Google Drive, OneDrive, Dropbox, S3, WebDAV, SFTP and more
WindowsmacOSLinux
Get Started Free →

Free core features. Plus automations available.

How Rclone Handles Interrupted Transfers

Rclone compares source and destination before transferring each file. When you re-run a Copy or Sync job after an interruption:

  • Files already transferred are skipped (by size+modification time, or checksum if enabled).
  • Partially transferred files are cleaned up and re-transferred from the beginning.
  • Not-yet-started files are queued and transferred in the resumed run.

This means there's no special "resume" button in most cases — just re-run the same job.

Step 1 — Re-Run the Same Job

After an interruption, open Jobs in RcloneView and click Run on the same job again:

Re-run a failed transfer job in RcloneView

RcloneView will:

  1. List the source and destination.
  2. Compare files already present in the destination.
  3. Skip successfully transferred files.
  4. Transfer only missing or modified files.

For a 10,000-file job where 8,000 succeeded, re-running takes a fraction of the original time.

Step 2 — Check Job History for Failed Files

Before re-running, review the Job History in RcloneView to understand what failed:

Review failed files in RcloneView job history

The log shows:

  • Which specific files failed
  • The error message for each failure
  • Whether failures were transient (network error) or persistent (permissions, path too long)

Persistent errors need a fix before re-running — transient errors will resolve on retry.

Step 3 — Handle Partially Uploaded Large Files

For very large files (multi-GB), an interruption mid-upload leaves a partial file in the destination. Rclone's behavior depends on the provider:

ProviderPartial File Behavior
Amazon S3 / S3-compatibleMultipart uploads: incomplete parts are orphaned, rclone retries from scratch
Google DriveResumable uploads: rclone can resume mid-file if the session is still valid
OneDriveResumable uploads: similar to Google Drive
Backblaze B2Large file parts: incomplete uploads are visible in B2 console

For S3 orphaned multipart uploads: These accumulate and cost money. Clean them up using:

  • RcloneView Terminal: rclone cleanup s3-remote:bucket-name
  • Or via the AWS console under S3 → Your Bucket → Multipart uploads

Step 4 — Use --retries and --low-level-retries

For jobs that fail on transient errors, configure RcloneView's job to retry automatically:

Add these to the Custom flags field:

--retries 5 --retries-sleep 10s --low-level-retries 20
  • --retries 5 — retry the entire job up to 5 times if errors occur
  • --retries-sleep 10s — wait 10 seconds between retries
  • --low-level-retries 20 — retry individual low-level operations (API calls) up to 20 times

Step 5 — Handle Checksum Mismatches

After an interrupted transfer, re-running with checksum verification ensures data integrity:

Verify transfer integrity with folder comparison

In RcloneView, enable Checksum verification in the job settings. This forces rclone to compare file contents (not just size/mtime) — slower, but guarantees that partially written files are caught and re-transferred.

Step 6 — Recover from a Sync That Deleted Files

If a Sync job ran in the wrong direction — copying the destination over the source instead of the other way around — you need to recover from cloud provider versioning:

  • Google Drive: Restore from Trash or version history (available for 30–180 days).
  • OneDrive: Restore from Recycle Bin or version history.
  • S3 with versioning: Restore from a previous version in the S3 console.
  • Backblaze B2: Restore from version history if enabled.

This is why using Copy mode (non-destructive) for initial large transfers is strongly recommended over Sync.

Prevention: Configure Transfers for Resilience

Build resilience into your transfer jobs from the start:

SettingRecommendation
Job modeUse Copy for initial migrations; Sync for ongoing maintenance
RetriesSet --retries 5 --retries-sleep 10s
ChecksumEnable for critical data
TransfersLower concurrent count for unstable connections
ScheduleRun during stable network windows (overnight, off-VPN)
BandwidthSet limits to prevent timeouts caused by network saturation

Getting Started

  1. Download RcloneView from rcloneview.com.
  2. Check Job History to identify what failed and why.
  3. Re-run the job — rclone skips completed files automatically.
  4. Configure retries and checksum verification for future resilience.

Most interrupted transfers require nothing more than clicking Run again. Rclone does the rest.


Related Guides:

Supported Cloud Providers

Local Files
WebDAV
FTP
SFTP
HTTP
SMB / CIFS
Google Drive
Google Photos
Google Cloud Storage
OneDrive
Dropbox
Box
MS Azure Blob
MS File Storage
S3 Compatible
Amazon S3
pCloud
Wasabi
Mega
Backblaze B2
Cloudflare R2
Alibaba OSS
Ceph
Swift (OpenStack)
IBM Cloud Object Storage
Oracle Cloud Object Storage
IDrive e2
MinIO
Storj
DigitalOcean Spaces