Skip to main content

Transfer Files Between Google Cloud Storage and AWS S3 Without CLI Using RcloneView

· 8 min read
Tayson
Senior Engineer

Managing data across Google Cloud Storage and AWS S3 usually means juggling gsutil, aws cli, and custom scripts. RcloneView lets you do it all from a visual interface — browse, compare, sync, and schedule transfers between GCS and S3 in minutes.

Multi-cloud is the reality for most engineering teams. Your ML training data sits in GCS buckets, your production assets are on S3, and someone needs to keep them synchronized. The traditional approach — writing shell scripts with gsutil and aws cli — works, but it's fragile, hard to monitor, and impossible for non-engineers to manage.

RcloneView connects to both GCS and S3 natively, giving you a unified GUI to browse, transfer, compare, and automate data movement between the two biggest cloud platforms.

RcloneView app preview

Manage & Sync All Clouds in One Place

RcloneView is a cross-platform GUI for rclone. Compare folders, transfer or sync files, and automate multi-cloud workflows with a clean, visual interface.

  • One-click jobs: Copy · Sync · Compare
  • Schedulers & history for reliable automation
  • Works with Google Drive, OneDrive, Dropbox, S3, WebDAV, SFTP and more
WindowsmacOSLinux
Get Started Free →

Free core features. Plus automations available.

Why Move Data Between GCS and S3?

Teams transfer data between Google Cloud Storage and AWS S3 for several common reasons:

Multi-cloud redundancy — Storing critical data on two major providers protects against provider-level outages and vendor lock-in. If one cloud goes down, your data is accessible from the other.

Cost optimization — GCS and S3 have different pricing for storage, egress, and operations. Moving cold data to whichever provider is cheaper for your usage pattern can save significant money.

Cross-platform workflows — Your data science team uses GCP (BigQuery, Vertex AI), but your production infrastructure runs on AWS. Data needs to flow between both.

Migration — Moving from GCP to AWS (or vice versa) without downtime requires reliable, resumable transfers.

Compliance and data residency — Some regulations require data copies in specific regions or providers.

Setting Up GCS and S3 Remotes

Add Google Cloud Storage

  1. Open RcloneView and click Add Remote.
  2. Select Google Cloud Storage from the provider list.
  3. Choose your authentication method:
    • Service Account JSON — Recommended for server-to-server transfers. Upload your service account key file.
    • OAuth (browser login) — Good for personal GCP accounts. Follow the OAuth login guide.
  4. Set your project ID and default bucket location if prompted.
  5. Save the remote — your GCS buckets are now browsable.

Add AWS S3

  1. Click Add Remote again.
  2. Select Amazon S3 from the provider list.
  3. Enter your Access Key ID and Secret Access Key.
  4. Select your region and endpoint.
  5. Save — your S3 buckets appear in the Explorer.

For detailed S3 setup, see the AWS S3 connection guide.

Add GCS and S3 remotes in RcloneView

Browsing GCS and S3 Side by Side

Once both remotes are connected, open them in RcloneView's two-pane Explorer. GCS buckets on the left, S3 buckets on the right (or vice versa). You can:

  • Navigate through buckets and folders on both sides simultaneously.
  • View file sizes, dates, and counts to understand what's where.
  • Drag and drop files directly from GCS to S3 — or use the built-in copy/move commands.

This side-by-side view gives you instant visibility into both clouds without switching between the GCP Console and AWS Console.

Browse GCS and S3 side by side in RcloneView

Transfer Scenarios

Scenario 1: One-Time Migration (GCS → S3)

Moving all data from GCS to S3 for a platform migration:

  1. Create a Copy job:
    • Source: GCS remote → select your bucket
    • Destination: S3 remote → select target bucket
  2. Configure for maximum speed:
    • Parallel transfers: 8–16 (both GCS and S3 handle high parallelism well)
    • Chunk size: 64MB–128MB for large files
    • Enable --fast-list flag to speed up directory listing
  3. Run the job and monitor progress in real time.

For large migrations, run the Copy job multiple times. After the first full copy, subsequent runs only transfer new or changed files — making it safe to resume if interrupted.

Scenario 2: Ongoing Sync (Bidirectional)

Keeping a GCS bucket and S3 bucket in sync continuously:

  1. Create a Sync job (GCS → S3) for the primary direction.
  2. Schedule it to run hourly or daily using Job Scheduling.
  3. Add a reverse Sync job (S3 → GCS) if you need bidirectional sync.
  4. Use Batch Jobs (v1.3) to run both directions in sequence.

Scenario 3: Selective Cross-Cloud Backup

Back up only specific data to the other cloud:

  1. Use Filter Rules to include/exclude specific file types or folders.
    • Example: Only sync *.parquet and *.csv files (ML datasets)
    • Example: Exclude tmp/ and logs/ directories
  2. Create a scheduled Copy job with these filters applied.

Comparing GCS and S3 Contents

Before and after transfers, use Folder Comparison to verify that both buckets contain the same data:

  • Files only in GCS — Highlighted for easy identification.
  • Files only in S3 — Shows what exists at the destination but not the source.
  • Different files — Files with the same name but different sizes or checksums.
  • Identical files — Confirmed matches across both clouds.

This is invaluable for migration verification: after copying terabytes of data, you can prove that every file arrived intact.

Compare GCS and S3 bucket contents

Optimizing Transfer Speed

GCS and S3 are both high-performance object stores, so you can push transfers hard:

SettingRecommended ValueWhy
Parallel transfers8–16Both providers handle concurrent requests well
Chunk size64MB–128MBReduces API overhead for large files
Checkers16–32Speeds up the comparison phase for large directories
Buffer size128MBSmooths out network latency between cloud regions
Fast-listEnabledDramatically reduces API calls for directory listing

Cross-region considerations

If your GCS bucket is in us-central1 and your S3 bucket is in eu-west-1, data must traverse the internet between regions. For best performance:

  • Keep source and destination in the same geographic area when possible.
  • Run transfers during off-peak hours to avoid network congestion.
  • Monitor egress costs — both GCS and S3 charge for data leaving their networks.
Monitor GCS to S3 transfer speed

Automating GCS ↔ S3 Workflows

Daily data pipeline

Set up a scheduled job that runs every night:

  1. Sync new ML training data from GCS → S3 for AWS-based training jobs.
  2. Copy results back from S3 → GCS for BigQuery analysis.
  3. Get notified via Slack when the pipeline completes.
Schedule GCS to S3 data sync

Disaster recovery replication

Maintain a live copy of critical S3 data in GCS (or vice versa):

  1. Create a Sync job from your primary bucket to the DR bucket.
  2. Schedule it hourly for RPO (Recovery Point Objective) under 1 hour.
  3. Use checksum verification to ensure data integrity.

Cost-based tiering

Move data to whichever provider is cheaper for its access pattern:

  1. Frequently accessed data → Keep on the provider closest to your compute.
  2. Cold/archive data → Move to GCS Nearline/Coldline or S3 Glacier based on pricing.
  3. Schedule periodic tiering jobs to keep costs optimized.

GCS vs S3: Using RcloneView as a Unified Layer

Instead of learning two different CLIs (gsutil for GCS, aws s3 for S3), RcloneView gives you a single interface for both. This means:

  • One tool to learn — Your team doesn't need to master two different command-line interfaces.
  • Visual operations — Drag-and-drop, right-click menus, and dialog-based configuration instead of flags and arguments.
  • Consistent monitoring — Same Job History and Transfer Monitoring regardless of which clouds are involved.
  • Portable jobs — A job that syncs GCS to S3 works exactly the same way as one that syncs OneDrive to Dropbox.
Unified job history for GCS and S3 transfers

Getting Started

  1. Download RcloneView from rcloneview.com (Windows, macOS, Linux).
  2. Add your GCS remote with a service account key or OAuth login.
  3. Add your S3 remote with access key credentials.
  4. Browse both side by side in the Explorer.
  5. Create a Copy or Sync job for your use case.
  6. Schedule and monitor for hands-free multi-cloud data management.

Stop juggling gsutil and aws cli. RcloneView unifies GCS and S3 management into one visual workflow — making multi-cloud data transfers accessible to your entire team, not just the engineers who know the CLI.


Related Guides:

Supported Cloud Providers

Local Files
WebDAV
FTP
SFTP
HTTP
SMB / CIFS
Google Drive
Google Photos
Google Cloud Storage
OneDrive
Dropbox
Box
MS Azure Blob
MS File Storage
S3 Compatible
Amazon S3
pCloud
Wasabi
Mega
Backblaze B2
Cloudflare R2
Alibaba OSS
Ceph
Swift (OpenStack)
IBM Cloud Object Storage
Oracle Cloud Object Storage
IDrive e2
MinIO
Storj
DigitalOcean Spaces