gsutil rsync hangs on large files
See original GitHub issueHi! We are using gsutil rsync
to upload our backups to coldline storage. After our files become larger (more than 30GB) rsync
starts hanging on each such file. The same situation is for gsutil cp
.
I’ve already sent a log file with gsutil -D rsync
to gs-team@google.com
gsutil version: 4.25
boto version: 2.42.0
python version: 2.7.12 (default, Nov 19 2016, 06:48:10) [GCC 5.4.0 20160609]
OS: Linux 4.4.0-77-generic
multiprocessing available: True
using cloud sdk: True
config path(s): /etc/boto.cfg
gsutil path: /usr/lib/google-cloud-sdk/platform/gsutil/gsutil
compiled crcmod: True
installed via package manager: False
editable install: False
Issue Analytics
- State:
- Created 6 years ago
- Reactions:5
- Comments:14
Top Results From Across the Web
gsutil rsync from google cloud storage to s3 hangs on large file
Create a folder with only that file. Then try again. Do not press CTRL-C. Wait until the command finishes or reports an error....
Read more >gsutil rsync hangs on file transfer · Issue #497 - GitHub
I am using gsutil rsync to sync directories from a regular POSIX fileserver over to GCP buckets. One on particular directory, gsutil rsync...
Read more >rsync - Synchronize content of two buckets/directories
The gsutil rsync command copies changed files in their entirety and does not employ the rsync delta-transfer algorithm to transfer portions of a...
Read more >Slow synchronisation stage on gsutil rsync? - Server Fault
Well I now know that it's each 32,000th file that creates a large pause. Which is the size of "buffer_size" in that file....
Read more >Slow synchronisation stage on gsutil rsync? - Kiwix Server
Is trying to use gsutil rsync with 2 million files to one bucket a bad idea? ... Well I now know that it's...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Hello, I’m trying to rsync from gcs to s3 buckets but some files are bigger than 5 GiB.
I receive this error: “exceeds the maximum gsutil-supported size for an S3 upload. S3 objects greater than 5 GiB in size require multipart uploads, which gsutil does not support.”
As you can see here https://docs.aws.amazon.com/AmazonS3/latest/dev/qfacts.html there is a limitation in S3 where you have to use multipart upload for files bigger than 5 GiB.
Please help me, how can I fix that?
Any help will be really appreciated! thanks @houglum for your time
Best Regards Fabio Rigato
I am also facing same issue