STOP and READ USE THIS TEMPLATE NO EXCEPTIONS - By not using this, you waste your time, our time and really hate puppies. Please remove these two lines and that will confirm you have read them.
What is the problem you are having with rclone?
Hi all,
I read a lot of articles, SO topics, rclone forum topics, I still have a problem and I'm out of ammo, so I try to ask for help here.
My request is "quite" simple : I need to sync a bucket on Ceph+radosgw with another bucket on Wasabi. This bucket has millions of objects, with a total size of around 100TB.
I ran the command you see in the following section, I tuned parameters after some tests on a smaller bucket, to ensure my Ceph was not too much stressed by the massive read.
Then, the timeline for that first sync is:
- day 0, 08:00 : sync is launched, requests are made to Ceph & wasabi endpoints
- day 0, 14:30 : no more requests, a CPU core is used at 100% (all the others are idle), the only logs I saw is these:
Transferred: 0 B / 0 B, -, 0 B/s, ETA -
Elapsed time: 1m0.3s
[...]
Transferred: 0 B / 0 B, -, 0 B/s, ETA -
Elapsed time: 20h59m0.3s
- day 1, 11:30 : data transfer is starting, using all the max bandwith I setup (250MB during the day, 500MB during the night)
- day 6, 10:00 : data transfer is done
Timing summary is:
- list files in buckets : 6h30m
- CPU thing (difference ? graph ?) : 21h00m
- data transfer (110TB) : 4 days + 22h30m
After that first sync, I made a second run, and timing were:
- list files in buckets: 30m
- CPU thing (difference ? graph ?) : 20h00m
- data transfer (1TB) : 1h30m
So, my question is: what is clone doing with the CPU during these 21 hours ? Is it possible to optimize it, multithread, etc ?
Run the command 'rclone version' and share the full output of the command.
rclone v1.62.2
- os/version: ubuntu 20.04 (64 bit)
- os/kernel: 5.4.0-148-generic (x86_64)
- os/type: linux
- os/arch: amd64
- go/version: go1.20.2
- go/linking: static
- go/tags: none
Which cloud storage system are you using? (eg Google Drive)
Source : S3 (Ceph + Radosgw)
Destination : S3 (Wasabi)
The command you were trying to run (eg rclone copy /tmp remote:tmp
)
rclone \
--config=/etc/rclone/rclone.conf \
--rc \
--rc-addr ":5572" \
--rc-enable-metrics \
--rc-web-gui \
--rc-user rclone \
--rc-pass ************ \
--max-delete 1000 \
--checkers 32 \
--fast-list \
--transfers 16 \
--buffer-size=1G \
--bwlimit "00:00,500M 05:00,250M" \
--s3-chunk-size 32M \
--s3-upload-concurrency 16 \
--s3-list-chunk 50000 \
--checksum \
--dump=headers --retries=1 --low-level-retries=1 --log-level=DEBUG --log-file=/tmp/rclone/rclone.log \
sync \
ceph:************ \
wasabi:************
The rclone config contents with secrets removed.
[ceph-production]
type = s3
provider = Ceph
region =
env_auth = false
access_key_id = ****************************
secret_access_key = ****************************
endpoint = **********************
acl = private
[wasabi]
type = s3
provider = Wasabi
env_auth = false
access_key_id = ****************************
secret_access_key = ****************************
endpoint = **********************
acl = private
A log from the command with the -vv
flag
It generates a lot of logs, not sure if it is relevant to paste it here.