I've set up a scaleway glacier backend which is working well with one exception: settier
# rclone -v --progress settier STANDARD scaleway:guybrush-scripts/0pdj43lcacui7p71d2cr7uumec
2020-06-24 01:18:31 ERROR : S3 bucket guybrush-scripts: Failed to do SetTier, EntityTooLarge: Your proposed upload exceeds the maximum allowed object size.
I get the same error with move/copy, presumably because glacier backends do not support server-side copy. Why I should get EntityTooLarge for settier is more of a mystery though. I can't see anything about about maximum object size in the scaleway console.
Has anybody seen this before?
Changing an object's storage class in the scaleway console works fine.
AFAICT the modification time is also in metadata. Can rclone read this on GLACIER objects, or is it syncing on size only? The Scaleway console won't let me read metadata on GLACIER objects, but it shows up for NORMAL objects.
rclone is like a compatibility layer for object stores, so I wouldn't be too fussed about adding some magic to settier. You could always do both, but then you have the problem that a "restore" command might make no sense to some/most backends. It's also asymmetric.. you have "restore" but no "freeze"...
My use case is restoring backups, which would be single object, directory, or whole bucket. Single objects can already be done with most web interfaces, so it's the directory/recursive/bucket case that needs tooling.
I tried using a bucket in Scaleway's nl-ams region instead of fr-par, and it worked fine. Scaleway seem to have a few issues with their service. Unfortunately I can't give it a more thorough test until they fix their fr-par endpoint.