Colons :
, new lines \n
, etc.
Yes, I did, this hasn't happened in any of the previous logs, I am trying to figure out what to make of it, what does it mean?
Colons :
, new lines \n
, etc.
Yes, I did, this hasn't happened in any of the previous logs, I am trying to figure out what to make of it, what does it mean?
to be clear, both rclone and rsync fail with the same weird symbols
?
and why do you need encoding
in the remote config?
I see, but still I prefer the multithreading capabilities of rclone.
For some reason both rclone and rsync were failing to sync files with those symbols in their name, and I couldn't make rsync sync them, however with the encoding capabilities of rclone it was able to successfully sync them.
Its a brand new SSD and I did a SMART test on the SSD and all seems good:
sudo smartctl -H /dev/sdc
[sudo] password for usertemp:
smartctl 7.4 2023-08-01 r5530 [x86_64-linux-6.8.0-40-generic] (local build)
Copyright (C) 2002-23, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
I researched how to debug Ubuntu Server the best I can, and here is what I came up with so far:
Ubuntu Server DEBUG information:
sudo cat /var/log/kern.log | grep error
sudo cat /var/log/syslog | grep error
Still those things seem to be unrelated according to what I found on the internet.
I will try a direct smb mount to rclone, like kapitainsky suggested, and report back.
In the mean time any help on how to debug this and find what exactly is failing will really mean a lot to me...
More DEBUG information:
Here are some screenshots of when the crash happens on the VM, the CPU just flatlines:
And also screenshots of the TrueNas Scale VM, it doesn't freeze up and is constantly accessible:
There is nothing yet pointing into rclone I am afraid. What is does is just reading a lot of data (all files) from both src and dst as you run check with --checksum
flag.
And it does kill your VM....
Do you have another SSD to test? Can you connect it via SATA? and not USB?
Also if you have enough disk space on your Proxmos server you could create ad-hoc 2TB disk and sync/check. If it works it would be very strong argument against rclone issues.
You could also try some other OS.
PS. Also try Proxmos forum - sometimes things do not work as they should when VMs are involved, e.g. Backup to Samba Share freezes | Proxmox Support Forum
Thanks for the reply, I appreciate it because I am not able to find anything else yet.
I figured it might be a good idea to post to ProxMox forums since you suggested it might not be an rclone, but a VM issue, and I posted yesterday, however noone has replied and by now my topic in the ProxMox forum is far behind new topics an has no visibility and will probably get no replies...
It is connected via SATA, and I took out the SSD and connected it to a USB v3.0 enclosure - same problem.
It does not seem to be the connection...
Unfortunately I don't have another SSD to test with, nor sufficient space on ProxMox to create a disk there.
I am trying to debug the ubuntu server vm and the ProxMox PVE, but so far I am not seeing anything in the logs, I am trying to find out what is pushed to the limit so it kills the VM... no luck so far...
Yeah it might be very pesky issue indeed.
What about to try the same rclone check
without any VM? Directly from Proxmos OS level? If it works than you would know that issue is VMs.
It would be good if rclone
has some features for throttling
local syncs, such as from one local dir to another local dir.
It seems in the case of local sync the --bwlimit
flag does not take effect.
Also it would be good to have the same throttling flags to both the sync
and check
command, for such scenarios, where the sync
command succeeds but the check
command goes blazing fast and overloads something.
As it is Linux you could use cgroups to throttle all aspects of rclone process including I/O bandwidth.
But in this case I think you will only try to mask existing issue with throttling - whatever the issue is. OS like Ubuntu should not die because some process is reading a lot of data....
Well said.
I have posted on AskUbuntu in StackExchange and on the Ubuntu forum... let's hope that the threads get some replies as well as the one on the ProxMox forum...
It really messes with me, if I can't find a problem and fix it, but rather patch it up with a "temporary" work around.
Hello again,
So, I made some time to do more debugging:
I passed the SMB shares directly to rclone
as suggested in the comments above.
There was no change, it still crashed.
However it seems that now the --bwlimit
flag is actually taking effect as I played around with throttling it.
Previously when the SMBs were locally mounted and I was passing them to rclone
as locals it seemed that the --bwlimit
had no effect at all.
I also re-wrote the script to use rsync
and it did not crash.
When it ran with rsync
it did not reach any high bandwidth throughput, which is I guess to the non-multithreaded nature of rsync
, it also ran slower, so it did not seem to push something to the limit, something related to I/O reading/writing large amounts of data at a very high speed/rate.
I do want to be able to use rclone
because of it's more advanced features, not just multithreading but encoding, SMB support etc...
With rsync
I also got the following errors in the log file:
rsync: [receiver] mkstemp "/mnt/sd25-4tb-bkp4-mnt4/tns-maindatads-smb/tns-maindatads-smb/main-data-subds-1/GoogleDrive-alk8915-bkp-2024-07-09-1800/Programirane neshta failove alk19890105/SoftUni/SoftUni AI 2021 Materials/SoftUni - AI - Data Science June 2021/softuni-ai-datascience-2021-project-FINAL-SUBMISSION/softuni-ai-datascience-2021-project-atkuzmanov-FINAL/resources/images/.Icon\#015.pu2lTm" failed: Invalid argument (22)
I don't get these problems with the rclone
encoding config I have posted in the beginning of the thread, this is another reason why I want to use rclone
, I was not able to find out how to get around those errors in rsync
.
I also got this weird output in the log:
2024/08/17 22:33:46 [9667] calling match_sums /mnt/tns-maindatads-smb/main-data-subds-1/Prog hranilishte alk8915 arhiv bkp-2024-07-10-1118/Rabotilnitsa-programirane-alk19890105-2022-08-02-1455.zip
calling match_sums /mnt/tns-maindatads-smb/main-data-subds-1/Prog hranilishte alk8915 arhiv bkp-2024-07-10-1118/Rabotilnitsa-programirane-alk19890105-2022-08-02-1455.zip
tns-maindatads-smb/main-data-subds-1/Prog hranilishte alk8915 arhiv bkp-2024-07-10-1118/Rabotilnitsa-programirane-alk19890105-2022-08-02-1455.zip
^M 32.77K 0% 97.86kB/s 38:59:03 rsync: [receiver] mkstemp "/mnt/sd25-4tb-bkp4-mnt4/tns-maindatads-smb/tns-maindatads-smb/main-data-subds-1/Prog hranilishte alk8915 arhiv bkp-2024-07-10-1118/.Icon\#015.0JwIpt" failed: Invalid argument (22)
^M 46.92M 0% 44.75MB/s 0:04:58 ^M 114.26M 0% 54.51MB/s 0:04:03 ^M 183.11M 1% 58.25MB/s 0:03:47 ^M 251.33M 1% 59.97MB/s 0:03:39 ^M 319.68M 2% 65.06MB/s 0:03:21 ^M 388.10M 2% 65.32MB/s 0:03:19 ^M 456.59M 3% 65.24MB/s 0:03:18 ^M 525.73M 3% 65.47MB/s 0:03:17 ^M 593.82M 4% 65.41MB/s 0:03:16 ^M 662.41M 4% 65.45MB/s 0:03:15 ^M 730.60M 5% 65.38MB/s 0:03:14 ^M 799.15M 5% 65.24MB/s 0:03:13 ^M 867.40M 6% 65.28MB/s 0:03:12 ^M 935.40M 6% 65.12MB/s 0:03:11 ^M 1.00G 7% 65.20MB/s 0:03:10 ^M 1.07G 7% 65.20MB/s 0:03:09 ^M 1.14G 8% 65.20MB/s 0:03:08 ^M 1.21G 8% 65.29MB/s 0:03:07 ^M 1.28G 9% 65.22MB/s 0:03:06 ^M 1.35G 9% 65.20MB/s 0:03:05 ^M 1.41G 10% 65.21MB/s 0:03:04 ^M 1.48G 10% 65.31MB/s 0:03:03 ^M 1.55G 11% 65.34MB/s 0:03:02 ^M 1.62G 11% 65.31MB/s 0:03:01 ^M 1.69G 12% 65.36MB/s 0:02:59 ^M 1.76G 12% 65.27MB/s 0:02:59 ^M 1.81G 13% 62.51MB/s 0:03:06 ^M 1.88G 13% 62.44MB/s 0:03:05 ^M 1.95G 14% 62.34MB/s 0:03:04 ^M 2.02G 14% 62.43MB/s 0:03:03 ^M 2.09G 15% 65.31MB/s 0:02:54 ^M 2.16G 15% 65.38MB/s 0:02:52 ^M 2.22G 16% 65.48MB/s 0:02:51 ^M 2.29G 16% 65.42MB/s 0:02:50 ^M 2.36G 17% 65.32MB/s 0:02:50 ^M 2.43G 17% 65.35MB/s 0:02:48 ^M 2.50G 18% 65.42MB/s 0:02:47 ^M 2.57G 18% 65.49MB/s 0:02:46 ^M 2.63G 19% 65.43MB/s 0:02:45 ^M 2.70G 19% 65.39MB/s 0:02:44 ^M 2.77G 20% 65.31MB/s 0:02:43 ^M 2.84G 20% 65.17MB/s 0:02:43 ^M 2.91G 21% 65.13MB/s 0:02:42 ^M 2.98G 21% 65.06MB/s 0:02:41 ^M 3.04G 22% 64.94MB/s 0:02:40 ^M 3.11G 22% 64.84MB/s 0:02:39 ^M 3.18G 23% 64.74MB/s 0:02:39 ^M 3.25G 23% 64.78MB/s 0:02:38 ^M 3.32G 24% 64.68MB/s 0:02:37 ^M 3.38G 24% 64.66MB/s 0:02:36 ^M 3.45G 25% 64.73MB/s 0:02:35 ^M 3.52G 25% 64.69MB/s 0:02:34 ^M 3.59G 26% 64.77MB/s 0:02:33 ^M 3.65G 26% 64.82MB/s 0:02:31 ^M 3.72G 27% 64.56MB/s 0:02:31 ^M 3.79G 27% 64.16MB/s 0:02:31 ^M 3.86G 28% 64.06MB/s 0:02:30 ^M 3.92G 28% 64.12MB/s 0:02:29 ^M 3.99G 29% 64.06MB/s 0:02:28 ^M 4.06G 29% 64.38MB/s 0:02:26 ^M 4.13G 30% 64.45MB/s 0:02:25 ^M 4.19G 30% 64.45MB/s 0:02:24 ^M 4.26G 31% 64.73MB/s 0:02:22 ^M 4.33G 31% 64.75MB/s 0:02:21 ^M 4.40G 32% 64.88MB/s 0:02:20 ^M 4.46G 32% 64.57MB/s 0:02:20 ^M 4.53G 32% 64.56MB/s 0:02:19 ^M 4.60G 33% 64.67MB/s 0:02:17 ^M 4.67G 33% 64.58MB/s 0:02:17 ^M 4.74G 34% 64.91MB/s 0:02:15 ^M 4.80G 34% 64.98MB/s 0:02:14 ^M 4.87G 35% 64.89MB/s 0:02:13 ^M 4.94G 35% 64.85MB/s 0:02:12 ^M 5.01G 36% 64.78MB/s 0:02:11 ^M 5.07G 36% 64.48MB/s 0:02:11 ^M 5.14G 37% 64.48MB/s 0:02:10 ^M 5.21G 37% 64.08MB/s 0:02:09 ^M 5.28G 38% 64.03MB/s 0:02:08 ^M 5.34G 38% 63.98MB/s 0:02:08 ^M 5.41G 39% 63.96MB/s 0:02:07 ^M 5.48G 39% 64.31MB/s 0:02:05 ^M 5.54G 40% 64.17MB/s 0:02:04 ^M 5.61G 40% 64.35MB/s 0:02:03 ^M 5.68G 41% 64.10MB/s 0:02:02 ^M 5.75G 41% 63.95MB/s 0:02:01 ^M 5.81G 42% 64.01MB/s 0:02:00 ^M 5.88G 42% 64.06MB/s 0:01:59 ^M 5.95G 43% 63.84MB/s 0:01:59 ^M 6.01G 43% 63.81MB/s 0:01:58 ^M 6.08G 44% 63.53MB/s 0:01:57 ^M 6.15G 44% 63.70MB/s 0:01:56 ^M 6.22G 45% 64.25MB/s 0:01:54 ^M 6.28G 45% 64.48MB/s 0:01:52 ^M 6.35G 46% 65.02MB/s 0:01:50 ^M 6.42G 46% 64.88MB/s 0:01:50 ^M 6.49G 47% 64.57MB/s 0:01:49 ^M 6.55G 47% 64.70MB/s 0:01:48 ^M 6.62G 48% 64.27MB/s 0:01:48 ^M 6.
After which it carried on fine.
Any help in debugging and fixing this is appreciated!
This topic was automatically closed 30 days after the last reply. New replies are no longer allowed.