
On 2023-06-04 09:48, Lennart Sorensen via talk wrote:
Personally I use rshapshot for backups with a target being a linux server at my parents house and then they backup to mine the same way.
No cloud providers, nothing complicated, it just works and it's automatically offsite.
Of course it's great for making a backup of your data, it is not for making a system backup should you need to restore the system, but I don't consider that to my a big task in general. I also tend to use at least RAID1.
I found that rsnapshot does a lot of filesystem churn on the target system, which can start to be an issue with data that has a huge number of small files that wanted hourly snapshots all hitting RAID6. It will do an rm -rf of older snapshots, resulting in each directory having to be cleaned out entry-by-entry, all the files having their link counts decremented, lots and lots of inode activity, and then it builds new snapshots and has to increment link counts and build new directories. The solution I found was to recycle older snapshots, letting rsync bring a recently-retired snapshot up to date making only the new changes before calling that the current snapshot. For absolutely vital data there's also a lot to be said for keeping it in a git repo to track changes and being able to revert damage, and then git push of a packfile can be your backup. There's still a lot to be said for tarball backups, since they hit backup target storage as a single coherent file and don't do the disk thrashing. Nowadays something in a squashfs image could let you mount that and copy out individual files without having to restore the whole tarball, so that's an interesting direction too. Anthony