slightly OT - backup strategy
Mark E. Fuller
mark.e.fuller at gmail.com
Wed Jul 20 13:24:25 IDT 2022
20 Jul 2022 13:15:13 Rabin Yasharzadehe <rabin at rabin.io>:
> Using ZFS with sanoid[https://github.com/jimsalterjrs/sanoid]
>
> ZFS will give you all the benefits of COW filesystem, compression, and snapshots (and much more),
> combined with sanoid utility will allow you to automate the snapshots and send them to a remote system,
> and because ZFS is block-level aware of the changes between snapshots, send&recive is much more efficient,
> because unlike Rsync which needs to stats and compare each file to determine what to sync,
> ZFS only need to compile a list of block which have changed between 2 snapshots and send only them.
> which also works if the volume is encrypted, which allows you to have a remote system, which is encrypted on rest,
> and keep pushing/sending snapshots to it without having to unlock it.
>
>
> --
> Rabin
>
>
> On Sun, 17 Jul 2022 at 16:50, Shlomo Solomon <shlomo.solomon at gmail.com> wrote:
>> I recently lost some files because of a bad disk - hardware problem.
>>
>> I do regular backups so I was not really worried, but I now see that I
>> have a problem with my backup strategy so I'd like to know how others
>> handle/prevent what happened to me.
>>
>> I backup files using rsync and I basically have 2 types of backups.
>>
>> My most important files are backed up every night. I do incremental
>> backups using: rsync -aqrlvtogS --ignore-errors --backup
>> I keep about 4 months of backups. So if a file is damaged,
>> missing or accidentally deleted, I can find a good file - even if, for
>> example I screwed up the file and only discovered the problem a few
>> days later.
>>
>> BUT, all the rest of my files - music, videos, pictures, etc are backed
>> up daily and weekly on 2 different physical drives using:
>> rsync -qrlvtogS --delete --ignore-errors
>> I use --delete to prevent accumulating garbage on my backup disks.
>>
>> So here's the problem: Because of a hardware problem, several files on
>> one of my disks were lost. As a result, the daily backup script
>> "thought" that those files should be deleted from the daily backup.
>> Unfortunately, I did not notice the problem. A few days later, those
>> same files were also deleted from the weekly backup. So they are lost.
>>
>> So on one hand, I need --delete to avoid keeping backups of old
>> garbage, but on the other hand, the --delete option does not know if I
>> deleted the file or if it's gone because of a hardware problem.
>>
>>
>>
>> --
>> Shlomo Solomon
>> http://the-solomons.net
>> Claws Mail 3.17.5 - KDE Plasma 5.18.5 - Kubuntu 20.04
>>
>> _______________________________________________
>> Linux-il mailing list
>> Linux-il at cs.huji.ac.il
>> http://mailman.cs.huji.ac.il/mailman/listinfo/linux-il
Coming to this late, I use "backintime" which is a Python wrapper and GUI for rsync.
One of the settings, "smart remove" is set to remove old snapshots, but keep some for exactly the kind of problem described here, e.g.
2/day for the last week, then 1/week for a month, then 1/month for the previous year, etc. (you set all these yourself)
I made a similar deletion error once which propagated through my Dropbox that I only noticed after I was past the 30 day backup there, but I easily pulled the files off my backup HDD from an old image
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.cs.huji.ac.il/pipermail/linux-il/attachments/20220720/53e3e49c/attachment.html>
More information about the Linux-il
mailing list