On Tue, 20 Feb 2007, Les Mikesell wrote:
You might want to look at backuppc (http://backuppc.sourceforge.net/) to
manage the on-line backups. It uses a scheme of compression and hard-linking
duplicate files to greatly reduce the storage space you need when keeping a
long history - especially if many files are unchanged over that time. With
something that size you might still have to play some tricks with snapshots
or break it down into smaller directory runs. It uses tar, smb, or rsync to
do the actual copies, then has some extra overhead for the compression and
linking jobs. If it is too slow, you might run it against your existing
uncompressed copy so it has all day to make the long-term archive version.
Interesting, especially for the Win side of things, thanks!
However one of the things I may need is to use the backup server as a
"live" server, should the original server ever catch fire (or whatever),
so pull the backup from the bunker, drive it across town and with minimal
changes, have it take the place of the live server (turn on nfs/samba on
the appropriate locations) until we get a new one installed... So it's
compression would be of minimal use, but it may have other uses...
Thanks,
Gordon
_______________________________________________
linux-lvm mailing list
linux-lvm@redhat.com
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/