[Beowulf] (no subject)
    Mark Hahn 
    hahn at mcmaster.ca
       
    Fri Feb 16 14:17:57 PST 2007
    
    
  
> not buy a tape drive for backups.  Instead, I've got a jury-rigged backup
tapes suck.  I acknowlege that this is partly a matter of taste,
experience and history, but they really do have some undesirable properties.
> scheme.  The node that serves the home directories via NFS runs a nightly tar 
> job (through cron),
> 		root at server> tar cf home_backup.tar ./home
> 	root at server> mv home_backup.tar /data/backups/
>
> where /data/backups is a folder that's shared (via NFS) across the cluster. 
> The actual backup then occurs when the other machines in the cluster (via 
> cron) copy home_backup.tar to a private (root-access-only) local directory.
>
> 	root at client> cp /mnt/server-data/backups/home_backup.tar 
> /private_data/
>
> where "/mnt/server-data/backups/" is where the server's "/data/backups/" is 
> mounted, and where /private_data/ is a folder on client's local disk.
did you consider just doing something like:
 	root at client> ssh -i backupkey tar cf - /home | \
 		gzip > /private_data/home_backup.`date +%a`.gz
I find that /home contents tend to be compressible, and I particularly
like fewer "moving parts".  using single-use ssh keys is also a nice trick.
> large (~4GB).  When I try the cp command on client, only 142MB of the 4.2GB 
> is copied over (this is repeatable - not a random error, and always about 
> 142MB).
might it actually be be sizeof(tar)-2^32?  that is, someone's using a u32
for a file size or offset?  this sort of thing was pretty common years ago.
(isn't scientific linux extremely "stable" in the sense of "old versions"?)
> only some of the file be copied over?  Is there a limit on the size of files 
> which can be transferred via NFS?  There's certainly sufficient space on disk
it's certainly true that old enough NFS had 4GB problems, as well as
similar vintage user-space tools.
    
    
More information about the Beowulf
mailing list