Ubuntu :: Backup Software That Can Clone A Directory?
Mar 28, 2010
I'm looking for software that can backup all the files in my /home directory including hidden files.
I liked Lucky Backup, but it puts everything in a tar file, meaning that the backup fails if the file gets too large (4 GB I think). I would prefer to avoid using tar/archives anyway, as often I only need to recover 1 file from a backup (an archive holding my 50 Gb of data would take ages to open).
Does anyone know of a program or a way to get rsync or the like to copy all the files in a directory, including hidden files, into another directory ( so I end up with effectively a carbon copy of the original). Disk space is not an issue so I don't need to compress anything. I'm not bothered whether its a fancy GUI-based program or a rsync command, just so long as it can save my previous files from.... myself.
I have ubuntu 10.10 installed on 1TB HD. I would like to back up the full hd, on a 2TB usb HD I have. I would like an easy way to clone/backup the full installation, so if my hd fails I would be able to book the usb hd to get files or do a full reinstall. I have tried dd, but not certain if it worked correctly.Also tried clonezilla and that did not work
I have a file server running in an office that's mostly used for file sharing and a scanner saves pdf files to the server. I'm running the latest LTS ubuntu server edition and I really only have ssh installed and samba. My question is that I've done so much to the server as far as premissions and configuration and I'd like to make a clone of this to another computer and not sure how I would do this?
I'm not sure if clonezilla or something like this can perform this task? I basically just have a very old computer and now I have another very old computer that I want to make into a spare just incase something happens to the original. Any recommedations on how I would accomplish this?
I am new Ubuntu. I have a new Dell Mini10 running that I have configured and running. I want to clone or backup the system at it's current state. I have been looking at the options and Clonezilla seems to be recommended often. Are there other choices that are easer or better? My real goal is to make a bootable DVD that will restore my system back to the point it was cloned.
I have a finely working CentOS server. I want to clone the complete OS (over network) so that I can use it for same functionality on several other machines..
i want clone my linux partition for create a backup. i want use dd command but i have some question.my linux partition is 30GB and linux only used 10GB of it if i use dd command for create a image i must have 30GB free space? can i use dd command in X window or i must first exit from linux and use live cd? in fedora i use dd command for create a backup when linux is running but after restore some command like su not work!!!
i use some tools like partimage for make a backup but it show me an error about block 0!!
The place I work has a web/dns server on opensuse 10 up and running hosting a few websites (our company one & a few vhost ones). The box was set up before I got here. Now we are wanting to create a new 11.2 server that is basically a backup / clone of 1st server in case it goes down.
using Back In Time to backup my home directory to a second hdd that is mounted at /media/backupThe trouble is, I can do this using Back In Time (Root), but not using Back In Time without the root option. This is definitely a permissions issue - it can't write to the folder, but when I checked by right clicking on the backup directory and looking at the permission tab, it said I was the owner
This script simply deletes files older than a certain age (in this case 7 days) from a certain location; I use it to purge old backups nightly, and it works as expected:
# delete backups older than 7 days find /mnt/backup/* -mtime +7 -exec rm -Rf {} ;
The problem is, every morning I get an email with an error message something like this:
find: `/mnt/backup/subfolder': No such file or directory
I have a couple of beginner questions;I have checked out some source code in a directory, not inside the home directory, using: Code:git clone git://git.videolan.org/ffmpeg.git ffmpeg-gitI moved the folder around (was not originally in the right directory), and now would like to delete it and start from scratch. How can I do it without creating issues with git? (I don't see ffmpeg-git in Synaptic)More generally, what is the best way to delete old folders related to obsolete or malfunctioning software, or even unsuccessful attempts at installing software ? Can I do 'delete' in the file manager? or just sudo rm <folder name> ? Or is there hidden issues?Another related question: When manually installing packages not available in Synaptic, for example, from downloaded tar.gz, what is the recommended directory structure. For example, is it better to use, for example: /home/<program name>/source to unpack the source code/home/<program name>/build to build the software?as opposed to using folders in the root directory? Is there a tutorial on the subject of best practice for directory structure?
After an upgrade gone wrong my 9.10/10.4 is unbootable and I see no way to fix it.
I can't mount the directory now (LiveCD or chroot), it probably has something to do with the breakage and/or pre-existing ecryptfs issues, as I have both login and mount passphrases.
And it's all on one partition too, so before reinstalling I have to back up. How can I back up the encrypted directory? And can I transplant it to a fresh Ubuntu installation?
cp complains about symlinks, returning 'not permitted', is this normal?
ETA: Now that I look at it from LiveCD .ecryptfs and Private appear as broken symlinks (what do they link to anyway??), but when I tried to copy them while chrooting, they were both accessible directories (though still couldn't mount or copy) - I don't get it.
How do I tar into tar.bz2 a locked backup directory from the simpleconfig backup program? The backup directory is located in /home/lukasz/Downloads/backup/
rsync: link_stat "/av" failed: No such file or directory (2) skipping directory home rsync error: some files/attrs were not transferred (see previous errors) (code 23) at main.c(1060) [sender=3.0.7]
I would like to make a backup of my /home directory onto a NAS device, and have whatever software is used for the purpose update (new and changed files) every night, or perhaps everytime there is a period of inactivity. Any suggestions for a GUI package that will do this?
I do not want a complete backup each time, just the new or changed files. Also prefer software that backs up to a mirror of the original (i.e., uncompressed folders and files)
I need to backup my /home directory because I want to switch from Fedora to OpenSUSE but I didn't put /home as a separate partition so I need to back it up. Problem is, I can't figure out how.I've tried tar and gzip through every google hit I can possibly find but not one has worked.
I have carefully made daily backups using rdiff-backup, so in the case of needing to restore I can do so.
But I deleted a directory yesterday, and made a backup in the evening. Therefore, the directory is not in the latest mirror, but in the incremental backup from yesterday.
Now I need to restore the directory. But I cannot figure out how to!
I can see the directory in yesterday's incremental backup; i.e., the following works:
Code:
Where [backupdir] is the backup (mirror) directory, and [nameofdir] is the name of the directory I'm trying to restore.
So, I have tried to restore. This is the type of thing I have tried:
Code:
Where to-restore.lst holds the name of the directory to restore (in rdiff-backup's format) and [restoredir]is where I want the restored directory to go to.
But, I get errors like:
Code:
Useful file specifications begin with the base directory or some pattern (such as '**') which matches the base directory. Well, obviously the file specification doesn't exist in the [restoredir]. That's because I'm trying to restore it! If I try to create an empty directory first, it complains:
Code:
How do I restore a deleted directory from a previous day's backup to a designated destination?
I would like to have dump backup just my home directory but am having problems the command I am using wants to back every thing and takes hours upon hours it has been running for about 10 hr and only 21% is done. This is the command dump -0u -f dp_hd /media/CENTON USB/ /how can I get this to back up only my home directory
What's a good cron script for backing up and zipping a directory of files, or multiple directories with files, to a backup directory on my server, on a daily basis?I found an easy to use mysql backup script, now I need to backup my site directory, but not all the directories in it. So I need a method in the script to omit certain directories from backing up, ie dirs that contain gigs worth of files.This seems like it should be one of the most common crons to set a server up with but two pages deep in google (and here) I have yet to find anything remotely resembling a solution.
I've got the following code and kludge. It's working okay except that there's a directory called "Archive" which I don't need to backup. I've tried various combinations of &&, -a, etc., and I'm kind of at a loss for getting this right.
Here's the code I have so far:
Code: #!/bin/sh # backup subdirs as individual tarballs with prepended timestamps and # move them somewhere else. # --exclude "/path/to/dir/*" (Doesn't seem to work)
[Code]...
How can I write that line so that I can loop through all of the subdirs, tarring them up, but excluding the "Archive" subdirectory?
So I've finally given up on saving my kubuntu install that wont boot. I've searched, and looked, but couldn't find a thing.My delema now is to make sure that I:
a) get all of the user data safely packed up onto my external USB drive. I believe it's all in the home directory. I'm not sure about getting hidden files though...
b) get the new install to go smoothly, and not mess up grub or the parralell XP install on the same hard drive.
c) get the user data back on the computer and recreate the user structure. Permissions were messed up already, so setting those up again is not an issue.
So, I've been poking around, and this is how I think things should go:
a) tar cvpjf backup.tar.bz2 /home to get my home directory backed up. Not exactly sure how to get from here to my external hd, but I'm sure I can figure it out
b) just run a live cd of kubuntu, delete the old partitions, and reinstall over them?
c) unzip the tar into by /home directory.
That's all I've been able to find so far. How do I set up the users? Will they show up as soon as I untar? Will the resinstall play nice with my windows install? Will I get all the hidden files too? Is there anything I'm missing?
This should be a quick one. I'm trying to backup a single directory and it's subdirectories on my Lucid Server to a freenas box across my network. This is what I'm using to do that..
rsync -r -a -v -z * --delete freenas:dSIBackups It almost works perfectly except for one problem. When a file is deleted at the source, this command doesn't seem to delete it on the receiving end. I assumed that the --delete would do that but aparently not. Can anyone think of a reason that this would happen?
i am in need of linux help. iam at college and i need this back/restore script to pass this final part of an assessment. i require a backup script that will not only backup but also restore files to the relevent directories. e.g. users are instructed to store all wordprocessor files in a directory named wp. so i am needing to create a backup directory and 3 directories within that and some files within the 3 directories and then back them up ot restore them. l know i should/have to do this myself by been trying to get/understand info for the last few days and came up with zero.