Rsnapshot is a software written in Perl to make backup of local and remote file system. The well proven rsync is behind this utility. rsnapshot does not need root user intervention to restore the data of a normal user. It does not take much space in your Backup server. It can be easily automated (scheduled) to make life easier. Just setup once and forget it configuration. Basically it takes snapshot of file system (or a part of) in regular interval such as hourly, daily, weekly and monthly.
This can be configured easily through a simple text based configuration file. The above task can be setup in a few easy steps in a few minutes. Two major tasks are configuring rsnapshot and openssh automatic login. To make the backup automatically, we need to automate the remote login in a secured way. This can be done through openssh tools. This scenario depicts backup of desktop (assuming that IP address is 192.168.0.100) data to a backup server. My desktop runs on Ubuntu 10.04 and backup server runs on Debian Squeeze. [URL]
This is my first post and I am a linux newbie.. took on the challenge of setting up an ubuntu server with proxy and firewall with vpn access as well.. sounds good? While messing with things, I installed ebox server, know called Zentyal and perform backups, however, when performing a full restore, following instructions and all it does not restore correctly. I ran many attempts with verious configurations and still allways got some errors, such as eboc-ebackup failing to start, etc. most noteably wa sthe ldap error, which i figured out how to fix by restarting once logged into and then rebooting.
I say the Clonezilla options for a full backup, and it sounds great, but does require me to perform a full abckup, while bring the server down to perform. What I wanted to know, was what do you recommend for performing a full backup (possibly to NTFS partition or USB key with FAT32) with automation or even if i had to perform the trigger, just without needing to bring the server down, so when i go into production i dont have any issues backing up anytime?
What's the current best practice for automounting a remote drive for automated backup? I want to use Back in Time and maintain snapshots but it can't do that remotely so I have to mount a folder outside of Back in Time. I have used sshfs from this howto in the past and it works mostly but seems to lose connection and not reconnect a lot. [URL]. Is there a more "modern" way? NFS is horribly unreliable and dog slow for me so it's OUT unless it's changed in the last year.
I have created an automated backup using the following:
When cron runs it creates a .tar like it is suppose to, but it creates archive-.tar. However, when I manually type in that command, I get archive-"whateverthedateis". Is there a step that I missed since the date keeps being omitted by this cronjob. I tried a another method
I have a machine on my network and that machine is a mass storage server that I will eventually use as a media server (to stream movies, videoclips and music on my home theater system). I use slackware 13 on ALL of my machines.
I am trying to automate the backup of the "/home" folder of my laptop onto the mass storage server. I currently use rsnapshot and it works great, but I would like to automate the whole process, even if I am not home or in front of my machines...
Here's what I imagined (in pseudo code):
1) Poll if server is active (up); 1.1) If not: 1.1.1) Wake up the server (WOL); 1.1.2) Wait for the server to boot; 1.1.3) Confirm the server has made it to the login prompt (normal boot); 220.127.116.11) If not, send an alarm via email;
I am looking for an automated backup system and I like bacula. I have 3 Notebooks and a Desktop computer that need regular backup. Now I don't want to let them run all night just to do the backuping, so I was thinking I could use wake-on-lan to have bacula wake up the machines, then do the backups, and shut them down afterswards. While this may work with devices on the ethernet, it won't work with the Notebooks on the wifi. So is it possible to have the Notebooks schedules to automatically wake up from suspend or shutdown ? Or is it possible to interject a shutdown command if it is after a cerain hour and call the bacula director to start the backup now?
I am trying to create a backup script that will back up a single folder for a class i am in. I was wandering if I could get some help. If possible I would also like to know how to write a script that can encrypt that same file . I will be putting the back up in my /home/usr/Backup directory. I am not trying to back up my whole system just a single folder. I am using Fedora 11
I have installed an application manager(monitoring application) on my linux server. Now, i need to have backup schedule for my application. The application itself has executive file to backup database.But when i put this file in my crontab to schedule the backup program it wont run!50 09 * * * root /opt/ME/AppManager9/bin/BackupMysqlDB.sh
using Back In Time to backup my home directory to a second hdd that is mounted at /media/backupThe trouble is, I can do this using Back In Time (Root), but not using Back In Time without the root option. This is definitely a permissions issue - it can't write to the folder, but when I checked by right clicking on the backup directory and looking at the permission tab, it said I was the owner
After I spent some time discovering The BIG BANG of Universe and The Meaning of Life :
I managed somehow to create a script to make some backup of files on server and TAR it and then FTP the archive to another location FTP server and then emails result.
It also measures time needed to complete it and deletes archive older than XX days (set in find -mtime +20) and makes incremental backup every weekday and FULL on Sundays (which suits me bcoz no heavy load).
Files for TAR to include and exclude are in txt files listed each line separate name:
I have a scheduled backup to run on our server at work and since the 7/12/09 it has be making 592k files instead of 10Mb files, In mysql-admin (the GUI tool) I have a stored connection for the user 'backup', the user has select and lock rights on the databases being backed up. I have a backup profile called 'backup_regular' and in the third tab along its scheduled to backup at 2 in the morning every week day. If I look at one of the small backup files generated I see the following:
-- MySQL Administrator dump 1.4 -- -- ------------------------------------------------------ -- Server version`
It seems that MySQL can open and write to the file fine, it just can't dump
I've tried to google but not much luck. What I would like to do is have anumber of folders on my desktop and their contents, replicated/duplicated into another folder on the same PC in real time. So for example, if I were to change an OpenOffice document in a specific folder on my Desktop it would be replicated/duplicated in real time. If I had three folders on my Desktop A, B and C they would also appear/be backed up (in real time) in a folder called /home/backup. Can this be done?
does anyone know of a good backup software for Ubuntu 10.4 that will let me select which folders to backup, rather than a complete backup? My install and settings etc can be replaced, but my photos and memories cannot!
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
I am new to Linux. I am using tar to backup my emails to a server. I would like to automate this process to routinely backup my emails periodicly, however, i keep running into a problem: I start in the dir I would like to create the tar file (dir size = 240MB). I enter the following command
tar cf bup.mail.llc.tar "/Users/d/Library/Mail/INBOX.mbox/Messages" file size = 234MB
When I would like to backup my emails into the previously created tar file I use the following command:
tar uf bup.mail.llc.tar "/Users/d/Library/Mail/INBOX.mbox/Messages"file size = 462MB The backup command works, except the size of the original tar file grows, around twice the size. When I extract the updated tar file (file size = 462MB), the unarchived file is 240MB the same size as the original directory.
Why does the size of the tar keep growing each time i perform 'tar uf'?? I don't understand this
O/S: Fedora 12 I am newbie in linux. What I want to do is: Make backup for my file system, cos I learn how to configure servers. So if I made some thing wrong, I want to be able to restore the default setting for my files. Instated of install new O/S.
I did a backup of the ssd on my eeepc using the following command from a Linux Mint on a USB key: dd if=/dev/sda1 of=/media/disk/eeepc_save/SYSTEM/system.bck (/media/disk in an external USB disk)
I deleted the ext2 partition using GPartEd on live USB key and created it back. I rebooted Linux Mint and restored the filesystem using the opposite command : dd if=/media/disk/eeepc_save/SYSTEM/system.bck of=/dev/sda1
I mounted /dev/sda1 and when I "ls" the root directory, I get several "NFS stale file handle" messages concerning directories (/dev and other). I tried "e2fsck -y", had a bundle of corrections that resulted in the deletion of the directories. I don't use NFS. I did the same for the user filesystem and had no problem (it's an ext3 partition). The two filesystems are the ones that came with the original Xandros installed on my eeepc and that were mounted with union-fs.
I have use backup-manager tool very often but now found some problems and goes to it official site backup-manager.org for search answers.But this site didn't open already more than week! url url url
At now DNS records don't have A record:
Code: $host backup-manager.org backup-manager.org mail is handled by 10 private.sukria.net. backup-manager.org mail is handled by 15 private.nxr.fr. backup-manager.org mail is handled by 5 mx.sukria.net. backup-manager.org mail is handled by 10 jupiter.unix-scripts.info. Does this project moved, renamed or died? Maybe it change the domain address? Or this is only temporary problems with hosting or domain?
I work for a school consulting company.We helped a school deploy about 1500 computers.The computers have windows XP but we have been using G4L for the restore partition on the drives.So far the software works great. We did however run into a problem in that many of the computers we deployed are missing the restore partition. The reason they are missing is long and convoluted and not really that important. What I have been charged to do is try and fix the restore partition problem. One solution that I had, which im not even sure if it will work, was to backup the recovery file, that g4l created, to DVD and write a basic script to recreate the partition and then copy the file over. This process would need to be as automated as possible since this disc will be inserted by the end user(the students). The backup file that g4l created is 5.9GB so it wont fit on just one disc and Dual layer discs are too expensive to use for this project, so the file will either need to be compressed again (not sure if that's a good idea or not) or split across two DVD's.
I have searched the forums here and I was not able to find anything to fix this problem. I was able to find some info on splitting files across two discs but im not sure how to use that to fix my problem.