CentOS 5 :: Backup Script With TAR - Incremental Backup With Simple FTP To Another Location And Email Status
Jan 15, 2010
After I spent some time discovering The BIG BANG of Universe and The Meaning of Life :
I managed somehow to create a script to make some backup of files on server and TAR it and then FTP the archive to another location FTP server and then emails result.
It also measures time needed to complete it and deletes archive older than XX days (set in find -mtime +20) and makes incremental backup every weekday and FULL on Sundays (which suits me bcoz no heavy load).
Files for TAR to include and exclude are in txt files listed each line separate name:
file: including.txt:
View 7 Replies
ADVERTISEMENT
Sep 15, 2009
I am trying to create a backup script that will back up a single folder for a class i am in. I was wandering if I could get some help. If possible I would also like to know how to write a script that can encrypt that same file . I will be putting the back up in my /home/usr/Backup directory. I am not trying to back up my whole system just a single folder. I am using Fedora 11
View 2 Replies
View Related
May 10, 2011
I install and tested Restore EE Backup server on a test PC with basic configuration and its working fine.
[URL]
The issue i have is where is the location these backup snapshots or files are saving? I want to add a separate Storage to save the backup?
View 1 Replies
View Related
Aug 10, 2011
I want to make a backup from my Email and my Favorites from Mozilla.
But which folders I have to make a backup from.
View 3 Replies
View Related
Jul 18, 2011
I am trying to find a backup program to incrementally backup some files to an external disk every week for example. I would prefer not to have to write a script as I am not really used to it.
View 4 Replies
View Related
Mar 16, 2011
I'm trying to take backup for my data for rhel, but I not able to take all backup. Could anybody help show me how I take incremental and full backups? What is the process?
View 3 Replies
View Related
Oct 20, 2010
when rsync is finished the update, or in the meantime - i need to move the updated files to a different location - like date +%Y%m%d something or what ..the reason is, because of the development, i need the modified files, but all of them, not just the last one - so i have to store them daily, but i dont want to store the whole dir - just that few files which are updated does it make sense?
View 5 Replies
View Related
Mar 8, 2011
incremental backup of folder.The problem with e.g. find&tar is, that I want backup not only files with modification time after x:y , but also older files, that have been copied into this folder after last backup.
View 6 Replies
View Related
Sep 18, 2010
A complete back up using tar takes consumes more time. so is there any way to take incremental backups using tar.And i also want to take incremental backup dump of my databases too.Any suggestions and links will be very helpful.i keep on googling for this,but could find any exact for this.
View 14 Replies
View Related
May 7, 2009
I had full backup in mysql. now i added some tables .i got new binary logs. how i can i use these logs to have incremental backup.
View 3 Replies
View Related
Sep 17, 2010
I want to backup data and upload to online hosting services.Since I'm uploading stuff online, I only want to upload encrypted data (so that even the hostiing service admins cannot look at the data).Thus, I first want to encyrpt my data locally that I want to backup. Since I will be making changes locally to the data, I want some sort of incremental imaging system where the incremental changes are stored in seperate files so that I only have to upload the incremental encrypted changes.
Duplicity is an option, but it uses GPG, which makes it a bit complicated; and I was wondering if there was any alternative which was simpler as I am only doing the encryption and backup locally.
View 1 Replies
View Related
Apr 3, 2010
I am trying to write as bash script in order to have backup done by cron on the webhosting server. I want all backup runs to have incremental number in front of them. I came up with an idea to store incremental number of a backup in txt file on a server so when next one runs is can check the number in the file. However I am having terrible issues. My script:
[code]....
View 7 Replies
View Related
Mar 15, 2011
I am currently using a script to backup my Ubuntu 10.04.1 system. The mySQL databases are backed up separately from the the system / data.
My problem is with the mySQL incremental / binary log backups.
The problem is that the binary log file(s) are always named xxxx-bin.1.
Up to about a month ago the binary logs were named xxxx-bin.000001, xxxx-bin.000002, etc.
I did make some changes at about the time that this change in file naming ocurred, but I can not identify what, if any, setting I may have changed that has caused all of the binary log files to always have the same name.
My back up script uses both mysqldump and mysqladmin flush-logs to create the binary logs.
All of the setting for mysqldump and mysqladmin are contained in the my.cnf file.
The my.cnf file contents that are relavent are as follows:
Code:
The statements in the backup script that do the backup are:
mysqladmin flush-logs
or
mysqldump | gzip > $DB_BACKUP_DIR/$ARCHIVE_FILE #Note: delete-master-logs in my.cnf
View 3 Replies
View Related
Aug 10, 2010
I want to backup data and upload to online hosting services. I first want to encyrpt my data locally that I want to backup. Since I will be making changes locally to the data, I want some sort of incremental imaging system where the incremental changes are stored in seperate files so that I only have to upload the incremental encrypted changes. Duplicity is an option, but it uses GPG, which makes it a bit complicated; and I was wondering if there was any alternative which was simpler as I am only doing the encryption and backup locally.
EDIT:I have only ONE computer on which the data resides, and on which the backup image image is made. That is, I have a directory foo on my computer, the backup of which will be made to back-foo on the same computer. I want back-foo to be in an encypted form Then back-foo will be uploaded (unencrypted) to microsft live storage or to spideroak storage etc. Since back-foo is encrypted, my upload is secure. And since I'm uploading, I want incremental backup support, that is, the backup utility should create new files which contain the incremental changes so that I can upload only the new files which contain the changes.
View 2 Replies
View Related
May 4, 2010
i am very sorry if this has been asked before... i'm sure it has.. but i have searched all over the net looking for an answer and i still cant find it...
I have a really simple cron job script like this:
When i run this manually it works fine but when i run it from my ROOT user in Plesk as a cron task is always creates a file that is just 45 bytes. Why doesn't it work... I am running it as a root user.. so surely i must have permission to access the file?
View 7 Replies
View Related
May 23, 2010
I have a second HD in my computer for doing a backup onto. All I want is to select a couple of my documents folders and have them copied to it daily. I would like it to add new files and update newly modified ones. I don't want it encrypted or archived, just a basic copy.
View 4 Replies
View Related
May 4, 2011
I use Simple Backup to backup all my machines across my home network. I have just upgraded my test system to 11.04 with a clean install and want to restore all mt stuff from my backup, however, looking in the software centre Simple Backup no longer exists..
View 5 Replies
View Related
Apr 22, 2010
I installed Simple Backup Suite [URL] on this Karmic box last week and it ran fine the first time. Now it locks up shortly after I tell it to run a backup. ps shows
Code:
3273 ? 00:00:01 gksu
3274 ? 00:00:01 simple-backup-c
3279 ? 00:00:02 sbackupd <defunct>
3287 pts/0 00:00:00 ps
I re-installed the application, which didn't help.
View 5 Replies
View Related
Jun 10, 2010
I'm trying to set up a simple backup script with cron.
In "crontab -e" (and sudo crontab-e - I tried both) I enter "0 22 * * * /home/USERNAME/.backup.sh", with the hope that it will run the script at 10pm each day. The srcipt work fine if I run in a terminal. why it won't work? It's bound to be something obvious....
View 5 Replies
View Related
Jun 3, 2011
uBuntu 10.10, Simple Backup My backup stopped working at certain date. Went back in "Simple backup-Configuration", it says, there is no configuration. What might cause trouble like this? Should I seek for more secure back up tool?
View 2 Replies
View Related
Feb 20, 2010
I am somewhat new to Linux and I am looking for a way to back up my HD with all my Linux files. I have a Toshiba laptop running Windows 7. The HD has been partitioned so that the computer can run Red Hat Scientific Linux. Using Grub I can dual boot to either Windows 7 or Linux on start up. I want a simple way of backing up the entire contents of my HD (both partitions - everything) - so that in the event of my laptop being damaged I can reconstruct my set up and data as before with all my files and settings in both Windows 7 and Linux intact. Is there a simple program that will enable me to copy everything to an external HD for back up. Can anyone recommend a package that will do this?
View 5 Replies
View Related
Nov 23, 2009
I have my webserver with cpanel , and I have a FTP location. I wanna backup to the FTP location directly, because I dont have enough space on the server to back up first then rsync over the data, I have this script I need a little help in getting it to backup directly to the ftp location , or is there a simpler way .
Code:
#!/bin/sh
# System + MySQL backup script
# Full backup day - Sun (rest of the day do incremental backup)
# Copyright (c) 2005-2006 nixCraft <url>
# This script is licensed under GNU GPL version 2.0 or above
code....
View 15 Replies
View Related
Jan 26, 2010
how to that covers a simple LAN NFS backup server? One that I only boot to backup from three Linux-only boxes then turn off.
View 9 Replies
View Related
Aug 14, 2010
Simple Backup no longer works. (Says backup is started in the background... then nothing) What are some good comparable alternatives?
View 3 Replies
View Related
May 16, 2011
Here's what I want to do: Copy the whole Ubuntu 10.04 partition/installation from my old laptop to the new one.
What I tried: Used Simple backup to back up my Ubuntu installation to a USB hard drive. It yields a 10.4 gig folder containing 7 files. Installed 10.04 on the new laptop, used Synaptic to install Simple Backup, plugged in USB drive, started Simple Backup Restore, tagged the backup directory in Simple Backup Restore, and get the error:
Error: no backups found in the target directory.
I also tried copying the backup to the local drive, same difference.
View 7 Replies
View Related
Jan 20, 2011
I am using Cron for nightly backups to a usb device. I was just wondering in my script for the backup, how do I find the location of my usb device.
View 2 Replies
View Related
Jul 17, 2010
I'm a beginner at backing up my Ubuntu system, but I've set Simple Backup to do a backup once a week. I deleted the oldest of these files, but now it's sitting in my Trash and I can't empty it. I get a permission denied error for the folders within the backup folder in the trash, yet I can't restore the folder either - Ubuntu says it 'failed to determine the original path' for the folder. I've just discovered this in Xubuntu Jaunty, but I'm confident the same will happen in any other WM I choose (I have several installed - I like variety ).
It's not a huge file, but it's hanging out there and I'd like to get it either deleted or restored. Possibly I oughtn't to have deleted it in the first place (it usually lives in /var/backup, which I can't access except as root). The files, which I probably deleted /as/ root, show up in my user trash rather than root's trash. I found the trash in ~/.local/share/Trash/files, but I'm not sure if just deleting them as root would be a good.
View 1 Replies
View Related
May 27, 2009
I have a mount called on /home for /dev/sda12..I want to mount /dev/sda12 onto /backup..I tried to do this by changing things in the fstab file i.e. i replaced /home with /backup. This change caused boot up problems and I had to change my fstab file back to get going again.
View 2 Replies
View Related
Dec 7, 2010
i am beginner to Linux, i have svn server installed in redhat 5 till now every thing is working fine now i need help from all of you to set up svn backup with email authentication. i need conformation of when it is backup and is the backup don perfect or not. can anyone give me guidelines how to setup (Step by step)
View 2 Replies
View Related
Apr 14, 2011
One of my clients needs a backup of his svn repository. I see that this is possible using svadmin dump command. I see where the location of the source repository is, but I don't see anything in documentation as to where the actual dump file is located. I need to know where the dump file is so I can scp or rsync the file to another server for backup.
View 5 Replies
View Related