I have an extensive DVD collection, and I want them backed up in case of a fire or other problem. I am trying to create a script to automatically prompts me for the name of the DVD and then backs the dvd up when I insert it. I am running kubuntu 10.10. When I run this in a terminal, nothing happens. Do you have any idea what I did wrong here?
I take responsibility for data backup my centos server. Please help me backup data using tar. following scenario:- Sunday - full backup- the rest of differential backup.We're back up to full as follows:tar-cvzf / backup / backup-date.tar.gz / data
I have requirement on shell scriptingre scripts for the following: 1) Full backup(application and DB): tar all the files and database and copy them in the backup server 2) Incremental back up of files: backup of all the updated filester the last backup. 3) Restoring of backup files: restore the backup files from the backup server to the application server.
I have a script that I'm trying to use to backup my hard drive every night. The full backups are working flawlessly. I haven't figured out the incrementals yet. I thought they were working when I first wrote the script a few weeks ago, but it doesn't seem to be working anymore. I have a long script that I have a whole bunch of statements like this.
[Code]...
To be honest, I'm not a Linux expert and I was really surprised when I started this venture that Linux filing system doesn't have an archive bit like NTFS does and we have to use a file to store the archiving information. But there's the file at /var/log/usr.snar. My theory is I just leave that file for a week so I can create one full and incrementals through the week. Then when I want to create another full backup (when I swap hard drives), I then delete that usr.snar file. I thought that worked a few weeks ago, but it doesn't anymore. When I start my backups, they backup every folder I have listed and not just the files that changed. My full backups are 165 GB, so I'd like to use the incremental feature.
I need a script that can undo a complete backed-up file system restructuring that took place after an upgrade to the latest version of SyncBackPro. Basically a windows machine running SyncBackPro runs a backup every morning to a SuSE 11.2 SFTP file server and after the monthly update of SyncBack I had some issues getting the existing profiles to run. Suffice it to say I did some tinkering and thought I reset the backup paths to their correct locale but the next day I found that the program took everything in the parent SFTP user directory on the SuSE server and did the backup in the main user directory.
Since the user directory is not the backup directory, SyncBackPro took all of the files in the normal user directories and moved them into versioned $SBV$ directories and inserted a unique 12-digit numerical value to the front of every file below ~/ in every directory (recursively). I found a program named krename to correct the renaming issue and it is still adding the ~3 million files to the rename structure (tests on smaller folders worked flawlessly) to correct the 100*********.'s that have been appended to the beginning of every file. My problem is that I need to take all files that have been moved into these $SBV$ directories and move them up one level to their original directory and then remove the $SBV$ directory after the move. I do have a little experience with bash/shell scripts as I have made and use them for extracting, timestamping and moving jobs that execute through TORQUE and G03 but fall far short of having a programming degree. Chemistry, now that's a different story...I have found another post which appears to move the files in a specified directory up one level but I need a solution that can search recursively for the $SBV$ folder and then move the files and delete the then empty $SBV$ folder. [URL]
I've got the following code and kludge. It's working okay except that there's a directory called "Archive" which I don't need to backup. I've tried various combinations of &&, -a, etc., and I'm kind of at a loss for getting this right.
Here's the code I have so far:
Code: #!/bin/sh # backup subdirs as individual tarballs with prepended timestamps and # move them somewhere else. # --exclude "/path/to/dir/*" (Doesn't seem to work)
[Code]...
How can I write that line so that I can loop through all of the subdirs, tarring them up, but excluding the "Archive" subdirectory?
I would like to develop a shell script that would automatically run a backup program and then shutdown the computers, but I do not know the first thing about script development or even writing scripting.
I have a shell script that need to create some files:
1) backup files of user passed in file ( that will be written by this shell ).
2) temp files that the shell will create and later delete/remove.
This shell script will be used from my local dir ( I am not a super or a sysadmin ). Users of this shell will call this script to run on their local files in their respective directories. When my script runs, it errors with the following:
for the temp I was able to avoid this error by creating the temp file in the /tmp directory. All I want is for this shell script to run, create/modify/backup files in user's local dirs.
I would like to create a bash menu script for my home server For instance if i were to type ./script It would then bring up 3 options
a. Create a backup b. Restore files from a backup c. Quit
If you were to select a or b it should then ask you were you want to backup or restore from. And if i were to type in an incorrect letter i should get an error and take me back to menu. I have attepmted this a view time now and have magaged to get the menu up using parameters
using Back In Time to backup my home directory to a second hdd that is mounted at /media/backupThe trouble is, I can do this using Back In Time (Root), but not using Back In Time without the root option. This is definitely a permissions issue - it can't write to the folder, but when I checked by right clicking on the backup directory and looking at the permission tab, it said I was the owner
After I spent some time discovering The BIG BANG of Universe and The Meaning of Life :
I managed somehow to create a script to make some backup of files on server and TAR it and then FTP the archive to another location FTP server and then emails result.
It also measures time needed to complete it and deletes archive older than XX days (set in find -mtime +20) and makes incremental backup every weekday and FULL on Sundays (which suits me bcoz no heavy load).
Files for TAR to include and exclude are in txt files listed each line separate name:
This script simply deletes files older than a certain age (in this case 7 days) from a certain location; I use it to purge old backups nightly, and it works as expected:
# delete backups older than 7 days find /mnt/backup/* -mtime +7 -exec rm -Rf {} ;
The problem is, every morning I get an email with an error message something like this:
find: `/mnt/backup/subfolder': No such file or directory
I have a scheduled backup to run on our server at work and since the 7/12/09 it has be making 592k files instead of 10Mb files, In mysql-admin (the GUI tool) I have a stored connection for the user 'backup', the user has select and lock rights on the databases being backed up. I have a backup profile called 'backup_regular' and in the third tab along its scheduled to backup at 2 in the morning every week day. If I look at one of the small backup files generated I see the following:
Code:
-- MySQL Administrator dump 1.4 -- -- ------------------------------------------------------ -- Server version`
[code]....
It seems that MySQL can open and write to the file fine, it just can't dump
I've tried to google but not much luck. What I would like to do is have anumber of folders on my desktop and their contents, replicated/duplicated into another folder on the same PC in real time. So for example, if I were to change an OpenOffice document in a specific folder on my Desktop it would be replicated/duplicated in real time. If I had three folders on my Desktop A, B and C they would also appear/be backed up (in real time) in a folder called /home/backup. Can this be done?
I am trying to create a backup script that will back up a single folder for a class i am in. I was wandering if I could get some help. If possible I would also like to know how to write a script that can encrypt that same file . I will be putting the back up in my /home/usr/Backup directory. I am not trying to back up my whole system just a single folder. I am using Fedora 11
I'm just setting up a partition on a seperate HDD in my system. I plan to use the partition to backup the important files on my main HDD (to guard against HD crash).
The question I have is about where would be the typical location to auto mount this partition? Which would it be normal to go for:
does anyone know of a good backup software for Ubuntu 10.4 that will let me select which folders to backup, rather than a complete backup? My install and settings etc can be replaced, but my photos and memories cannot!
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
I am now preparing myself to upgrade lenny to squeeze and decided to do a backup on my system. I used backup-manager to do the job and it worked fine. how do you restore said backup data?
I have use backup-manager tool very often but now found some problems and goes to it official site backup-manager.org for search answers.But this site didn't open already more than week! url url url
At now DNS records don't have A record:
Code: $host backup-manager.org backup-manager.org mail is handled by 10 private.sukria.net. backup-manager.org mail is handled by 15 private.nxr.fr. backup-manager.org mail is handled by 5 mx.sukria.net. backup-manager.org mail is handled by 10 jupiter.unix-scripts.info. Does this project moved, renamed or died? Maybe it change the domain address? Or this is only temporary problems with hosting or domain?
I have installed an application manager(monitoring application) on my linux server. Now, i need to have backup schedule for my application. The application itself has executive file to backup database.But when i put this file in my crontab to schedule the backup program it wont run!50 09 * * * root /opt/ME/AppManager9/bin/BackupMysqlDB.sh
I create Backup partition with dd and save my backup on NTFS partition. i want recover this backup,what should i do? can i use boot cd and use dd command for recover it?