General :: Difference Between Archiving And Backup Of Data?
Feb 24, 2010I am a little confused in distinguishing Backup, compressing and Archiving data. Help me to figure out how these can be useful.
View 2 RepliesI am a little confused in distinguishing Backup, compressing and Archiving data. Help me to figure out how these can be useful.
View 2 Repliesi am using red hat linux 2.4 . I have 3 folders dir1 dir2 dir3 I have tarred them like this.
1.tar cvfz tarball_1.tgz dir1 dir2 dir3
2.tar cvfz tarball_2.tgz dir1 dir2 dir3 2>& /dev/null (So that it does not display any error message or operation details to the user)
[usr@machine]$ ls -lrt
-rw-r--r-- 1 usr grp 199843988 May 17 13:39 tarball_1.tgz
-rw-r--r-- 1 usr grp 199837488 May 17 13:53 tarball_2.tgz
But can any one explain the size difference as seen in list output...
I want to create retrievable archives of a my old emails say monthly to avoid the old emails running me out of memory. If I use the 'Backup Settings' procedure as explained in Evolution Help what happens when I wish to consult the /home/dbus/evolution-backup.tar.gz archive file?
1- Will it simply over-write my current Evolution data? [in which case its not what I need]
2- If not, how do I return the archive file to dead storeage and resuscitate my current data?
3- If it will overwrite using the 'Restore Evolution' procedure given in Evolution Help is there a workaround ... perhaps by ...
3a- renaming the archive file,
3b- or 'restoring' it in another version of Evolution,
3c- or archiving CURRENT data as a 2nd backup with a different name [eg: /home/dbus/evolution-backupJan11.tar.gz?] then restoring that?
4- Will I be able to retrieve successive archives if I rename them, say '/home/dbus/evolution-backup.tarDec10.gz' etc once Evolution's saved them?
Alternatively the following came from a dead thread [from commonlyUNIQU3 ] ... is it still valid? does it avoid the problem of potentially running out of memory?
A. Make a subfolder to the "Inbox" under the "On This Computer" tree (I call mine "Archive")
B. Drag and drop the emails you want to archive into this folder.*
This will move the selected emails off of your Exchange server/account and into this folder (and into local storage) - unless you do a copy & paste instead of drag & drop. You may need to have the setting for downloading emails for offline access enabled for this to work as desired. If I recall, the new integrated backup feature creates a compressed (.zip) archive from which you can later restore the email (haven't tried that just yet).
I have a problem with a script i wrote, the script runs fine if manually executed however it doesn't run *fully* when executed via cron
here's the script :
Code:
#!/bin/bash
FILENAME=mysql_full_dump_`date '+%m.%d.%y'`.sql
`which mysqldump` --all-databases -uroot -p************ -h127.0.0.1 > /root/$FILENAME
RETVAL=$?
[code]....
the script resides in /root/bin and the cron entry is as follows:
Code:
0 0 * * * root "/root/bin/mysql_daily.sh"
the result is the .sql file, but it doesn't archive it.
What are the options to backup data to CD/DVD from command line interface. Don't want to use K3B and Nero Linux. Want to backup using "tar" like utility. Please suggest good articles or options.If we can use tar itself to backup data to CD. Then how?For example I want to backup my /home directory to a DVD. I don't want to use GUI tools for this.
View 2 Replies View RelatedI am now preparing myself to upgrade lenny to squeeze and decided to do a backup on my system. I used backup-manager to do the job and it worked fine. how do you restore said backup data?
View 4 Replies View RelatedHow to backup my Linux server for preserving data in case if hard disk crashes?
View 2 Replies View RelatedI have couple hard drives from different linux distro's. on the hdd's are logical volumes created. what i want to do is to backup the data from those hdd's. but when I connect some of them and boot up the system it
gives me an error that there are volume group duplicated and it list the UUID. My host system is centos 5. What is the best solution to rename those volume groups or to mount the drives from other systems, so the data will not get lost?
My hard disk is failing and I amnot able to boot into the system! Currently I have logged into the system uing Live CD! Any way to compress and back up the data in my hard disk in an efficient way!
View 6 Replies View Relatedi setup the following script ro run each night (12am) as a cron on the main server:
if mount | grep -q '/home';
then
rsync -ranv --delete /home/ 138.73.56.12:/home;
[code]....
the main server is running on a dell poweredge 2600. The script rsyncs to a virtualised duplicate running on a hp dl380. when i set this script up and begn running data started going missing from the main server. if new files had been created the files by staff they would go missing, if data was added to existing files prior to activating the script the new changes made to that file would be lost.i just cant understand why this happened. as soon as i turned the script off after a few days it was all back to normal but the data that had gone missing had gone missing.
i just wanted to know if this could be a disk read/write issue, was the script running too soon not allowing data to be written before it can be backed up? could it bee memory? i just dont know. another developed occured after a few days of all this happening, which was one of our hard disk from the main server started mis behaving and flashing amber (attention).
I'm relatively new to Linux in general but have learned to do the basics with the CLI.
Well my main problem is writing my first "real" script in VIM. I just have no idea where to start. I was hoping you guys could point me in the right direction.
Well this is what the script needs to do.
"As the IT administrator for a large manufacturing company you have been tasked with producing a script for archiving the daily log files. Each week the daily log files are placed in a directory named weekXX, where XX is the number of the week. Each week directory should be archived and compressed and stored in a folder called log_directories. When the script has completed the task it should display on the monitor exactly what files and directories it has archived.
The script should be started by a user with the required week numbers added as arguments (e.g prog 13 14 15 should start the program and archive the daily log files in week13, week14 and week15).
A basic manual is required, showing how to start the program, the hardware and / or software requirements, a hard copy of the script and a brief description of the test strategy and test data."
I need to tar this logs, but i dont how to make it simplier to me. Everyday there are created this five logs. I need to make five tar files from every day from this files at the end of the month
For example
Till now i have tar it manualy (copied every file)
Which are the Open Source "file and email archiving" software for both Linux and Windows equivalent to Enterprise Vault Symantec?
View 2 Replies View RelatedI started unarchiving a RAR file that this several gigabytes big. The computer is now going really slow, it's almost frozen. Sometimes I can move the mouse a little, but that's it. The unarchiving process seems to have halted, so now all I can do is restart the system. I don't think I can unarchive this file in Linux.
View 5 Replies View RelatedI need to back up a fold on a remote machine to my local box; the remote hd does not have enough space archive it, neither does my local box. I know there's a cantrip to pipe scp through gzip (or similar), but I don't remember the syntax.
View 1 Replies View RelatedI have a script which periodically backs up a directory using the command "tar -czvf [name] [directory]" but my problem is that the script has recently been putting a lot of stress on the server (Minecraft SMP) and tends to lag players as it backs up, which recently has been taking nearly 5 minutes.So I need to know if there's a way to control the GZip compression rate at the same time that it archives and backs up the files?I understand that I can first tar the files and then GZip them separately with a different compression rate afterwards, but this would not work because it names the files with the current server time, which sometimes changes in between commands.
View 1 Replies View RelatedI need a script that accepts two parameters inputDir and outputDir.
This script should copy all the log files in the inputDir to a folder like <BackupLogs-currentDaysDate>
The new folder with the log files should be tarred and gzipped <BackupLogs-currentDaysDate>.tgz
And this new <BackupLogs-currentDaysDate>.tgz file should be copied to the outputDir.
Also all the log files in the inputDir should be deleted.
I sometimes get confused by the varying command line options I need to run common Unix archiving and compression software (e.g. gzip, bzip2, zip, tar).
Is there a program out there that can just Do What I Mean for common cases? For example:
Now, i thinking solution: Backup Data on Server CentOS to HDD Box. When server error, i can restore Server from HDD Box.
View 1 Replies View RelatedI am trying to backup data files from a Debian Lenny server (in my office) to a second Debian Lenny box at my home. I was going to try rsnapshot configured for daily & weekly backups.My question has to do with access to the backup files. I can get into my hmoe server via ssh (specifically putty & winscp). If I need a specific file, is it easy to log into my home machine and simply find the file(s) I need in the directory tree? Is there some other (simple) method you would recommend instead?
View 1 Replies View RelatedHow to backup in such way that, data will be backed up in every 1 hour...but every time not full data;; only the data/file which have been modified within last 1 hour. Its on locally. And how to back up local data to remote system. Because I have tried with ssh-keygen also , but everytime its requiring password.
View 1 Replies View RelatedI want to have brasero or k3b do a data check after burning a backup cd. I know that in Nero for Windows you just click on the box that says "verify after burning". For brasero I see that you need an md5 file but it will be to tedious to use an md5 utility to make a checksum for every file and then put in in a txt.
View 3 Replies View RelatedI have installed also Vista on my PC, but "Documents and Settings" f.e. is void if I mount the Vista partition. In this case its not very important. But a week ago i'd run a XP-Laptop with 9.04 live-CD for backup. That worked fine (The XP was very damaged). I put the 40GB data on my PC through the WLAN during one night without any aid else. All the XP files are well accessible. And I put it back later on a fresh XP.But what would have happen with a Vista-Laptop? Anyone an idea how to reach all Vista or 7 files with a live CD? Or at least from my installed 8.04 to installed Vista.
I think it works alternatively from Win7 running using a USB hard drive adapter if you change the rights on Vista or 7 on the connected hard drive. Not tested.
how to backup both the system and the data from my Red Hat Server to a share on a windows SBS server where the regular network backup will take it off-site every night. The Red Hat Server is basically a web server on the LAN using RAID 1 to ensure some redundancy in the discs and is used as an internal web server for an administration system.
The Windows SBS server is the domain controller and rest of the network is totally windows based. The SBS server has large discs which host all of the general data to be backed up and is controlled by a 3rd party IT company. We have a share created on the SBS Server which is to backup all of the information from the Red Hat box to ensure we can potentially restore the system quickly in case of a major problem.
The thorn here is that I am working on a different site to the Red Hat Box and have to do everything over SSH. the Red Hat server has Samba running but I have not used it to mount an external windows share? Is there an 'out of the box' backup software on the Red Hat server that I can control over ssh to do the job? The biggest worry for me is the backing up the system to a state where it can be popped on a disc and the system restored from the backup - or most of it anyway - enough to get the server back on its legs.
I installed a fresh version of squeeze, with kde, and it worked fine. After I untarred my backup /home and /etc, things came to a standstill. I got a final ERROR code after /home. After a restart I got a 'fix your system' warning requiring a Control-D, which I couldn't type. I had used tar -xzvpPf ...... If I had included a -w, which asks for confirmation before overwriting data, could I have prevented this problem.
View 1 Replies View RelatedCurrently I have bacula backing up the contents of my server on a daily basis (Mon-Fri) on to a tape. There are in total 9 backup tapes so far. One for each day of the week (Mon-Thurs) and one for every Friday in the month (so there are 5 friday tapes). The tapes are all from the same pool and once the tapes are full they should then get recycled. As I understand and from what I have seen, the data from my server backups up to the tape and is then marked with the append status. Every time that tape is used, data is then appended onto the tape until the tape is full. At this point it is marked as full but wont be recycled until all the tapes from the pool are marked as full (so until all 9 tapes are full).
My question is, is there a way of overwriting the data on the next backup on the tape, rather than appending or requiring that all tapes in a pool are full before a tape can be recycled. As the mon-thurs tapes currently cant be recycled until all the Friday tapes have been used. The reason why I dont have seperate pools for the Friday backups and the other days of the week or even seperate pools for each day of the week, is that if a tape is accidently not changed or someone puts in the incorrect tape by accident a backup will still occur. Which is better than no backup!
I am not able to detect my USB hard drive in Linux rescue mode after passing "fdisk -l" as i want to take the data backup in my USB hard disk.
View 1 Replies View RelatedI accidentally deleted the Thunderbird 3.0 profile in both Ubuntu 10.04 and Windows Xp (home) along with their backups. So i really have 2 questions.1. How to identify which files to Undelete, perhaps using NtfsUndelete?2. How To get Thunderbird to switch to using an old backed up profiles folder.So far the problem with my 2 is that Thunderbird just claims that a version of Thunderbird is already open (it isn't) and then closes down.
View 4 Replies View RelatedI managed to damage my 11.2 installation so it starts in the GUI mode only in failsafe mode. Actually I tried before to repair the installation, using the install DVD, but the automatic repair procedure failed. More than that, since then boot loader also seams to be "repaired" so that the Windows installation doesn't appear in the boot menu, but this is another thing.For me, now, the fastest way to get a stable system is to make a new installation. The biggest problem is that I cannot save/backup the emails and accounts settings in an elegant way. I'm using Thunderbird. Of course I would also like to save other apps settings.So is there a way to save user application settings so that I can used them after a new install? I had a look to the yast backup tool but these seams to be a way to archive files, or am I wrong?
View 9 Replies View RelatedMy primary drive is 250GB and has the root, home and var (I'd read it's good to put var on a separate partition for MythTV which I'm planning on doing) on separate partitions. I have a second 1TB drive that I'll be using to backup the 250GB drive and also host less critical data. I've created two partitions on this drive, one 250GB and the other covering the rest of the drive. I'd like to move the Videos directory out of Home on the 250GB onto the 1TB drive but can't find a definitive way of doing this. Should I just follow this guide for moving the home folder to a new partition? Next question is when performing a backup of the 250GB drive how do I make sure it's going to the 250GB partition on the 1TB drive? Can the different partitions be mounted separately?
View 2 Replies View Related