General :: Tar Backup File Keeps Growing?
Jan 7, 2010
I am new to Linux. I am using tar to backup my emails to a server. I would like to automate this process to routinely backup my emails periodicly, however, i keep running into a problem: I start in the dir I would like to create the tar file (dir size = 240MB). I enter the following command
tar cf bup.mail.llc.tar "/Users/d/Library/Mail/INBOX.mbox/Messages"
file size = 234MB
When I would like to backup my emails into the previously created tar file I use the following command:
tar uf bup.mail.llc.tar "/Users/d/Library/Mail/INBOX.mbox/Messages"file size = 462MB The backup command works, except the size of the original tar file grows, around twice the size. When I extract the updated tar file (file size = 462MB), the unarchived file is 240MB the same size as the original directory.
Why does the size of the tar keep growing each time i perform 'tar uf'?? I don't understand this
View 6 Replies
ADVERTISEMENT
Jun 8, 2010
I have this little problem. I wanna backup a big file that is constantly growing. Is there any way to make one backup and then have some way to take incremental backup of it?
View 2 Replies
View Related
Apr 23, 2010
How to find continuously growing files in the file system?
View 11 Replies
View Related
Sep 13, 2010
Can some one give me a sample of a crontab for backing a directory please, System is Ubuntu 9.04Quote:
#!/bin/bash
# this file is an automated backup script, backup.sh.
# this backs up my domain site.
[code]....
View 7 Replies
View Related
May 20, 2010
Currently running 10.04 and Cacti latest version and although my RAM is only just over 3/4 used my swap file is growing. If I leave the box running for about a week the swap fills and the system grinds to a halt.Is there some way of seeing what's in the swap so I can debug the problem and get my system more stable again?
View 5 Replies
View Related
Mar 29, 2010
I've already asked this in the mythbuntu and didn't get an answer there so I'm trying here. OK just added another 2Tb WD drive to my mdadm controled RAID5 array, and the reshape is finished:-
Code:
Code:
Mar 28 18:28:36 alpha2 kernel: [104357.343421] md: md1: reshape done.
Mar 28 18:28:36 alpha2 kernel: [104357.525114] RAID5 conf printout:
Mar 28 18:28:36 alpha2 kernel: [104357.525119] --- rd:4 wd:4
Mar 28 18:28:36 alpha2 kernel: [104357.525122] disk 0, o:1, dev:sda2
[code]...
View 6 Replies
View Related
Nov 10, 2010
This script simply deletes files older than a certain age (in this case 7 days) from a certain location; I use it to purge old backups nightly, and it works as expected:
# delete backups older than 7 days
find /mnt/backup/* -mtime +7 -exec rm -Rf {} ;
The problem is, every morning I get an email with an error message something like this:
find: `/mnt/backup/subfolder': No such file or directory
View 2 Replies
View Related
Aug 10, 2010
I was wondering if it is possible to create a growing image for a loopback device. Like a file that you create that will grow with the data that is placed on it instead of a file that has to be the size of the entire file system?
View 4 Replies
View Related
Sep 30, 2010
currently my disk-space is growing very-very fast and in the same time I have a very limited amount of it.Last time I had this kind of problem, I had MySql persistence replication is on and disabling the feature fix the problem. I don't know what happened between now and then, the space is shrinking rapidly (600Meg in couple of days) and I only downloaded files for less than 10Meg in the same period.
Could anybody give me a pointer to a tool that can oversee growing directories or files or maybe a script that able to do this (possibly involving Cron). I try using "Find" but I cannot find any files that are suspiciously growing. I suspect it's a directory that is growing, but I don't know.
View 5 Replies
View Related
Jul 20, 2009
An external service I dont manage pushes mediafiles into a shared directory on my server. I need to move these files into their correct directories automatically. The problem is that if I run my script as a cronjob once every 3 minutes, I notice that the script copies files which are still on their way into my server. So I need to figure out how to have the script check that the files are complete (done downloading) before the script moves the files. This is what I got so far:
Code:
#!/bin/sh
# Script by proximity 280709.
# Locate correct directory
[code]....
View 8 Replies
View Related
Aug 18, 2010
The fifth field in the /etc/fstab file specifies dump frequency isn't it ?now this value can be theoretically used by backup products isn't it ?
View 2 Replies
View Related
Mar 23, 2011
I'm going to be using this command to back up my files:
Should I change anything or is it ok?
View 4 Replies
View Related
Apr 13, 2010
O/S: Fedora 12
I am newbie in linux. What I want to do is: Make backup for my file system, cos I learn how to configure servers. So if I made some thing wrong, I want to be able to restore the default setting for my files. Instated of install new O/S.
View 4 Replies
View Related
Oct 19, 2009
I did a backup of the ssd on my eeepc using the following command from a Linux Mint on a USB key:
dd if=/dev/sda1 of=/media/disk/eeepc_save/SYSTEM/system.bck
(/media/disk in an external USB disk)
I deleted the ext2 partition using GPartEd on live USB key and created it back. I rebooted Linux Mint and restored the filesystem using the opposite command :
dd if=/media/disk/eeepc_save/SYSTEM/system.bck of=/dev/sda1
I mounted /dev/sda1 and when I "ls" the root directory, I get several "NFS stale file handle" messages concerning directories (/dev and other). I tried "e2fsck -y", had a bundle of corrections that resulted in the deletion of the directories. I don't use NFS. I did the same for the user filesystem and had no problem (it's an ext3 partition). The two filesystems are the ones that came with the original Xandros installed on my eeepc and that were mounted with union-fs.
View 3 Replies
View Related
Nov 2, 2009
I work for a school consulting company.We helped a school deploy about 1500 computers.The computers have windows XP but we have been using G4L for the restore partition on the drives.So far the software works great. We did however run into a problem in that many of the computers we deployed are missing the restore partition. The reason they are missing is long and convoluted and not really that important. What I have been charged to do is try and fix the restore partition problem. One solution that I had, which im not even sure if it will work, was to backup the recovery file, that g4l created, to DVD and write a basic script to recreate the partition and then copy the file over. This process would need to be as automated as possible since this disc will be inserted by the end user(the students). The backup file that g4l created is 5.9GB so it wont fit on just one disc and Dual layer discs are too expensive to use for this project, so the file will either need to be compressed again (not sure if that's a good idea or not) or split across two DVD's.
I have searched the forums here and I was not able to find anything to fix this problem. I was able to find some info on splitting files across two discs but im not sure how to use that to fix my problem.
View 5 Replies
View Related
Feb 11, 2011
It's always a good to backup a configuration file like sources.list before you edit it. To do so, issue the following command: sudo cp /etc/apt/sources.list /etc/apt/sources.list.backup Where does it backup to and how do I access it? I want to put the backup on removable disk or upload it
View 2 Replies
View Related
Jun 5, 2010
I am IT Manager. Currently I am responsible for a mostly IBM shop, running a mix of AIX, OS/400. VMware, and Windows.
My P-series running AIX and Oracle and my AS/400 both back up to tape drives daily.
But my VMware system is not being backed up, the virtual machines are, but my tape drive just broke down.
I just ordered a server grade machine with an i5 CPU, running 3 2TB SATA drives in RAID5, this is going to be my new backup server.
I am hoping I can put Linux on it, and some sort of open source file replications, real time backup software that will back up my Windows servers, and my Windows desktops in real time.... any sort of animal exist?
View 5 Replies
View Related
May 20, 2010
I am administering a live web server i want to keep a backup of the access log file without disturbing the server performance. can anyone guide me how to to this. the size of teh log file run in GB so i will need to take a daily backup
View 7 Replies
View Related
May 22, 2010
Does the dump command back up entire file-systems or is it capable of backing up subsets of a file-system? And is tar capable of taking device names (for file systems) as input to be archived?
View 1 Replies
View Related
Apr 9, 2011
I have installed an application manager(monitoring application) on my linux server. Now, i need to have backup schedule for my application. The application itself has executive file to backup database.But when i put this file in my crontab to schedule the backup program it wont run!50 09 * * * root /opt/ME/AppManager9/bin/BackupMysqlDB.sh
View 1 Replies
View Related
Mar 3, 2011
How do you create a cron file that will regularly perform a level 0 backup once per month?
View 2 Replies
View Related
Feb 22, 2011
I am having problems gaining access to the BackupPC admin web page. The error:
Quote:
[Tue Feb 22 16:43:59 2011] [error] [client 192.168.0.2] (13)Permission denied: Could not open password file: /etc/BackupPC/htpasswd [Tue Feb 22 16:43:59 2011] [error] [client 82.30.227.113] access to /backuppc failed, reason: verification of user id 'myhtaccessuser' not configured
Why is this occuring, is there anyway of getting around this? I just cant remember originally on my old system how I set this up.
View 3 Replies
View Related
Sep 15, 2009
I am trying to create a backup script that will back up a single folder for a class i am in. I was wandering if I could get some help. If possible I would also like to know how to write a script that can encrypt that same file . I will be putting the back up in my /home/usr/Backup directory. I am not trying to back up my whole system just a single folder. I am using Fedora 11
View 2 Replies
View Related
May 21, 2011
I'm just setting up a partition on a seperate HDD in my system. I plan to use the partition to backup the important files on my main HDD (to guard against HD crash).
The question I have is about where would be the typical location to auto mount this partition? Which would it be normal to go for:
1. /backup/
2. /media/backup/
3. /mnt/backup/
4. /home/chris/backup/
View 7 Replies
View Related
May 13, 2011
I made a script to backup file from each host with general password in local network. This script using SSH Pass and Rsync with this
syntax:
rsync --rsh="sshpass -p password ssh -l root" hostath destinationpath
Everything is okay under 9.10 version until I migrate to Ubuntu 11.04, there is always give an error:
rsync error: received SIGINT, SIGTERM, or SIGHUP (code 20) at rsync.c(541) [Receiver=3.0.7]
I am using bash version: GNU bash, version 4.2.8(1)-release (i686-pc-linux-gnu) and 2.6.38-8-generic kernel
View 3 Replies
View Related
Sep 20, 2015
I want to back up an entire Linux system on a 3Tb external Western DIgital USB3 drive.
I do not want to reformat it from what it is, apparemtly NTFS.
Is there a utility that can act like a file manager like mc, that will permit me to create an ever expanding (to 320Gb) TAR file that will retain all the original file permissions. I have had nothing but disappointment with Linux backup utils with a FAT32 external drive, and I am concerned if I just try an tar the entire drive at once, with around 3 million files, I might run out of memory.
View 5 Replies
View Related
Jul 18, 2010
i am new to linux and i nid a bash script to Delete selected backup files and also restore selected backup file to a user define directory.
View 3 Replies
View Related
Feb 22, 2011
Does webilder just keep growing the collection? Or does it ever get rid of old photos? It seems to me an ideal webilder I could set max 100k, it would delete oldest photos to make room for new. Or, like flickrwall (windows), it would just pull one flickr photo per hour, over writing last photo.
View 1 Replies
View Related
Aug 3, 2010
I was growing a 3 disk raid 5 to a 4 disk raid 6.I used mdadm --grow /dev/md0 --level=6 --raid-devices=4 backup- file=/home/user/Desktop/md0.backup for the grow.The drive I added was /dev/sde1. Well it was going strong for like 2-3 days and last time I checked it was at 73% rebuild.Well the frikin power went out on me and when I came back home I can't get the array assembled.
I just have a fear that i lost my superblocks but I'm also a moron when it comes to this stuff.. I would like to think not, but I'm sure there are much smarter people than I on this issue. Any help would help cause of course I have a lot of stuff on my raid with 3 2tb drives.. the irony is I was converting to raid 6 to make sure I had really good safety net.
View 4 Replies
View Related
Feb 15, 2011
I am trying to build a media server for my home and still in the process of evaluating my OS options (Ubuntu Server, Fedora Core, or Win Server). I am planning to use four 1TB drives initially for the RAID5 array. Once it fill up i will add more 1TB drives.
My question is can Fedora Core create a RAID5 array and grow latter without having to back up data to external hard drive and re-create the array? I am looking for something that is easy to use and manage. If Fedora Core doesnt have this option, can you recommend other distributions that can do this?
View 1 Replies
View Related