Software :: File Archiving Software - For Backups ?

Jul 29, 2010

I'm looking for ... or will develop if I cannot find any ... software that can keep archives of a tree of files. The tree of files is on a computer dedicated to backups. Various servers and desktop users periodically run rsync to update their slice of that tree of files. But that means old files are replaced. What I want to do is keep the old files around, with some limitations (such as a finite number of versions, finite age, etc). This would be similar to a repository like Subversion, except that data only comes in (there are no checkouts, though obviously a restore mechanism is needed). It would be like rolling backups, except the "rollover" would be on a per-file basis.

What I currently do, but want to get away from, is making a hardlink replica tree every few days with a command like "cp -al synctree/. archtree-for-today". But that eats up inodes very rapidly and is much harder to find things in. Making a file listing is very much slower this way, too.

View 3 Replies


ADVERTISEMENT

General :: Get Mail And File Archiving Software?

Jun 23, 2010

Which are the Open Source "file and email archiving" software for both Linux and Windows equivalent to Enterprise Vault Symantec?

View 2 Replies View Related

General :: Ubuntu Slows To Crawl When Un-archiving Big File

Aug 4, 2010

I started unarchiving a RAR file that this several gigabytes big. The computer is now going really slow, it's almost frozen. Sometimes I can move the mouse a little, but that's it. The unarchiving process seems to have halted, so now all I can do is restart the system. I don't think I can unarchive this file in Linux.

View 5 Replies View Related

Ubuntu :: Archiving Many (>100000) Files Into A Single File Via Cli

Mar 4, 2011

I'm trying to zip or rar >100000 files into a single file so that I can upload it to my server much faster than ftp downloaded it. Total they're all only 4gb, but because of the number of files Nautilus freezes just opening the folder they're in. They're all .jpgs and all in the same folder and I've tried a few commands but I keep getting error messages.

Anyone have a command that will archive all the files from a folder into a single zip (or rar, tar etc)? I can't just archive the folder because then I would have to move all the files out of that folder and just opening the folder to move them would crash it, and I don't have ssh into that server.

View 3 Replies View Related

General :: Remote File Copy Through Archiving Software ?

Oct 6, 2010

I need to back up a fold on a remote machine to my local box; the remote hd does not have enough space archive it, neither does my local box. I know there's a cantrip to pipe scp through gzip (or similar), but I don't remember the syntax.

View 1 Replies View Related

Software :: Mysql Full Backup Script Not Archiving The Dump File?

Mar 20, 2011

I have a problem with a script i wrote, the script runs fine if manually executed however it doesn't run *fully* when executed via cron

here's the script :

Code:

#!/bin/bash
FILENAME=mysql_full_dump_`date '+%m.%d.%y'`.sql
`which mysqldump` --all-databases -uroot -p************ -h127.0.0.1 > /root/$FILENAME
RETVAL=$?

[code]....

the script resides in /root/bin and the cron entry is as follows:

Code:

0 0 * * * root "/root/bin/mysql_daily.sh"

the result is the .sql file, but it doesn't archive it.

View 2 Replies View Related

CentOS 5 :: Scheduled Unattended Backups With Alerts If The Backups Fail

Feb 3, 2011

I've been a DOS/Windows guy for 20 years, and recently became a SW test lab helper. My company uses CentOS for a lot, so I've become familiar with it, but obviously not as comfortable as I am with Windows.

Here's what I have planned:

machine: Core 2 Duo E8400, 8GB DDR2, 60GB SSD OS drive, ATI 4650 video card, other storage is flexible (I have 3 1TB drives and 4 750GB drives around that can be used in this machine.)

uses: HTPC, Network Storage, VMWare server host: SMTP, FTP server, and Web server virtual machines

I've figured out how to do much of this, but I haven't figured out how to do backups in Linux. I've been spoiled with Windows, with the built in backup system so simple to use. I find myself overwhelmed with the array of backup software, and unable to determine which to use. none of them seem to do everything I need them to do, but some come close, I think. I'm hoping someone here can help me out in figuring out which program to use and how to use it.

Here is what I need the backup software to do:
1. scheduled unattended backups, with alerts if the backups fail
2. a weekly full backup with incremental every 12 hours
3. removing the old backups when the new full backup runs, I would prefer to keep 2 weeks of backups, but that's not necessary
4. a GUI would be preferable, since my arthritic fingers don't always do as I want them to do. I typo things a lot, and the label worn off my backspace can attest to that.

View 7 Replies View Related

General :: Rsync Incremental Backups Rather Than Full Backups?

Nov 12, 2009

How do you get Rsync to do incremental backups rather than full backups? At the moment I have a script that will create a backup folder (if it doesnt already exist) then copy the source files into the backup directory with the command

rsync $VERBOSE --exclude=$TARGET/ $EXCLUDE --exclude '/Ls-wtgl1c8/**' -rt --delete $source/ $TARGET/$source/ >> $LOG_FILE

Target is where the files will be backed up to Sources is the dir(s) to be backed up Exclude files is the list of files not to backup
log file is where the output will be saved to. At the moment it only does full backups, but I would only like to do incremental, how would this be achieved? Am I missing out an option in the Rsync that is required.

View 9 Replies View Related

Fedora :: File Backups - With Suffix

Dec 1, 2009

As you know everywhere we create a new file ,immediately a backup file with suffix ~ will be created with it if we open the file , i deleted a file but always i have problem with its backup how can delete them with deleting the file?

View 7 Replies View Related

Ubuntu :: Getting The File System For Backups?

Nov 25, 2010

I'm having some trouble to find a file system that allows me to backup my data to an external HD, and then access this data in other Linux (and sometimes windows and mac) machines.

The file system that I'm looking for must:

- have no user permissions: anyone can do anything with the data;

- have support to large files: I've used FAT for a while but it just sucks;

- maybe support access in windows and or mac with additional drivers;

- have journaling (or something of the sorts) to reduce the risk of data loss.

View 8 Replies View Related

Debian :: File Integrity For Backups Of Home Data ?

Nov 17, 2010

I'm currently backupping our home data (pictures, videos, our CDs ripped to FLAC which I spent a lot of time to tag accurately ), totalling almost 300 Gb, on 2 external USB drives, one of which is meant to stay at a friend's. I left the factory msdos filesystem as it was, thinking it could be useful to be able to connect the drives to a windows machine with no problems. It's certainly useful to have «normal» data that I can take with me e.g. when visiting my family.

I'm simply using rsync manually, checking for suspicious changed or deleted file before commiting the change. I do that every 2 weeks or so.

Now I want to add a file integrity management to my backupping scheme: I want to be able to check that new data I'll be committing has not been tampered with (integrity check before updating tags on my main drive), and I want to be able to check that backupped data is still sane on my USB drives, especially if I need to recover from data corruption on my main drive.

Since I'm essentially mirroring the data, I thought run of the mill integrity software would let me just rsync the integrity database, and I'm done.

But after browsing through the docs of tripware, afick and the like, I fear they work only with absolute paths, so the database for my main drive wouldn't work for my USB drive, that's mounted elsewhere when I plug it in, obviously.

So, I feel I'm missing something. It looks to me I'm trying to solve a very common problem, how do people do it?

Did I miss a file integrity software that works with backups?

Is there a trick like using a symbolic link pointing to whatever file hierarchy I want to check, and have tripware/afick/... monitor that link?

Should I run a more elaborate backupping system than plain rsync? Which one? (Storebackup for instance looks promising since it involves md5 sums, but it's targetting a completely different problem, and I'm not sure I can use it at all for what I need.)

View 3 Replies View Related

Red Hat :: Count How Many Backups Are Successful/failed Based On The File

Jun 28, 2010

I just wanted to count how many backups are successful/failed based on the file. My file look like following..Quote:

# Directory Backup was successful on server1
# Directory Backup was successful on server2
# Directory Backup was failed on server3

[code]....

View 4 Replies View Related

Server :: Setup House Mostly (for The First Time) For Backups And File Sharing?

Jan 28, 2011

**Edit: path for mount was incorrect Distro Server: CentOS 5.5

Clients:
Fedora(latest)
OSX(latest)

Backround I am attempting to setup a server in my house mostly(for the first time) for backups and file sharing. It is important to me that file permissions are preserved. So its my understanding that I must use idmapd in order for this to work. As of now I'm only working with the linux distros while osx will be dealt with once these two work together. portmapper is up and running, along with lockd on both machines. Firewalls are also down on both machines for now. The server side was all setup using the GUI interface with no extra options selected. Problem When attempting to "mount -t nfs4 10.0.0.2/$sharedfolder /mnt" I get an error operation not permitted with no error printing in /var/log/message. If I use "mount -t -o nolock nfs4 10.0.0.2/$sharedfolder /mnt" it mounts just fine. Ive checked both machines multiple times to make sure that lockd is up and running. In the idmapd.conf file I the domain as "localdomain" for both machines but I doubt that is right; like I stated above this is my first attempt at a server. I'm assuming the problem is a whole missing step that involves some kind of id mapping server I need to setup.

View 5 Replies View Related

Ubuntu Security :: Rsync Automated Backups Of Entire File Structure Over Ssh?

May 16, 2011

I am in the process of writing an rsync script to run unattended backups of my entire file system to another system located on my local network using ssh and password-less rsa keys.

I will absolutely will not use password-less keys with the root account and this is the limitation preventing me from accomplishing my goal because root is required by rsync to access the / tree and copy it to another location. I decided that if I compiled the script into a binary that I didn't have a problem with the password being contained within the binary itself but from what I've read there is no way to elevate to root and then back down to user level from within the script/binary.

I can create the script as the user and use chroot to make it owned by root but retain execution permission for the user but it will still cause the ssh login to be under root and therefore require either that I am there to enter my password or the use of password-less keys under the root account which I reiterate I will NOT do. Currently the script is executed by the user on the machine containing the files to be backed up.

View 9 Replies View Related

General :: Can Switching File Systems Affect Email Backups And Other Data?

Nov 27, 2010

A while back I installed Dreamlinux 3.5 Gnome edition using ext2. When I attempted to use the email address books I imported from the Dreamlinux3.5 XFCE edition, which had been ext3, I discovered that none of the email addresses could be mailed to. I had to manually type in the addresses.

When I reinstalled Dreamlinux 3.5 Gnome using ext3, the same backup files that did not work in ext2 now work just fine. The question is, was this a "broken data" problem caused by the switch to ext2 file system or something else? Has anyone else experienced this?

The mail program is Thunderbird.

View 2 Replies View Related

Security :: Secure And Automated Backups - Add Public Key To Authorized_hosts File On Prod Server?

Mar 13, 2010

I'm trying to find a secure way to backup files on my Prod Server to Backup Server. It must be automated, so I will need to run a command with cron which will login to Prod Server from Backup Server and backup data. 1. Do you think it would be secure enough to do this by creating an passwordless RSA private key on Backup Server and adding it's public key to authorized_hosts file on Prod Server? I can't think of a way to Automate this without having to enter any passwords without passwordless RSA key. Is there another. more secure way? 2. Should I create a special user for backup, which will only have read access to all files in the directory that I am backing up? If so, How can I run a check that this new backup user indeed has read access to ALL files in the folder that I intent to back up? How can I ensure the backup process will not skip files due to some permission problem? 3. I'm thinking of using rsnapshot tool, which uses rsync.

View 10 Replies View Related

Ubuntu Networking :: Create File Sharing Between MacBook Pro And Lucid Box For Time Machine Backups?

Mar 19, 2011

I've been trying to create file sharing between my MacBook Pro and my Lucid Lynx box for time machine backups and media server purposes. I followed this guide:[URL]..Everything seems to work with these exceptions: I can see my LucidLynx box in my finder app in my Mac but only when I run these commands from Ubuntu:

Code:
sudo /etc/init.d/netatalk restart
sudo restart avahi-daemon

If I restart my LucidLynx box then I can't see anything in finder. I can't log into my LucidLynx box from finder. I don't get a bad username or password error it just tells me the connection failed. *Note if I do enter an incorrect username or password it WILL tell me it's incorrect. I've looked at this link below since some people have used it in theses forums but it's a bit dated[URL]..

View 2 Replies View Related

Software :: PDF Editing And Archiving

Jul 22, 2010

I'm looking for an application that will give me some advanced tools for editing PDFs.

Here are the features I'm looking for:
Editing metadata (tagging with keywords)
Merge multiple PDFs
Rearrange page order
PDF bookmarking
Optical Character Recognition

I can give further clarification on these items if needed. My goal is to convert all of my paper files into digital files that I can store on my server. In order to effectively do this, I need the tools listed above.

Is there anything in the linux world that will give me these PDF editing abilities?

View 4 Replies View Related

Ubuntu :: Evolution 2.28 - Archiving And Backup

Jan 31, 2011

I want to create retrievable archives of a my old emails say monthly to avoid the old emails running me out of memory. If I use the 'Backup Settings' procedure as explained in Evolution Help what happens when I wish to consult the /home/dbus/evolution-backup.tar.gz archive file?

1- Will it simply over-write my current Evolution data? [in which case its not what I need]
2- If not, how do I return the archive file to dead storeage and resuscitate my current data?
3- If it will overwrite using the 'Restore Evolution' procedure given in Evolution Help is there a workaround ... perhaps by ...
3a- renaming the archive file,
3b- or 'restoring' it in another version of Evolution,
3c- or archiving CURRENT data as a 2nd backup with a different name [eg: /home/dbus/evolution-backupJan11.tar.gz?] then restoring that?
4- Will I be able to retrieve successive archives if I rename them, say '/home/dbus/evolution-backup.tarDec10.gz' etc once Evolution's saved them?

Alternatively the following came from a dead thread [from commonlyUNIQU3 ] ... is it still valid? does it avoid the problem of potentially running out of memory?
A. Make a subfolder to the "Inbox" under the "On This Computer" tree (I call mine "Archive")
B. Drag and drop the emails you want to archive into this folder.*

This will move the selected emails off of your Exchange server/account and into this folder (and into local storage) - unless you do a copy & paste instead of drag & drop. You may need to have the setting for downloading emails for offline access enabled for this to work as desired. If I recall, the new integrated backup feature creates a compressed (.zip) archive from which you can later restore the email (haven't tried that just yet).

View 5 Replies View Related

General :: Archiving Bash - Script In VIM ?

Dec 14, 2010

I'm relatively new to Linux in general but have learned to do the basics with the CLI.

Well my main problem is writing my first "real" script in VIM. I just have no idea where to start. I was hoping you guys could point me in the right direction.

Well this is what the script needs to do.

"As the IT administrator for a large manufacturing company you have been tasked with producing a script for archiving the daily log files. Each week the daily log files are placed in a directory named weekXX, where XX is the number of the week. Each week directory should be archived and compressed and stored in a folder called log_directories. When the script has completed the task it should display on the monitor exactly what files and directories it has archived.

The script should be started by a user with the required week numbers added as arguments (e.g prog 13 14 15 should start the program and archive the daily log files in week13, week14 and week15).

A basic manual is required, showing how to start the program, the hardware and / or software requirements, a hard copy of the script and a brief description of the test strategy and test data."

View 14 Replies View Related

General :: Automate Archiving Of Log Files Using Tar?

Mar 29, 2011

I need to tar this logs, but i dont how to make it simplier to me. Everyday there are created this five logs. I need to make five tar files from every day from this files at the end of the month

For example

Till now i have tar it manualy (copied every file)

View 2 Replies View Related

General :: Difference Between Archiving And Backup Of Data?

Feb 24, 2010

I am a little confused in distinguishing Backup, compressing and Archiving data. Help me to figure out how these can be useful.

View 2 Replies View Related

General :: Control Gzip Compression Rates While Archiving Files With Tar?

Jun 4, 2011

I have a script which periodically backs up a directory using the command "tar -czvf [name] [directory]" but my problem is that the script has recently been putting a lot of stress on the server (Minecraft SMP) and tends to lag players as it backs up, which recently has been taking nearly 5 minutes.So I need to know if there's a way to control the GZip compression rate at the same time that it archives and backs up the files?I understand that I can first tar the files and then GZip them separately with a different compression rate afterwards, but this would not work because it names the files with the current server time, which sometimes changes in between commands.

View 1 Replies View Related

General :: Difference In Size Of The Tarred Files After Archiving Using Tar Command?

May 17, 2011

i am using red hat linux 2.4 . I have 3 folders dir1 dir2 dir3 I have tarred them like this.

1.tar cvfz tarball_1.tgz dir1 dir2 dir3

2.tar cvfz tarball_2.tgz dir1 dir2 dir3 2>& /dev/null (So that it does not display any error message or operation details to the user)

[usr@machine]$ ls -lrt
-rw-r--r-- 1 usr grp 199843988 May 17 13:39 tarball_1.tgz
-rw-r--r-- 1 usr grp 199837488 May 17 13:53 tarball_2.tgz

But can any one explain the size difference as seen in list output...

View 4 Replies View Related

General :: Archiving Log Files - Script That Accepts Two Parameters InputDir And OutputDir

Feb 28, 2011

I need a script that accepts two parameters inputDir and outputDir.

This script should copy all the log files in the inputDir to a folder like <BackupLogs-currentDaysDate>

The new folder with the log files should be tarred and gzipped <BackupLogs-currentDaysDate>.tgz

And this new <BackupLogs-currentDaysDate>.tgz file should be copied to the outputDir.

Also all the log files in the inputDir should be deleted.

View 16 Replies View Related

General :: Command Line - Unix - Program That Can Handle All Popular Compression/archiving Formats - E.g.tar, Gzip, Bzip2, Zip?

Jun 22, 2011

I sometimes get confused by the varying command line options I need to run common Unix archiving and compression software (e.g. gzip, bzip2, zip, tar).

Is there a program out there that can just Do What I Mean for common cases? For example:

View 2 Replies View Related

Software :: Document And Project Archiving Software ?

Oct 26, 2010

At my work place, my colleagues or I create/start bunch of small projects.

Each of these small projects contain some document (pdf, word and etc.) and/or different types of documents.

I am looking for a simple software (preferred GPL) which can be run on Linux and/or Windows and can manage to maintain those small projects.

It can be a web based software, which would assign an unique number to the each new created project and then would create the necessary links, which would point to the necessary documents, as I mentioned above.

View 2 Replies View Related

Ubuntu :: DVD / Can't Read Backups?

Nov 17, 2010

I backed my photographs with K3B in Suse, prior to installing Ubuntu. Now I can't access them. Well I can but not with Ubuntu or Windows 7. Suse (KDE) on the family computer reads them with no problem?? I'm confused. Is there something that KDE has that Debian doesn't. there's nothing wrong with my system exept for this. Reads cd's and dvd's.

View 2 Replies View Related

Server :: Cron Job That Does The Backups?

Mar 21, 2010

I have a crontab related question which I am hoping someone can answer. I recently took over a Redhat Enterprise 5 Server, and I was told by the previous Server Admin that there is a cron job that does the backups. I ran the following command to get a list of all users:

Code:
cat /etc/passwd | grep "/home" |cut -d: -f1

I then ran the following command for each of those users to see if they have any crontabs associated with them:

Code:
crontab -u USER -l

It doesn't show any crontab entries for any users (including root). But I am positive that there is a scheduled job somewhere because the backups are still running every night.

View 6 Replies View Related

Server :: Use FileSystem For Backups?

Jul 8, 2010

With so many filesystems available which one should I use to make backups? All I care about is reliability and stability. I don't care at all about portability.

View 8 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved