General :: Backup Script Causing Data To Be Lost On Main Server- RHEL?
Feb 15, 2010
i setup the following script ro run each night (12am) as a cron on the main server:
if mount | grep -q '/home';
then
rsync -ranv --delete /home/ 138.73.56.12:/home;
[code]....
the main server is running on a dell poweredge 2600. The script rsyncs to a virtualised duplicate running on a hp dl380. when i set this script up and begn running data started going missing from the main server. if new files had been created the files by staff they would go missing, if data was added to existing files prior to activating the script the new changes made to that file would be lost.i just cant understand why this happened. as soon as i turned the script off after a few days it was all back to normal but the data that had gone missing had gone missing.
i just wanted to know if this could be a disk read/write issue, was the script running too soon not allowing data to be written before it can be backed up? could it bee memory? i just dont know. another developed occured after a few days of all this happening, which was one of our hard disk from the main server started mis behaving and flashing amber (attention).
I'm just setting up a partition on a seperate HDD in my system. I plan to use the partition to backup the important files on my main HDD (to guard against HD crash).
The question I have is about where would be the typical location to auto mount this partition? Which would it be normal to go for:
I just installed natty. I use Deja dup to backup my photos to the server in my basement, but now I can't find them. I really cant understand why theyre gone, but they are, maybe because I'm unorganized. Anyways, my hard drive may not be totally overwritten. Can i rescue my old data, or at least parts of it? How?
iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.
Since today morning we are getting this error, could not track down the cause
We are running : Postfix Amavist and Clam
70A6C45A1CE: to=<actualemailid@yahoo.com>, relay=none, delay=0, status=deferred (delivery temporarily suspended: lost connection with 127.0.0.1[127.0.0.1] while sending end of data -- message may be sent more than once).
I already have an ubuntu backup server in my location and need this one server to be backed up remotely in another state. this other location is a helpdesk so there's a danger that they can gain access to confidential data. I'll be setting up this new server as an ftp server but need to set the ftp folder to only allow access to the backup server and me. Because its remote on the helpdesk side, they'll need some access to the file system but need to be completely blocked off from the ftp folder where all the data is at. How can I make sure I can keep them away from my data and still be able to retrieve or copy files over without permission issues between both servers?
we've been trying to become a bit more serious about backup. It seems the better way to do MySQL backup is to use the binlog. However, that binlog is huge! We seem to produce something like 10Gb per month. I'd like to copy the backup to somewhere off the server as I don't feel like there is much to be gained by just copying it to somewhere else on the server. I recently made a full backup which after compression amounted to 2.5Gb and took me 6.5 hours to copy to my own computer ... So that solution doesn't seem practical for the binlog backup.Should we rent another server somewhere? Is it possible to find a server like that really cheap? Or is there some other solution? What are other people's MySQL backup practices?
I installed amanda on a client that uses Red Hat Enterprise Linux 4. When I do an amcheck from the server, I get this: Amanda Backup Client Hosts Check -------------------------------- WARNING: 192.168.3.23: selfcheck request failed: recv error: Connection reset by peer Client check: 1 host checked in 0.022 seconds. 1 problem found. I am using bsdtcp. I get an ACK error if I switch the client to bsd authentication. I read that I should enter servers adress to hosts file in /etc/ directory?
What are the options to backup data to CD/DVD from command line interface. Don't want to use K3B and Nero Linux. Want to backup using "tar" like utility. Please suggest good articles or options.If we can use tar itself to backup data to CD. Then how?For example I want to backup my /home directory to a DVD. I don't want to use GUI tools for this.
I am now preparing myself to upgrade lenny to squeeze and decided to do a backup on my system. I used backup-manager to do the job and it worked fine. how do you restore said backup data?
My wifes computer, IBM desktop, running XP, was hit by malware. I managed to save the data and tarred it and copied to a USB memory stick using my Linux box Ubu 9.1). I checked and saw it had been truly saved. Two days later I put stick in USB port and it was not recognized (usually any hot plugable device shows up immediately on desktop). I have tried plugging it into a windows machine and nothing was found. We are both very desperate at this apparent loss but because something similar happened with an Iomega zip disk.
A directory /test/test1 is created & under /test1 there are some files & subdirectories with data in it. I had copied the files (text & script files) with command as,
cp -irv /test/test1/.* /test as per the requirement but what i see in destinarion i.e, /test that no files or directory has been copied in /test & the files/directories is also removed from source i.e, /test/test1
so my query is how can i get the files/directories back with data?
I have couple hard drives from different linux distro's. on the hdd's are logical volumes created. what i want to do is to backup the data from those hdd's. but when I connect some of them and boot up the system it gives me an error that there are volume group duplicated and it list the UUID. My host system is centos 5. What is the best solution to rename those volume groups or to mount the drives from other systems, so the data will not get lost?
My hard disk is failing and I amnot able to boot into the system! Currently I have logged into the system uing Live CD! Any way to compress and back up the data in my hard disk in an efficient way!
I didn't realize my system drive had filled up... and TaskWarrior overwrote my 'pending.data' with a 0-byte copy. d-: I had worked SO hard to get all my tasks imported from various other notes (many of which I deleted along the way), and hadn't yet prepared for the possibility of data loss. Should've set it up to use my Dropbox...why doesn't TaskWarrior make ANY backup buy default? That data is so important, and yet so small and trivial to backup. Anyway, my 'undo.data' is totally intact and seems to contain all the information theoretically needed to reconstruct 'pending.data'. Unfortunately right now I have no sample 'pending.data' to look at. d-: OR: does anybody know of a simple Linux utility I can use to recover previous file versions? It's possible no data was overwritten, since the replacement is empty.
i'm using fedora core 12 on my laptop vostro 1310. I don't know why every gnome window has lost the main menu. Does someone know the way to turn on again those menus
Ive been running crunchbang on my system, and yesterday decided to try opensuse so i can install the novell client. I ran the live gnome cd and during installation I shrank my main partition down and created a new partition for opensuse. Installation went well other than sometimes programs close and make the part of my desktop that they occupied useless.
Question 1) I can see my crunch bang files through opensuse but i cannot find or boot into crunch bang.
question 2) Why are these programs closing and then killing my desktop? Say i have a Firefox in the top left of my desktop and it errors and closes. I go open Firefox again, anywhere that the old Firefox window was in ( not click able). My cursor will turn to the text cursor as soon as it enters the x y position. I have to restart the laptop to be able to click file or insert a web address.
About 4 years ago I selected a new solution to back up our Linux servers, and have had no problems except that the backups fail twice a year due to DST [1]. After a recent upgrade, I am experiencing what appears to be a memory leak in the beremote process, which grows larger every day. At one point after all backups were finished, the idle beremote process took 2gb of RAM and interfered with the normal function of the computer. Fortunately, the problem is intermittent.
Technical support informed me that they only officially support [2] RHEL up to version 5.4. I was advised to restart the Linux agent every day after the backup has finished.
This could be taken to mean that they do not officially support computers that have automatic updates enabled, since they would bring a system up to version 5.5.
is possible to edited the default RHEL CD to have it automatically install RHEL based off of a kickstart file that I will store locally on the CD. My plan would be to put a cd in a server and have the OS automatically being installed.
I encountered problem on my NEW PROD box. I have remaining space of 300GB and i decided the create a /dev/mapper/VolGroup00 using Redhat Gui. It is successful. Then, i decided to create logical volumes out of it..
We are planning to migrate our LINUX server from RHEL 3to RHEL 5. What are the configuration difference between RHEL 3 to RHEL 5 for webserver installations?
I cannot ssh into an RHEL 5.5 server (192.168.20.104) from another RHEL 5.5 server (192.168.20.101) unless server debug is turned on 192.168.20.104, and even then, I have to wait several minutes before the connection is established. scp to and from the 104 server is also not working.Here is the debug output on the 101 server when server debug is not enabled on the 104 server-:
uBuntu 10.10, Simple Backup My backup stopped working at certain date. Went back in "Simple backup-Configuration", it says, there is no configuration. What might cause trouble like this? Should I seek for more secure back up tool?
I have a database server running RHEL 5.1 32 bit that suffered some catastrophic failures about 6 months ago. We were able to patch it back together and keep it running, but now the manufacturing site it supports is going to shut down for two weeks and I would like to replace it permenantly. Does anyone have any guidance for that sort of thing? I'd like to have the new server up and running before hand, basically changing the hostname/ip and restoring the databases only on conversion day. I've done this in the past with HP UX - Red Hat conversions, but this is my first red hat to red hat move. Any advice or shortcuts?I forgot to add the other wrinkle. The new server will be running 64bit linux.
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.